Abstract
In this article we present the , , approximation properties of activated singular integral operators over the real line. We establish their approximation to the unit operator with rates. The kernels here come from neural network activation functions and we employ the related density functions. The derived inequalities use the high order modulus of smoothness.
Keywords:
activation functions from neural networks; Lp approximation; singular integral; Lp modulus of smoothness MSC:
26D15; 41A17; 41A30; 41A35
1. Introduction
The approximation properties of singular integrals have been established earlier in [1,2,3,4]. The classic monograph [5], Ch. 15, inspires us and is the driving force in this paper. Here we study some activated singular integral operators over and we determine the degree of their , , approximation to the unit operator with rates by the use of smooth functions. We derive related inequalities involving the high , , modulus of smoothness. Our studied operators are not in general positive. The surprising fact here is the reverse process from applied mathematics to theoretical ones. Our kernels here are derived by density functions coming from activation functions related to neural networks approximation, see [6,7]. Of great interest and motivating the author are also the articles [8,9,10,11,12]. In recent intense mathematical activity by the use of neural networks in solving numerically differential equations our current work is expected to play a pivotal role, as in the classic case played the earlier versions of singular integrals.
Regarding the history of the topic we make reference to the 2012 monograph [5] from 2012, which was the first comprehensive work to address the traditional theory of approximation by singular integral operators to the identity-unit operator in its entirety. The fundamental approximation features of the generic Picard, Gauss-Weierstrass, Poisson-Cauchy and Trigonometric singular integral operators over the real line were presented. These are not positive linear operators. They specifically looked into the rate at which these operators converge to the unit operator and their associated simultaneous approximation. This is provided by use of high order modulus of smoothness of the high order derivative of the engaged function via inequalities. It has been shown that some of these inequalities are sharp, in fact they are attained.
2. Essential Background
Everything in this section comes from [5], Ch. 15. In the following we mention and deal with the smooth general singular integral operators defined as follows.
For and , we set
that is . Let , and let be Borel probability measures on .
Let and , ; we define for , the integral
The operators are not in general positive operators; see [5].
We notice that , c constant, and
We need the rth -modulus of smoothness
where
see [13], p. 44. Here, we have , .
We need to introduce
Call
Notice also that
According to [5], we get
Thus,
Using Taylor’s formula, one has
where
Assume
Using the above terminology, we derive
where
We mention the first result.
Theorem 1
([5]). Let , such that , and the rest as above. Furthermore, assume that
Then,
If , ∀, , and as we get that .
The counterpart of Theorem 1 follows in case of .
Theorem 2
([5]). Let and , . Assume that
Then,
Additionally, assume that
∀. Hence, as , we obtain .
The case follows.
Proposition 1.
Let , such that , and the rest as above. Assume that
Then,
Additionally, assume that , , ∀; then, as , we obtain unit operator I in the norm, .
We finally need
Proposition 2.
Assume
Then,
Additionally, assuming that
∀, we obtain as that in the norm.
We will apply the above theory to our activated singular integral operators; see Section 5.
3. Basics of Activation Functions
Here everything comes from [14].
3.1. On Richards’s Curve
Here, we follow [7], Chapter 1.
A Richards is curve is
which is strictly increasing on , and it is a sigmoid function; in particular, this is a generalized logistic function. And it is an activation function in neural networks; see [7], chapter 1.
It is
We consider the function
which is , and all .
It is
and
We also have
We also get
and G is a bell symmetric function with maximum
Theorem 3.
It holds that
Theorem 4.
It holds that
So, G is a density function.
We make
Remark 1.
So, we have
(i) Let . That is, . Applying the mean value theorem, we get:
where .
Notice that
(ii) Now, let . That is, . Applying again the mean value theorem we get:
where .
Hence, we derive that
Consequently, we proved that
Let ; it holds that
Clearly, by Theorem 4, we have that
So that is a density function, and let , that is is a Borel probability measure.
We give the following essential result.
Theorem 5.
Let , and
Then, are finite and as .
In fact it holds that
for .
Next we present
Theorem 6.
It holds that
for
Also, this integral converges to zero, as .
In fact, it holds that
3.2. On the q-Deformed and -Parametrized Hyperbolic Tangent Function
We consider the activation function , and study its related properties; all of the basics come from [7], ch. 17.
Let the activation function be
It is
and
with
We consider the function
∀, . We have , so that the x-axis is a horizontal asymptote.
It holds that
and
The maximum is
Theorem 7.
We have that
Theorem 8.
It holds that
So, is a density function on ; .
Remark 2.
(i) Let . That is, . By the mean value theorem we obtain
for some .
But , and
.
That is,
Set , then
(ii) Let now . That is, . Again, we have
.
We have
and
Hence,
Therefore, it holds that
That is
Set ; then,
We have proved that
∀.
Let ; it holds that
By Theorem 8, we have
So that is a density function, and let
that is is a Borel probability measure.
We give
Theorem 9.
Let
Then, are finite and as .
In fact, it holds that
It also follows
Theorem 10.
It holds that ( ; )
and it converges to zero, as .
3.3. On the Gudermannian Generated Activation Function
Here, we follow [6], Ch. 2.
Let the related normalized generator sigmoid function:
and the neural network activation function be
We mention
Theorem 11.
It holds that
So that is a density function.
By [6], p. 49, we found that
But
∀.
Therefore, it is
So here it is
the related Borel probability measure.
We give the following results, their proofs as similar to Theorems 5, 6 are omitted.
Theorem 12.
Let , and
Then, are finite and , as .
Theorem 13.
It holds
; .
Also, this integral converges to zero, as .
3.4. On the q-Deformed and -Parametrized Logistic Type Activation Function
Here, all come from [7], Ch. 15.
The activation function now is
where .
The density function here will be
We mention
Theorem 14.
It holds that
By [7], p. 373, we have
So, here, it is
the related Borel probability measure.
We give the following results, their proofs as similar to Theorems 9, 10 are omitted.
Theorem 15.
Let
Then, are finite and , as .
Theorem 16.
It holds that
where ; .
Also, , as .
3.5. On the q-Deformed and -Parametrized Half Hyperbolic Tangent Function
Here, all come from [7], Ch. 19.
The activation function now is
where .
The corresponding density function will be
It holds
Theorem 17.
By [7], p. 481, we have that
Thus, here, it is
the related Borel probability measure.
We state the following results; their proofs as similar to Theorems 9, 10 are omitted.
Theorem 18.
Let
Then, are finite, and , as .
Theorem 19.
It holds that
where ; .
Also, , as .
4. More on Activation Probability Measures
We present
Theorem 20.
Let , , , , , be the ceiling of the number, and . It holds that
Proof.
We have, in general:
□
We continue with
Theorem 21.
Let , . It holds that
Proof.
We have that
□
We continue with
Proposition 3.
Let . It holds that
Proof.
We have that
□
Proposition 4.
Let , . Then,
Proof.
We have that
□
We continue with the following results.
Theorem 22.
All as in Theorem 20. Then,
where .
Proof.
Similar to Theorem 20 and (70). □
Theorem 23.
Let , . Then,
Proof.
Similar to Theorem 21 and (70). □
Proposition 5.
Let . It holds that
Proof.
Similar to Proposition 3 and (70). □
Proposition 6.
Let , . Then,
Proof.
Similar to Proposition 4 and (70). □
We continue with more related results.
Theorem 24.
Let , , , . Then, there exists such that
Proof.
Similar to Theorem 20. □
Theorem 25.
Let , . Then, there exists such that
Proof.
Similar to Theorem 21. □
Proposition 7.
Let . Then,
Proof.
As in Proposition 3. □
Proposition 8.
Let , . Then,
Proof.
As in Proposition 4. □
More needed results:
Theorem 26.
Let , , , , . Then, there exists :
Proof.
Similar to Theorem 22. □
Theorem 27.
Let , . Then, there exists :
Proof.
Similar to Theorem 23. □
Proposition 9.
Let . Then,
Proof.
As in Proposition 5. □
Proposition 10.
Let , . Then,
Proof.
As in Proposition 6. □
Furthermore, we have the following.
Theorem 28.
Let , , , ; . Then, there exists
Proof.
Similar to Theorem 22. □
Theorem 29.
Let , . Then, there exists
Proof.
Similar to Theorem 23. □
Proposition 11.
Let . Then,
Proof.
As in Proposition 5. □
Proposition 12.
Let , . Then,
Proof.
As in Proposition 6. □
5. Main Results
Here, we describe the , approximation properties of the following activated singular integral operators, which are special cases of ; see (2). Their definitions are based on Section 3 and Section 4. Basically, we apply our listed results in Section 2.
We give the following results, grouped by operator.
Theorem 30.
Let , .
Call
Then,
and , as .
Proof.
By Theorems 1, 5, and 20. □
Theorem 31.
Let , . Then,
and , as .
Proof.
By Theorems 2, 5 and 21. □
Proposition 13.
Let . Then,
and in norm, , as .
Proof.
By Propositions 1 and 4, and Theorem 5. □
Proposition 14.
It holds
and in norm, as .
Proof.
By Propositions 2 and 3, and Theorem 5. □
We continue with the set of results for operator, .
Theorem 32.
Let , .
Call
Then,
and , as .
Proof.
By Theorems 1, 9, and 22. □
Theorem 33.
Let , . Then,
and , as .
Proof.
By Theorems 2, 9, and 23. □
Proposition 15.
Let . Then,
and in norm, , as .
Proof.
By Propositions 1 and 6, and Theorem 9. □
Proposition 16.
It holds that
and in norm, as .
Proof.
By Propositions 2 and 5, and Theorem 9. □
We continue with the set of results for operator.
Theorem 34.
Let , .
Call
Then,
and , as .
Proof.
By Theorems 1, 12, and 24. □
Theorem 35.
Let , . Then,
and , as .
Proof.
By Theorems 2, 12, and 25. □
Proposition 17.
Let . Then,
and in norm, , as .
Proof.
By Propositions 1 and 8, and Theorem 12. □
Proposition 18.
It holds that
and in norm, as .
Proof.
By Propositions 2 and 7, and Theorem 12. □
We continue with the set of results for operator, .
Theorem 36.
Let , .
Call
Then,
and , as .
Proof.
By Theorems 1, 15, and 26. □
Theorem 37.
Let , . Then,
and , as .
Proof.
By Theorems 2, 15, and 27. □
Proposition 19.
Let . Then,
and in norm, , as .
Proof.
By Propositions 1 and 10, and Theorem 15. □
Proposition 20.
It holds that
and in norm, as .
Proof.
By Propositions 2 and 9, and Theorem 15. □
We finish with operator results, .
Theorem 38.
Let , .
Call
Then,
and , as .
Proof.
By Theorems 1, 18, and 28. □
Theorem 39.
Let , . Then,
and , as .
Proof.
By Theorems 2, 18, and 29. □
Proposition 21.
Let . Then,
and in norm, , as .
Proof.
By Propositions 1 and 12, and Theorem 18. □
Proposition 22.
It holds that
and in norm, as .
Proof.
By Propositions 2, 11, and Theorem 18.
6. Conclusions
Here, we presented the new idea of going from the neural networks main tools, the activation functions, to singular integrals approximation. That is the rare case of employing applied mathematics to theoretical ones.
Funding
This research received no external funding.
Data Availability Statement
No new data were created or analyzed in this study.
Conflicts of Interest
The author declares no conflicts of interest.
References
- Anastassiou, G.A.; Gal, S. Convergence of generalized singular integrals to the unit, univariate case. Math. Inequalities Appl. 2000, 3, 511–518. [Google Scholar] [CrossRef]
- Gal, S.G. Remark on the degree of approximation of continuous functions by singular integrals. Math. Nachrichten 1993, 164, 197–199. [Google Scholar] [CrossRef]
- Gal, S.G. Degree of approximation of continuous functions by some singular integrals. Rev. Anal. Numér. Théor. Approx. 1998, 27, 251–261. [Google Scholar]
- Mohapatra, R.N.; Rodriguez, R.S. On the rate of convergence of singular integrals for Hölder continuous functions. Math. Nachrichten 1990, 149, 117–124. [Google Scholar] [CrossRef]
- Anastassiou, G.; Mezei, R. Approximation by Singular Integrals; Cambridge Scientific Publishers: Cambridge, UK, 2012. [Google Scholar]
- Anastassiou, G.A. Banach Space Valued Neural Network; Springer: Berlin/Heidelberg, Germany; New York, NY, USA, 2023. [Google Scholar]
- Anastassiou, G.A. Parametrized, Deformed and General Neural Networks; Springer: Berlin/Heidelberg, Germany; New York, NY, USA, 2023. [Google Scholar]
- Aral, A. On a generalized Gauss Weierstrass singular integral. Fasc. Math. 2005, 35, 23–33. [Google Scholar]
- Aral, A. Pointwise approximation by the generalization of Picard and Gauss-Weierstrass singular integrals. J. Concr. Appl. Math. 2008, 6, 327–339. [Google Scholar]
- Aral, A. On generalized Picard integral operators. In Advances in Summability and Approximation Theory; Springer: Singapore, 2018; pp. 157–168. [Google Scholar]
- Aral, A.; Deniz, E.; Erbay, H. The Picard and Gauss-Weiertrass singular integrals in (p,q)-calculus. Bull. Malays. Math. Sci. Soc. 2020, 43, 1569–1583. [Google Scholar] [CrossRef]
- Aral, A.; Gal, S.G. q-generalizations of the Picard and Gauss-Weierstrass singular integrals. Taiwan. J. Math. 2008, 12, 2501–2515. [Google Scholar] [CrossRef]
- DeVore, R.A.; Lorentz, G.G. Constructive Approximation; Springer: Berlin/Heidelberg, Germany; New York, NY, USA, 1993; Volume 303. [Google Scholar]
- Anastassiou, G.A. Quantitative Uniform Approximation by Activated Singular Operators. Mathematics 2023, 12, 2152. [Google Scholar] [CrossRef]
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content. |
© 2024 by the author. Licensee MDPI, Basel, Switzerland. This article is an open access article distributed under the terms and conditions of the Creative Commons Attribution (CC BY) license (https://creativecommons.org/licenses/by/4.0/).