1. Introduction
With the advancement of science and technology and the development of societal needs, computer technology has penetrated various fields such as industry, daily life, and the economy, becoming a primary computational tool. However, as a computational tool, computer technology has inherent limitations, namely limited computational resources. Even today, for vast amounts of data, direct processing by computers is exceedingly slow, and the associated costs are undoubtedly enormous. Therefore, processing data through different algorithms becomes crucial, and finding the algorithm with the lowest cost is even more important. This has driven the development of the field of computational complexity. Computational complexity can be understood as the study of finding the minimal computational cost required to solve a problem within a reasonable error margin (see [
1]).
To better understand and address computational complexity, researchers have proposed various computational models and settings, with the most common ones being the worst case, average case, and probabilistic settings. These three options provide different perspectives for analyzing and optimizing algorithm performance.
Width is an important concept in the theory of function approximation, and it plays a vital role in the study of computational complexity (see [
2]). Width is commonly used to measure the optimal or minimum error achievable when approximating a known function in a function space using a finite number of information points. Essentially, studying the width is equivalent to studying computational complexity. In different settings, the width provides a quantifiable measure of the complexity of a problem. Therefore, width is not only a theoretical tool for quantifying the limits of algorithmic approximation capabilities but also closely linked to computational complexity. By studying the width under different settings, we can gain a clearer understanding of a problem’s complexity and design more efficient algorithms to solve these problems.
This paper primarily investigates the approximation problem of weighted Sobolev spaces on a sphere , considering both probabilistic and average case settings. The foundational concepts of the n-width will be introduced and reviewed first.
Definition 1 ([
2,
3]).
Let W be a bounded subset of a normed linear space . Then, the Kolmogorov n-width, the linear n-width, and the Gel’fand n-width of W in X are respectively defined as follows:where ranges over all linear subspaces of X with a dimension of n at most, denotes all linear operators from X to X with a rank of n at most, and refers to all linear subspaces of X with a codimension no greater than n. Clearly, the above definition characterizes the best approximation errors for the most difficult elements in W when approximated by n-dimensional subspaces, linear operators with a rank n, and subspaces of a codimension n. If these errors fall within a given tolerance, then this implies that all elements of W are considered acceptable.
We now shift our focus to probabilistic and average case settings. In these scenarios, it becomes essential to establish certain measurability assumptions. Let W be equipped with a Borel algebra , consisting of the open subsets of W, and a probability measure v defined on . Specifically, v is a nonnegative, -additive measure on which satisfies .
Definition 2 ([
4,
5,
6]).
Let W, X, and v be defined as above, and let δ be a number in the range of . The probabilistic Kolmogorov -width, linear -width, and Gel’fand -width of W in X are respectively defined as follows:where represents subsets within such that . By comparing the definitions in the probabilistic setting with those in the worst case setting, it is easy to see that by removing a set in with a measure not exceeding , the errors caused by extreme elements could be significantly reduced in certain cases. Compared with the worst case setting, this approach better reflects the intrinsic structure of the classes. On the other hand, from the perspective of computational complexity, the probabilistic approach can, in some cases, greatly reduce the model complexity, thereby lowering the required cost.
Remark 1. The probabilistic Gel’fand n-width was first introduced in [4]. The definition in this paper indicates that the excluded measure set must satisfy additional measure conditions, which are detailed in the same paper. However, as mentioned in [7] (Remark 1.2), the proof in this article only requires some basic results for the probabilistic Gel’fand width. Consequently, the additional measurement conditions are similarly excluded in this context. Definition 3 ([
6,
7,
8]).
Let W, X, and v be defined as above. Then the p-average Kolmogorov n-width, p-average linear n-width, and p-average Gel’fand n-width are respectively defined as follows:where , ranges over all linear subspaces of X with dim , denotes any linear operator from X to X with a rank not exceeding n, and refers to all linear subspaces of X with a codimension no greater than n. The Kolmogorov width and linear width in the average case setting have been studied for over 20 years. In contrast, the Gel’fand width in the average case setting was only recently defined in [
7], and there is limited research on this topic. These concepts are used to characterize the ability to optimally approximate the majority of the elements in the function class under consideration using subspaces or linear operators. Obviously, this method better reflects the actual situation of the class.
The paper employs the following notation. For two nonnegative functions and , where , if there exists a positive constant independent of y such that holds, then this relationship is denoted by . Conversely, if there exists a positive constant independent of y satisfying , then this relationship is denoted by . If and , then this relationship is denoted by .
Many papers have studied the problem of n-widths in the worst case, probabilistic, and average case settings, yielding numerous remarkable results. Here, we review some of the known findings:
- (1)
Consider a classical univariate Sobolev space
equipped with a Gaussian measure
and satisfying the conditions
,
, and
, where
ensures a well-defined measure
. Maiorov in [
5,
6] proved that
Fang and Ye extended Maiorov’s results in [
8] and obtained
However, no results were available for Gel’fand widths until recently, when Tan et al. obtained the following result in [
4]:
Subsequently, Liu et al. obtained the following result in [
7]:
- (2)
Consider a multivariate Sobolev space
with mixed derivatives equipped with a measure
and satisfying the conditions
,
,
, and
. Similarly,
ensures that the Gaussian measure
is well defined. The authors in [
7,
9,
10] obtained the following for
:
where
represents
,
, and
,
represents
,
, and
, and
are defined as mentioned above. Additionally, for the case
, Equation (
1) still holds for the Kolmogorov width. However, as noted in [
7,
10], the linear width and Gel’fand width are somewhat different:
where
represents
and
,
represents
and
, and
are defined as mentioned above.
- (3)
The weighted Sobolev spaces
on the ball
and
on the sphere
have been extensively studied in the literature. Wang and Huang obtained the asymptotic order of the linear width and Kolmogorov width of the weighted Sobolev space
on the ball under the deterministic setting in [
11]. Huang and Wang obtained the asymptotic order of the linear width, the Gel’fand width, and the Kolmogorov width of the weighted Sobolev space
on the sphere under the deterministic setting in [
12]. Due to the length of the results, they are not listed here. Additionally, Wang obtained the linear width for the weighted Sobolev space
on the ball in the probabilistic and average case settings (see [
13]) for
and
, where
,
, and
:
and
where
v represents the Gaussian measure on
. Note that
is required to ensure that the Gaussian measure
v is well defined. However, the width of weighted Sobolev spaces on a sphere in probabilistic and average case settings has thus far been inconclusive.
Aside from the papers mentioned above, there are many excellent papers on this topic, such as [
3,
14,
15]. However, due to space limitations, they will not be listed here.
Inspired by these studies, this paper will investigate the Gel’fand and linear n-widths of the weighted Sobolev space on the sphere in the probabilistic and average case settings.
The structure of this paper is as follows. In
Section 2, we introduce the primary results of this article, along with essential notations and the Gaussian measure
v defined in the weighted Sobolev space, including its key properties.
Section 3 presents several lemmas which are fundamental to the proof of the main theorems.
Section 4 will present the discretization theorem, which primarily transforms the widths between different function classes, thereby simplifying the computation process. In
Section 5, we integrate the discretization theorems to provide the proof of the main theorems. In
Section 6, a summary of and outlook for this paper are provided.
2. Preliminaries and Main Results
Let
denote the unit sphere in
, endowed with the standard rotation-invariant measure
. Given
with
and
, let
denote the Jacobi weight on
and
The space
for
is defined as the set of measurable functions on
for which the following norm is finite:
For
, we replace the space
with
, which is the space of continuous functions on the sphere
endowed with the uniform norm. When
, the space
is endowed with the inner product
which clearly makes it a Hilbert space.
The Dunkl operators corresponding to the weight function given in Equation (
2) are expressed as follows:
where
. A spherical
h-harmonic
P of a degree
n on
is obtained by restricting a homogeneous polynomial
P of a degree
n to the sphere such that
. Here,
denotes the
h-Laplacian, defined as
, which plays a role analogous to that of the standard Laplacian. For further details, refer to [
16,
17]. The notation
denotes the space of all polynomials in
d variables with a degree of
n at most. Likewise,
represents the subspace consisting of
n-degree polynomials which are orthogonal to those of a lower degree in
. Thus, under classical Hilbert space theory, the following relationship is clearly valid:
Moreover, it is known that
Let
be a fixed orthonormal basis for
. Then,
is an orthonormal basis for
.
We use
to represent the orthogonal projection from
onto
and
to denote the orthogonal projection from
onto
; that is, for any
, we have
Remark 2. The operator is also known as the Fourier partial summation operator, and has the following representation:where is the reproducing kernel for . Moreover, the reproducing kernel has the explicit representation, which can be found in [12,18]. It is well known that spherical
h-harmonics are eigenfunctions of the operator
, meaning
is also referred to as the Laplace–Beltrami operator.
Given
, we define the fractional power
by
in a distributional sense. We define
to represent the derivative of order
r of the distribution
f.
The following introduces the spaces primarily studied in this paper. For
, the weighted Sobolev space
is defined by
with an inner product
It is obvious that this space is a Hilbert space.
This paper primarily studies the width problem of in , where . In order to ensure that this problem is well defined, we need to be continuously embedded in . According to Lemma 2, when , the embedding is guaranteed, which ensures that the definition is meaningful.
Next, we endow the space
with a Gaussian measure
v having zero mean, and its correlation operator
has eigenfunctions
where
, and
, and eigenvalues
Here,
is required to ensure that the trace of the correlation operator
is finite. In other words, this condition guarantees that the Gaussian measure assigned is well defined. In fact, we have
Therefore, when
, the series converges.
It is well known that
. Additionally, the Cameron–Martin space
associated with the Gaussian measure
v is
, as established in [
19]; in other words, we have
We then choose an any orthonormal system of functions
in
and let these
belong to
. Also, we let
be an any Borel subset of
. Then, the Gaussian measure
v of the cylindrical subset
is equal to
For further details on Gaussian measures in the context of Banach spaces, please refer to [
20,
21].
This article aims to investigate the approximation problem in weighted Sobolev spaces defined on . Now, we derive the following main results of this article.
Theorem 1. Let Then, we havewhere denotes or , , , and . By employing the techniques from [
7,
22] and utilizing Theorem 1, we arrived at the following result. The proof details are omitted.
Theorem 2. Let Then, we havewhere denotes or . Remark 3. Regarding the results for the Kolmogorov width, when , using a similar approach to the one for proving the linear width in this paper, the following result can be obtained: For the case where , the proof method for the upper bound presented in this paper does not apply. Therefore, this issue will be addressed in future work. We conjecture, however, that for , the Kolmogorov width result is as follows:where the definitions of s, r, ρ, and δ are as described above. 3. Main Lemmas
This section presents some useful lemmas which play a vital role in proving the main theorems of this paper. We first introduce a nonnegative function on , which is supported on , equal to one on , and positive on .
Let
where
. For functions
and
we write
Clearly, for
, we have
For all
the following properties hold (see [
23]):
- (1)
For each
, we obtain the following result:
- (2)
For each
, we obtain the following result:
In this paper, we still use the following classical metric on the sphere
:
Let
, where
,
. According to [
24], a weight function
w defined on the unit sphere
is referred to as an
weight if there exists a constant
such that for any spherical cap
and measurable set
, the following condition holds:
From [
24], it follows that
satisfies the above equation, and thus it is an
weight.
Lemma 1 ([
24]).
For , and , . Then, we havewhere . Lemma 2. Let Then, can be continuously embedded into the space .
Proof. Then, for any
, under Theorem 3.3 in [
23] and Equation (
7), we have
It can then be concluded through Lemma 1 that for
, we obtain
Therefore, for
, we have
This concludes the proof. □
Lemma 3 ([
2]).
Let X be a normed linear space and , , and W be subsets of X. If , then the following relations hold: Lemma 4 ([
2]).
Let Y a normed linear space and X be a Hilbert space. If there exists a continuous embedding of X into Y, then the relationholds true.
The following can be readily obtained by combining the proof of Theorem 1.5 in [
7] with Lemmas 3 and 4. The proof is omitted here.
Lemma 5. Let Y be a normed linear space and X be a Hilbert space. If there exists a continuous embedding of X into Y, where , and v is a probabilistic measure on H, then the following result holds: According to Lemma 5, to obtain Theorem 1, it is enough to consider the case of linear widths, which significantly reduces the amount of work required. The following will focus on the relevant lemmas concerning linear widths.
Consider a finite-dimensional
space
(
), which is the space
equipped with the
norm, defined as follows:
Next, we review the standard Gaussian measure
in
, which is defined by
where
G is an arbitrary Borel subset in
.
Let
,
, and
. Let
be a diagonal operator of the order
m, where
. The linear
-width of
D is defined as follows:
where
denotes any Borel subset of
and satisfies
and
represents linear operators from
to
having a rank
.
Lemma 6 ([
5,
8,
22]).
The following results hold:- (1)
Let , and . Then, the following result holds: - (2)
Let , , and . Then, the following result holds:
Lemma 7 Then, for we have 4. Discretization
This section focuses on deriving the discretization theorems which simplify the calculation of the probabilistic widths of a function class to that of a finite-dimensional set under the standard Gaussian measure.
For any
, we define
where
is given in Equation (
3). We denote by
the reproducing kernel of the Hilbert space
Then, we have
Moreover, if
, we also have
The following lemma is crucial for establishing the upper bound result in Theorem 1. Before presenting the lemma, we first introduce some related concepts.
A finite subset is defined as maximal -separable if it satisfies and
Lemma 8 ([
24,
25]).
Let w be a doubling weight on . There exists a constant , which depends solely on d and the doubling constant of w, such that for every and any maximal -separable subset , a sequence of positive numbers for exists such that for any , we haveand moreover, for we haveandwhere the constants of equivalence are determined solely by the doubling constant of w, as well as the parameters p and d. It is well known that weights are special doubling weights, and thus the weight satisfies the above lemma. Additionally, the above formulas are referred to as the cubature formula and the Marcinkiewicz–Zygmund inequalities, respectively.
Lemma 9 ([
12]).
Let For , the following inequality holds:where the definitions of Λ and can be found in Lemma 8. Let
be the constant as shown in Lemma 8 and denote
as a maximal
-separated subset of
where
According to [
12], it can be concluded that
. According to Lemma 8, for all
, the following cubature formula holds:
with
. Additionally, for any
, it holds that
Here,
is given as
and for any
, we have
Obviously,
is a linear operator. Now, we introduce the linear operator
as
Here,
and
is given as shown in Equation (
5). It is demonstrated in Equation (5.7) of [
12] that for any
one has
It follows directly from Equations (
6), (
10), and (
11) that
for any
.
One of the main theorems in this section is presented below, which is essential for estimating the upper bound in Theorem 1 and is referred to as the upper bound estimation discretization theorem.
Theorem 3. Let , along with sequences and satisfying the conditions and Then, we havewhere . The following lemma will be crucial in assisting with the proof of Theorem 3.
Lemma 10. For any , we havewhere is as previously defined and is defined by Equation (9). Proof. We know that
and thus, through the Riesz representation theorem, the expression in Equation (
10), and the Cauchy–Schwartz inequality, we obtain
This concludes the proof. □
We are now prepared to begin the proof of Theorem 3.
Proof of Theorem 3. First, we write
where
is the standard Gaussian measure in
. Let
be a linear operator from
to
with a rank
and
Recalling the definition of
(see Equation (
8)), it follows that
for
. Therefore,
.
We denote
where
We also have the following equation:
Here,
, and the second inequality follows from Equation (
12). For
, we have
where
and
is the
rth partial derivative of
with respect to the first variable. Since the random vector
is a centered Gaussian random vector with a covariance operator
, the random vector
, obtained through a linear transformation, remains a centered Gaussian random vector. Its covariance matrix is denoted by
. Note that
Hence, its covariance matrix
is expressed as follows:
It is quite straightforward to observe that for any
we have
and
The subset
of
is defined as follows:
where
and
are the positive constants given in Equations (
14) and (
15). From Equation (
14), we obtain
It is easy to see that the set
is both convex and symmetric for any
. Moreover, since Equation (
15) holds, according to the Gaussian measure comparison theorem (see [
19], p. 29), the following equation holds:
where
is a centered Gaussian measure in
with a covariance matrix
. We define
, and the linear operator
on
is expressed by
Finally, according to the definitions of
we obtain
This concludes the proof. □
We now turn to the lower bound estimation. Let
m be large enough and
, where
,
are constants which do not depend on
n or
m. Let
, with
, such that
We choose to be small enough to ensure that . Let be a smooth function defined on which is supported on the interval and takes the value of one on the subinterval . Additionally, let be a nonnegative smooth function on , which is supported on the interval and equals one on the smaller subinterval .
We then consider
where
is a constant chosen to satisfy
It is clear that
and
with
For
with
, we have
From [
12], it is known that for a positive integer
, there exists
Therefore, under the Kolmogorov-type inequality (see [
26]), we obtain
where
denotes the greatest integer
b such that
.
For
and
, we let
and
The operator
serves as an orthogonal projection from
onto
. For
, we have
. Additionally, the following conclusion holds:
This result can be derived as follows. The operator
is given by
where
. Using Equation (
16), we find that
where the last inequality holds because
when
.
With the Hölder’s inequality, we estimate
where
and
. Thus, the inequality in Equation (
18) is verified.
If
, then it is clear that
. Using Equation (
17), the following inequality holds:
Another main theorem in this section is presented below, which is vital for estimating the lower bound in Theorem 1 and is referred to as the lower bound estimation discretization theorem.
Theorem 4. Let , and N be defined as described previously, satisfying and . Then, we havewhere denotes the standard Gaussian distribution in . Proof. For convenience, we write
We denote
as a bounded linear operator acting on
, and the rank of
does not exceed
n such that
If
A is a bounded linear operator acting on
and
, then the image measure
of
v under
A remains a centered Gaussian measure on
, and its covariance takes the following form:
where
is the adjoint operator in
,
is the covariance of
v, and
is given in Equation (
4) (see [
19], Theorem 3.5.1). Additionally, if
A satisfies
then
According to [
19] (Theorem 3.3.6), if for any Borel set
it is absolutely convex, then the following inequality holds:
From Equation (
19), we have
Thus, the constant
exists such that
Next, the set
is absolutely convex for
, and therefore
Let
and
be defined as follows:
It is clear that if .
For
, we have
Using Equation (
16) and the fact that
for
, we obtain
Therefore, for any
, we deduce
Since
forms an orthonormal system in
and
, the random vector
follows a standard Gaussian distribution
in
. Hence, we obtain
where
is a constant and
> 0. The rank of the operator
does not exceed
n, and
Thus, we have the inequality
which leads to
This concludes the proof. □
5. Proof Main Results
The main focus of this section is to demonstrate the proof of the main theorem using some lemmas and the relevant theorems obtained in the previous section.
Proof of Theorem 1. It suffices to prove , while follows directly from Lemma 5.
To begin with the lower bound estimation, by using Theorem 4 and Lemma 6, we can obtain the result for
:
The upper bound of the theorem remains to be estimated, and we now begin its proof.
For
and a fixed natural number
n, assume that
, where
will be determined later. We can choose a sufficiently small
such that
, where
is given in Lemma 9, and define
where the definition of
can be found in Theorem 3. Then, we have
It is evident that by selecting a large enough
value, we can ensure that
. Under Lemma 9, we know that for
,
, we have
For
,
, and hence
For
, by taking
and using Lemma 7, we find that for
, we have
and for
, we have
We will now proceed to estimate the upper bounds for
. For
, according to Equations (
13) and (
20), and noting that
, we obtain
Hence, for
, we have
For
, it follows from Equations (
13) and (
21), and noting that
, that
Hence, for
, we obtain
For
, by the definition of
, it is evident that
This concludes the proof. □
6. Conclusions
Sobolev spaces play a crucial role in both theoretical research and practical applications, while weighted Sobolev classes provide theoretical guidance for addressing issues arising from non-uniform data points. This paper investigated the width problem in the weighted Sobolev space on the sphere under the average case and probabilistic settings, deriving the exact orders of the Gelfand and linear n-widths in both settings. These results fundamentally reveal the computational complexity and provide an effective framework for quantifying algorithm performance, offering a solid theoretical foundation for designing efficient algorithms. Meanwhile, certain aspects of the Kolmogorov width remain unresolved, and future research will focus on these open problems to further advance the study of the classical width for weighted Sobolev classes.