1. Introduction
Suppose
and
are real (or complex) Hilbert spaces and
D is a nonempty subset of
E. Given a constant
, a function
is said to be an
ε-isometry if and only if
f satisfies the following inequality:
for all
, where
is the norm generated by the inner product
, i.e.,
.
Hyers and Ulam [
1] proved the stability of the surjective isometry defined on the ‘whole’ space by using properties of the inner product of Hilbert space:
For every surjective ε-isometry that satisfies , there exists a surjective isometry satisfying
for all .
In 1978, Gruber [
2] proved that if the Hyers–Ulam theorem holds for all surjective
-isometries
, where
E and
F are real Banach spaces, then
can be replaced by
in inequality (
2) (see also Gevirtz [
3]). Finally, Omladič and Šemrl [
4] succeeded in changing
to
in inequality (
2) and showing that the resulting upper bound is sharp.
For the case when the domain of the relevant
-isometries is bounded, Fickett [
5] answered the question as to whether there exists a true isometry that approximates the
-isometry defined on a bounded set. Now we introduce Fickett’s theorem:
Assume that is an integer, D is a bounded nonempty subset of , and is any constant. If is an ε-isometry, then there is an isometry that satisfies
for all .
Comparing (
2) and (
3), we see that as
approaches 0, the convergence rate on the bounded set of high-dimensional spaces becomes slower than the convergence rate on the (whole) Hilbert space. Vestfrid [
6] proved that for every
-isometry
, there is an isometry
such that
for all
, where
. Very recently, Choi and Jung were able to improve the upper bound of inequality (
3) to
, where we set
(see [
7], Theorem 3):
Given an integer , let D be a bounded subset of the n-dimensional Euclidean space such that for some . If is an ε-isometry with , where ε is some real number with , then there exists a linear isometry such that
for all .
So, as
, the
of the results in [
6,
7] converges by
, while the
of our result converges by
. As we can see, the rate with which the upper bound of the relevant inequality varies with the dimension
n of the space is as important as the speed with which the upper bound converges to 0 as
approaches 0.
In this paper, we focus on finding a sufficient condition to obtain a convergence rate proportional to
by applying singular value decomposition. Indeed, we prove the stability of isometry defined on the closed ball
, where
, under an additional condition of approximate linearity with the constant
K (whose definition is given in
Section 4). Throughout this paper,
denotes the
n-dimensional Euclidean space, where
n is a fixed positive integer.
2. Change of Basis
Assume that
is a basis for the
n-dimensional Euclidean space
and
, where every
and
x are written as a column vector. The
-
coordinates of
x are the weights
such that
where
are uniquely determined real numbers that depend only on the choice of
. We use the symbol
to denote the
-coordinates of
x. More precisely, if
x is represented by (
4), we have
where
denotes the transpose of the row vector
, i.e.,
is a column vector. For simplicity, we write
x instead of
, where
denotes the standard basis for
.
We now denote by
the
n-dimensional real vector space
in the
-coordinate system. That is,
Let us define the
matrix
by
Then the vector equation, Equation (
4), is equivalent to
The square matrix
is called the
change-of-coordinates matrix from
to the standard basis
. Since the columns of
form a basis for
,
is invertible. Thus, Equation (
6) is equivalent to
where
denotes the inverse of
. In other words,
converts the
-coordinates of
x into its
-coordinates.
3. Singular Value Decomposition
A square matrix is said to be diagonalizable if it is similar to a diagonal matrix, that is, if there is an invertible matrix and a diagonal matrix such that . However, there are matrices that cannot be diagonalized, but fortunately the factorization is possible for any matrix , where and are some suitable invertible matrices. A special factorization of this kind, the so-called singular value decomposition, is one of the most useful matrix factorizations in applied linear algebra.
Suppose
is an arbitrary
matrix whose entries are all real numbers, i.e.,
. We use the symbol
to denote the transpose of the matrix
. We note that
is symmetric and can therefore be orthogonally diagonalized. Let
be an orthonormal basis for
consisting of eigenvectors of
, and let
be the corresponding eigenvalues of
. Then we have
for every
, where
denotes the Euclidean norm on
. Therefore, all eigenvalues of
are nonnegative. By renumbering if necessary, we can assume that the eigenvalues are arranged in decreasing order:
The
singular values of
are the square roots of the eigenvalues of
, denoted by
, and they are arranged in decreasing order. That is,
where
for every
.
According to the singular value decomposition, for every
, there is an
orthogonal matrix
, a diagonal matrix
whose diagonal entries are the singular values of
, namely
, and an
orthogonal matrix
such that
where
and
are column vectors with
n components each, and the column vectors
of
form an orthonormal basis for
consisting of eigenvectors of
.
The SVD method is a way of carrying out diagonalization, and the reason we use this method is to dramatically reduce the number of variables we have to consider, so we can obtain good estimates.
4. Stability of Isometries on a Bounded Domain
As before, let
be the standard basis for the
n-dimensional Euclidean space
, where
. Let
represent a closed ball in
with radius
and its center at the origin of
, i.e.,
where
is the Euclidean norm generated by the inner product
with
for
and
, i.e.,
.
In this section, let
be an
-isometry that satisfies
. We define the
matrix
by
where every
is written as a column vector. We note that
for every
.
According to the singular value decomposition or (
10), there is an
orthogonal matrix
, a diagonal matrix
whose diagonal entries are the singular values of
, namely
, and an
orthogonal matrix
such that
. Note that the sets of column vectors
and
form orthonormal bases for
, respectively.
Gevirtz showed in [
3] that every surjective
-isometry
behaves asymptotically like a linear surjective isometry. When
D is a proper subset of a real Banach space
E, this result does not imply that each
-isometry
behaves asymptotically as a linear isometry. Nevertheless, this Gevirtz result ensures the reasonableness of the following definition.
Definition 1.
A function will be called approximately linear with constant K if and only if there is a constant such thatfor all , where ε and are given in and , respectively. In particular, a linear function is approximately linear with the constant . In the following two examples, we will construct functions that are approximately linear and functions that are not.
Example 1.
We define a function byfor all . Sincewe have Moreover, we obtainfor all . Therefore, we see that f is approximately linear with the constant . Example 2.
As a counter example, we let . And we define a function byfor all , where the closed unit ball is defined by . Then we have Furthermore, we apply the Lagrange multiplier method to obtainfor all . Therefore, there exists an that does not satisfy the condition for the approximate linearity of f with the constant . Theorem 1.
Assume that n is a positive integer, d is a real number that is not less than 1, and K is a nonnegative real number. Suppose is a function that satisfies the following three conditions:
;
f is approximately linear with the constant K;
f is an ε-isometry; more precisely, it satisfies inequality for all and for some constant ε with .
Then there exists an isometry such thatfor all . Proof. Let
be the change-of-coordinates matrix from the basis (coordinate system)
to the standard basis
. Similarly, let
be the change-of-coordinates matrix from
to
. On account of (
5), (
6), (
7), (
10), and (
12), it holds that
,
, and
for any
. Let
be the set of
-coordinates of all
, i.e.,
From a set-theoretic perspective alone, we can see that
. For simplicity, we write
instead of
.
Considering (
13), we define the linear function
by
for any
. Therefore, we obtain
i.e., the
-coordinates of
is
for all
. (From now on, all vectors will be expressed as row vectors for convenience.) Also comparing (
13) and (
14), we have
for any
.
Since the orthogonal matrix
preserves the Euclidean norm of each vector in
, it follows from (
7),
, and (
14) that
for all
and for the constant
given by
.
For any
with
, let
, where
and
denote the
-coordinates of
x and the
-coordinates of
, respectively. Since
and
and
are orthogonal matrices (since
and
and
are orthogonal), it is obvious that
Thus, from (
1),
, and (
17), we have
for all
and
. Moreover, since
has
-coordinates, from the inverse triangle inequality and (
16), we obtain
Since
from (
15) and
from (
17), it follows from (
19) that
for all
.
Hence, using (
17) and the first inequality of (
18), we have
and from (
16) and the (inverse) triangle inequality, we note that
i.e.,
for every
. Furthermore, using (
15), the second inequality of (
18), and (
22), we obtain
for all
.
It now follows from (
21) that
since
,
, and
by
. Similarly, it follows from (
23) that
for all
, since it holds from (
15) and (
22) that
and
We use (
25) to obtain
for any
. Therefore, it follows from (
20), (
24), and (
26) that
for all
.
We note that
. Since
, we can use (
20) and (
27) to obtain
for
.
We now define an isometry
by
for all points
with
, i.e.,
for all
with
. It then follows from (
28) that
for all
with
. We recall that
for all
. Let
be the transformation matrix for the linear operator
I. Then we have
for all
, where
is the (orthogonal) transformation matrix for some isometry
. □
We note that a function is linear if and only if it is approximately linear with constant . Therefore, if we assume that f is linear, we can easily verify that the following corollary is an immediate consequence of Theorem 1.
Corollary 1.
Assume that n is a positive integer, d is a real number that is not less than 1, and ε is a constant satisfying . If is a linear ε-isometry, then there exists an isometry such thatfor all . Example 3.
Given , we define a function byfor all . Then we haveHence, it follows thatfor all , i.e., f is linear (and therefore . In addition, we obtainfor all , where we set . That is, it holds thatfor all , which implies that f is an ε-isometry. According to Corollary 1, there exists an isometry such thatfor all . 5. Conclusions and Discussion
One of the characteristics of this paper is that it is the first to use the technique of singular value decomposition to study the stability of isometries. Although we additionally assumed approximate linearity, we were able to obtain a better result than previous results by other mathematicians by utilizing the SVD technique, as can be seen in
Table 1 below.
We compare the result of this paper with previously published notable research results and present them in
Table 1. More precisely, we compare the coefficients
of the upper bounds
for the difference between the given function
f and the sought isometry
I.
The values in the first row of
Table 1 were obtained by substituting
into the formula presented in the proof of [
8] (Theorem 4.1). The values in the second row of
Table 1 result from the assumption that
and
in [
6] (Theorem 1). The values of the third and fourth rows are due to the formulas presented in [
9] and [
7] with
, respectively. The last row gives approximate values obtained from Theorem 1 in this paper, assuming
and
. According to
Table 1, the results of this paper are superior to those of [
6] for all integers
.
In an extension of this paper, we plan to continue to study the following topics:
Whether similar results can be achieved with norms other than the Euclidean norm will be investigated in more detail.
We will investigate whether similar results to those in this paper can be obtained under conditions other than approximate linearity.
We will examine the relationship between the concept of -isometry and the concept of approximate linearity.
We will study whether the results of this paper can be extended to infinite-dimensional real Hilbert spaces.