The fundamental concept of compressed sensing theory is the integration of signal acquisition and compression, thereby avoiding the conventional approach in signal processing where the signal is first sampled and then compressed for transmission. The mathematical model of compressed sensing theory can be expressed as follows [
12]:
where
P is the observation matrix of
,
,
D is the sparse dictionary (sparse basis),
is the compressible original signal of length
N,
is the observed signal of
M dimensions, and
is the coefficient vector under the sparse basis. In reference [
13], it is assumed that the sparse basis
D is an orthogonal matrix, which has led many works related to compressed sensing to define
D as an orthogonal matrix. In reality, in reference [
14], M. Elad and D. L. Donoho, among others, have extended the sparse basis to non-orthogonal cases, defining this matrix as a dictionary.
Compared to the Nyquist–Shannon sampling theorem, the sparse dictionary is learned and provides a more accurate representation of the function being approximated relative to the Shannon function. For understanding the sparse dictionary, we can explain it through the projection of points, lines, and planes, for example, the following:
3.1. The Sparse Representation and Reconstruction Model of Points
In the context of this manuscript, the “method of changing plane” refers to a core technique in descriptive geometry for simplifying the projection analysis of spatial geometric elements. Its fundamental principle is that when a spatial figure (such as a line, plane, or solid) is in a general position (i.e., not parallel, perpendicular, or coinciding with the original projection planes), a new projection plane is introduced to replace one of the original projection planes (e.g., replacing the horizontal projection plane H or the vertical projection plane V). This new projection plane is positioned to be parallel or perpendicular to the key features of the target figure, thereby converting the general position of the figure into a special position (e.g., a line perpendicular to the new projection plane or a plane parallel to the new projection plane). Such a transformation retains the essential geometric properties of the figure (e.g., true length of lines, true shape of planes) while simplifying its projection expression, which aligns with the sparse representation logic of “retaining key information and reducing redundancy” discussed in this paper.
This technique is a classic approach in descriptive geometry, and its specific operational rules and mathematical properties can be further referenced in relevant theoretical frameworks and application studies of projective geometry.
As shown in
Figure 4, in 3D Euclidean space, let the coordinates of point
A be
. During the first plane change, the new projection axis
passes through the horizontal projection
a of point
A. In the new projection plane system
, point
A actually lies on the projection plane
. In the second plane change, the new projection axis
passes through
, and in the newly formed projection plane system
, point
A lies on the projection axis
, with its coordinates expressed as
.
The new projection plane system
can have infinite variations. The simplest form is the projection plane system obtained by translation, where the axis
passes through point
A, such that the three-projection plane system can still be represented in the form of a unit matrix
.
In the theory of compressed sensing, this matrix is referred to as a sparse matrix or a sensing matrix. Professor David L. Donoho from Stanford University suggests that this matrix does not need to be an orthogonal matrix [
14], which simplifies our subsequent analysis.
Theorem 1. A measurement matrix corresponding to a point in 3D Euclidean space allows for the precise reconstruction of the original signal from the measurement signal.
Proof. Without loss of generality, let the measurement matrix (projection matrix) be denoted as . According to the theory of compressed sensing, as long as the measurement matrix and the vector formed by the three coordinate values of a known point A are not orthogonal, we have the following:
(1) The measurement signal:
(2)
, where
is a sparse matrix;
x is the original signal, corresponding to a point in 3D Euclidean space; and
is the sparse coefficient. Therefore, we have
because
It is evident that by setting and , Equation (6) can be satisfied, and , thereby meeting the sparsity requirement. Next, we use the sparse coefficients obtained in this way to reconstruct the original signal , which represents the coordinates of point A.
According to the original signal reconstruction steps in compressed sensing theory [
13], we have
It is evident that an exact reconstruction of the original signal x has been achieved. Hence, the proof is complete. □
3.2. The Sparsity and Reconstruction Representation of a Straight Line
According to the descriptive geometry theory, when the new projection axis is parallel to a straight line, in the new projection system, this straight line becomes a projection plane parallel line. This is because the parallelism between the projection axis and the straight line ensures that the straight line maintains a fixed positional relationship with the corresponding projection plane in the new system, thus showing the characteristics of a projection plane parallel line. Its projection onto the projection plane parallel to it will reflect the actual length of the straight line, which is a key feature of projection plane parallel lines in descriptive geometry. In addition, for the situation where the examined plane is first made perpendicular to the new projection plane and then parallel to the latest projection plane, it is essentially a step-by-step transformation using the plane-change method. Making the plane perpendicular to the new projection plane first simplifies the projection of the plane, and then further changing the projection plane to make it parallel to the latest projection plane allows us to obtain the true shape of the plane, which is crucial for accurately analyzing and solving geometric problems involving planes. As for why a straight line is used to position the plane into a special relation with a new projection plane, it is because a plane can be uniquely determined by a straight line and a point not on the line, by two intersecting straight lines, etc. By controlling the positional relationship between the straight line and the new projection plane, we can indirectly control the positional relationship between the plane and the new projection plane. This method simplifies the positioning process and is in line with the basic principles and research methods of descriptive geometry.
It should be noted that the signal being processed is a straight line, and a line segment can be determined by two points, which already confirms the sparsity of the signal. In descriptive geometry, the plane-change method is essentially rotations around coordinate axes, which adjust the orientation of projection planes to align with the spatial line, and to further achieve the sparse representation of a straight line, one can use two plane transformations to convert a general straight line into a line perpendicular to the projection plane, and then by translating the coordinate system so that one of its axes coincides with the line, the maximum sparse representation of the line can be realized. As shown in
Figure 5, the straight line segment MN composed of points
and
is a general position straight line, which, after two plane transformations, becomes a line perpendicular to the projection plane in the new projection plane system.
For convenience of description, let
,
, and
. Assume that after two changes in the projection plane, in the new projection plane system
, the projection axis
coincides with the line segment
, which can be represented in vector form as
; in the original projection plane system
, the corresponding line can be expressed in vector form as
. Therefore, the sparse matrix
can be defined as
According to compressed sensing theory, the original line signal can be represented using a sparse basis as
By direct observation, the solution for the sparse coefficient can be obtained as follows: , . We shall not delve into the methods for solving the sparse coefficients at this juncture, but it is important to note that this represents the sparsest solution for the sparse coefficients, .
Next, we present the compressed sensing theorem concerning a line segment in the form of a theorem [
17].
Theorem 2. The measurement matrix corresponding to the line segment formed by the two points and in three-dimensional Euclidean space allows for the accurate reconstruction of the original signal from the measured data.
Proof. Without loss of generality, let the measurement matrix (projection matrix) be
According to the compressed sensing theory, we have the following:
(1) The measurement signal:
(2)
, where
is a sparse matrix;
x is the original signal, corresponding to two points in 3D Euclidean space; and
is the sparse coefficient. Therefore,
because
where
It is not difficult to see from Equation (11) that by choosing , , Equation (9) can be satisfied, and , which meets the sparsification requirement. Using Equation (12), the sparse coefficient can be obtained to reconstruct the original signal . This completes the proof of the theorem. □
It is easy to understand that in the sensing matrix , the first and fourth column vectors are the most crucial, while the second, third, fifth, and sixth column vectors only need to be uncorrelated with the first and fourth column vectors. For this reason, the observation matrix only needs to retain the information from the first and fourth column vectors, so an observation matrix with two rows and six columns will suffice to meet the requirements.
For a line segment, another more general vector expression method involves transforming the line segment into a line parallel to the coordinate axes, using a linear combination of the vector formed by the endpoint coordinates and the coordinate axis vectors. As shown in
Figure 6, the length of the line segment
is
, and the red unit vector
can be expressed as
, where
,
,
. Its direction is parallel to the line segment
.
In this method of representation, the local coordinate axes that make up the perception matrix are parallel to, but do not coincide with, the direction of the known line segment. A straight line segment can be expressed as the vector
, and the corresponding perception matrix is represented as
The sparse expression of the line segment
is given by
, and its corresponding matrix expression is
According to compressed sensing theory, its sparsest solution can be expressed by the following optimization formula:
It can be directly observed that
,
,
,
, and
. At first glance, the sparse coefficient vector
does not appear sufficiently sparse. However, the matrix possesses a certain degree of generality, enabling it to represent all lines parallel to line
, such as the magenta parallel lines shown in
Figure 6. Furthermore, when additional points are added on line
, the number of non-zero elements in the sparse coefficient vector
does not need to increase. For instance, when point C is added on the extension of line AB, making points A, B, and C evenly distributed, the signal containing the three points is still represented.
The sparse expression is
, and the corresponding matrix form is
Direct observation reveals that
Similar to the analysis of Theorem 2, the key to the aforementioned sensing matrix lies in the first four column vectors. As long as the observation matrix can capture the first four column vectors of the sensing matrix, it guarantees the accurate reconstruction of the original signal x from the observed signals.
In descriptive geometry, the core purpose of the method of changing planes is to transform geometric elements in general position (such as straight lines and planes) into special position lines in a new coordinate system. This process is analogous to the dictionary learning process in sparse representation theory: both achieve the sparsification of geometric element expression by leveraging the intrinsic structural characteristics of the target.
Specifically, the method of changing planes adjusts the projection orientation through successive transformations, reducing redundant coordinate information while retaining essential features (e.g., true length of lines or true shape of planes). Similarly, dictionary learning in sparse representation constructs an optimal basis that aligns with the signal’s intrinsic structure, minimizing non-zero coefficients in the expression. Their shared essence lies in extracting key features to simplify expression, thus establishing a natural connection between classical descriptive geometry practices and modern sparse representation theory.
3.3. Sparsity of Planar Polygons
In three-dimensional Euclidean space, a point requires three coordinate values to determine its position; a line segment requires two endpoints, which amounts to six coordinate values to define its location. However, when a point in space lies on one of the coordinate axes, only one of its coordinate values is non-zero; similarly, a line segment would have only two non-zero coordinate values. For instance, point A on the X-axis has the coordinates ; the line segment on the X-axis has the coordinates .
For a planar polygon formed by
k characteristic points, the method of plane projection can be employed so that the plane coincides with a projection plane, such as the horizontal projection plane. In this case, the
z-coordinates of these
k characteristic points are zero, and the vector composed of the coordinate values of these
k points is
-sparse. As shown in
Figure 7, the four characteristic points
of a quadrilateral lie on the same plane, forming a vector with twelve elements, of which four elements are zero.
In the theory of the sparse representation of lines, only one dimension of the three-dimensional coordinate system is utilized, and the other two coordinate axes are effectively redundant. For simplicity, we will directly use planar figures as examples to illustrate the principle of compressed sensing for planar figures with certain geometric structures. As shown in
Figure 7, a seemingly complex planar polygon has edges that actually follow only two directions. In the local coordinate system formed by the two red-marked vectors, all contour edges are either horizontal or vertical lines. Compressed sensing theory exploits this particular geometric structure.
As illustrated in
Figure 8, the principal components are immediately apparent; however, the structure is not symmetrical. This dataset pertains to planar data, but the principal components should encompass three directions, which can be further extended to additional directions. The model can be described as follows:
where
Clearly, this is a problem of sparse representation and compressed sensing.
3.4. Sparsity of Solids
Solids of revolution are defined as three-dimensional figures formed by rotating a planar curve (generatrix) around a fixed straight line (axis of rotation) in the same plane, with the axis of rotation remaining unchanged during the rotation process. Basic solids of revolution (cylinders, cones, spheres) and plane solids with symmetrical properties (prisms and pyramids), despite having numerous lines in their diagrams, can achieve compressed sensing through their axis of symmetry and axis of rotation. It can be demonstrated that the eigenvectors of the covariance matrix, formed by the characteristic points of a solid, correspond to the axis of symmetry or the axis of rotation of the solid. This is the primary reason for performing singular value decomposition on data matrices in sparse representation, compressed sensing, and deep learning [
18]. For symmetric geometric bodies involved in descriptive geometry and engineering drawing, the axis of rotation or symmetry can be directly obtained from the eigenvectors of the data matrix.
Store the given m pieces of n-dimensional data in the matrix D, then perform zero-mean normalization on each row of the matrix D. Next, construct the covariance matrix of the dataset and compute its eigenvalues and eigenvectors. Arrange the eigenvectors in a matrix, sorted by their corresponding eigenvalues from largest to smallest. Select the first k rows to form the matrix P. By multiplying the matrix P with the original dataset matrix, the dimensionally reduced data can be obtained.
If we take the characteristic points that define a line or a plane as a dataset, can we use principal component analysis (PCA) to analyze the principal components of this dataset? Without loss of generality, let us consider the rectangular cuboid shown in
Figure 9 as an example. The axis of the rectangular cuboid (depicted by the blue dashed line) is in a general position with respect to the coordinate system shown. Based on the symmetry of the rectangular cuboid’s vertices relative to the midpoint
O, it can be deduced that
According to the principles of principal component analysis, the set of points formed by the eight vertices of the rectangular cuboid can be represented in the following matrix form:
The covariance matrix corresponding to matrix
D is
Matrix
C’s eigenvalues can be determined as follows:
To simplify calculations, we employ the method of changing plane to transform the central axis of the rectangular cuboid (indicated by the blue dashed line) into the vertical line of the projection plane, as shown in
Figure 10. Clearly, directly computing the corresponding eigenvalues and eigenvectors is rather challenging.
Without loss of generality, the new projective plane system (purple coordinate axes
) shown in
Figure 10 can be easily obtained by selecting the axonometric axes and adopting the method of “changing the plane.” The coordinate values of the eight vertices under the new coordinate system are shown in
Figure 10. Let the length, width, and height of this rectangular cuboid be 2a, 2b, and 2c, respectively, and then we have
In the new coordinate system, the set of points formed by the eight vertices of the rectangular cuboid can be represented in the following matrix form:
The covariance matrix corresponding to matrix
D’ is
The eigenvalues of a matrix
can be determined by the following method:
Hence, it is not difficult to deduce that
Next, we seek the corresponding eigenvectors.
(1)
According to the definition of matrix eigenvectors, we have
Substituting Equation (26) into Equation (28), we obtain
Solving equation system (29), we obtain the eigenvalues and their corresponding eigenvectors as follows: the eigenvector corresponding to eigenvalue
is
; similarly, we have
for
and
for
. These three eigenvectors have clear geometric interpretations, as they correspond to the three purple coordinate axes in
Figure 10.
Next, we will use the method of changing plane to find the eigenvalues and eigenvectors in the original coordinate system. According to the definition of eigenvalues, for the same point set, the eigenvalues remain unchanged in different coordinate systems.
As shown in
Figure 11, the vector
represents the characteristic vector of the vertex data matrix of the rectangular cuboid shown in
Figure 10. Without loss of generality, in the new coordinate system (formed by the characteristic vectors of vertices of the symmetrical geometric object), the coordinates of points
P and
Q are (−L/2,
,
) and (L/2,
,
), respectively. Hence, the coordinates of
, as shown in
Figure 11, are (−L/2,
), and the coordinates of
are (L/2,
). Below, we will transform projections
and
of the line segment MN in the new projection plane system back to the original projection plane system [
19].
Step 1: Change the H-plane. Rotate by an angle
around the
Y-axis:
The first step can be represented by the corresponding rotation transformation matrix as follows:
Step 2: Change the V-plane. Rotate by an angle
around the
Z-axis:
where
, and the corresponding rotation transformation matrix can be represented as
Step three involves transforming the line segment PQ’s projections,
and
, from the new projection plane system back to the original projection plane system.
The result indicates that, regardless of how a rectangular cuboid is rotated, the vectors corresponding to its edges are the eigenvectors of the matrix obtained by multiplying its vertex data matrix with its transposed matrix. Therefore, the geometric interpretation of multiplying the vertex data matrix with its transposed matrix can be explained using the method of changing plane.
Hence, it can be observed that for basic three-dimensional solids with a symmetric structure, the choice of axonometric axis and the results of principal component analysis are consistent
For asymmetrical structures, the corresponding eigenvector (principal component) directions are relatively complex. A combination of clustering and singular value decomposition (such as K-SVD) methods may help to identify sparse vectors that can represent the characteristics of the shape.