Next Article in Journal
Investigation of Trajectory Tracking Control in Hip Joints of Lower-Limb Exoskeletons Using SSA-Fuzzy PID Optimization
Previous Article in Journal
HCT-Det: A High-Accuracy End-to-End Model for Steel Defect Detection Based on Hierarchical CNN–Transformer Features
Previous Article in Special Issue
A Novel Multi-Sensor Nonlinear Tightly-Coupled Framework for Composite Robot Localization and Mapping
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Multiple CR Spatiotemporal Compressive Imaging System

1
School of Optics and Photonics, Beijing Institute of Technology, Beijing 100081, China
2
Key Laboratory of Photo-Electronic Imaging Technology and System, Ministry of Education of China, Beijing 100081, China
3
National Key Laboratory on Near-Surface Detection, Beijing 100072, China
*
Author to whom correspondence should be addressed.
These authors contributed equally to this work.
Sensors 2025, 25(5), 1334; https://doi.org/10.3390/s25051334
Submission received: 20 January 2025 / Revised: 15 February 2025 / Accepted: 19 February 2025 / Published: 21 February 2025
(This article belongs to the Special Issue New Trends in Optical Imaging and Sensing Technologies)

Abstract

:
Higher spatial and temporal resolutions are two important performance parameters in an imaging system. However, due to hardware limitations, the two resolutions are usually mutually restricted. To meet this challenge, we propose a spatiotemporal compressive imaging (STCI) system to reconstruct high-spatiotemporal-resolution images from low-resolution measurements. For STCI, we also designed a novel reconstruction network for multiple compression ratio (CR). To verify the effectiveness of our method, we implemented simulation and optical experiments, respectively. The experiment results show that our method can effectively reconstruct high-spatiotemporal-resolution target scenes for nine different CRs. With the maximum spatiotemporal CR of 128:1, our method can achieve a reconstruction accuracy of 28.28 dB.

1. Introduction

High-spatiotemporal-resolution imaging finds widespread application in diverse fields, including medicine, security, and commercial settings. However, simultaneously improving both the temporal and spatial resolution of imaging systems presents a significant challenge. High-speed, high-resolution imaging puts forward higher requirements on the hardware performance of the systems. The massive data associated with high-speed, high-resolution image signals, coupled with continuous high-frame-rate acquisition, impose stringent requirements on the data transmission bandwidth of the system and the performance of back-end storage devices [1]. Consequently, in the development of imaging systems, spatial and temporal resolutions always exhibit mutual constraints [2]. Although some high-speed cameras can simultaneously achieve high spatial and temporal resolutions, they frequently encounter limitations in data storage capacity, hindering prolonged operation. Furthermore, the high performance of such detectors typically comes at a considerable cost. Framing cameras, which are capable of extremely high-speed imaging, are essential for capturing transient phenomena. However, their complex structure, high cost, and limited sequence depth have hindered widespread adoption. Compressive imaging (CI) is a computational imaging method based on compressive sensing theory [3]. CI can realize high-resolution imaging that exceeds the limitation of the imaging system by combining back-end reconstruction algorithms [4,5,6]. Based on the spatiotemporal correlations inherent in natural image sequences [7], STCI aims to achieve high-speed and high-resolution imaging of target scenes using CI techniques.It encodes and compresses the target scene in both spatial and temporal dimensions and reconstructs the high-resolution images through the reconstruction algorithm, which effectively alleviates hardware limitations. STCI significantly reduces the amount of data acquired, which alleviates the strain on both the data transmission bandwidth of the system and back-end storage devices, thereby enabling prolonged high-speed, high-resolution imaging. Before our work, there have been some initial studies on STCI. In [8], Treeaporn et al. demonstrated the feasibility and advantages of STCI under high-readout-noise conditions through simulation experiments. The paper also discusses three electro-optical space-time compressive imaging architectures. However, it is difficult to apply to actual imaging systems because of the high complexity of the system. In [9], Harmany et al. combined the coded aperture and the keyed exposure technique to improve spatiotemporal resolution and analyzed the mathematical model. However, the encoding matrix designed using the coded aperture is complex and computationally intensive. In addition, a feasible optical structure is not given for the imaging idea. Ke et al. [10] combined block-wise compressive imaging (BCI) [11] and temporal compressive imaging (TCI) techniques for high-spatiotemporal-resolution imaging. A two-step reconstruction strategy is adopted for reconstruction. However, the quality of the reconstructions is limited. In [12], Tsai et al. discussed CI in the spectral and temporal domains, which presents excellent performance. With advances in light field modulation technology, Zhang et al. proposed end-to-end snapshot compressive super-resolution imaging with deep optics, which adopted the diffraction optical element (DoE) and the TCI technique to realize the improvement of spatial and temporal resolutions [13]. The researchers also proposed an End-to-End network to optimize the DoE and the reconstruction network jointly. The system achieved 128 times data expansion in spatial and temporal domains. The introduction of the DoE element realizes the spatial encoding for target scenes but also increases the complexity of the system.
Reconstruction algorithms play an important role in CI. In previous research, CI reconstruction algorithms have been well studied, consisting mainly of the traditional model-based reconstruction algorithm [14,15,16] and the deep learning (DL)-based reconstruction algorithm [17,18]. For STCI, it encodes and compresses the original data in both temporal and spatial dimensions. Successive image signals exhibit strong spatiotemporal correlations. Therefore, the utilization of spatiotemporal correlations is of significant importance for the design of the STCI algorithm. Furthermore, we found that high-speed motion scenes exhibit more pronounced sparsity in the temporal domain than in the spatial domain. Based on this observation, we discuss a new STCI model and design a novel reconstruction algorithm based on the unfolding network from the point of view of encoding strategy design. In addition, we notice that most reconstruction algorithms are designed for fixed CR. However, for different imaging tasks, such as object detection and recognition, the requirements for imaging quality or resolutions vary depending on the scene and task requirements. Adaptively adjusting the CR based on the imaging task and scene variations can avoid redundant data acquisition. In such cases, a reconstruction algorithm that can handle measurement data with varying CRs is very valuable. Based on the above discussion, we believe that a simple and easy-to-implement STCI system and a reconstruction algorithm that can handle different CRs are of great significance for the development of STCI. The specific contributions of this work are as follows.
  • Considering the need to balance the encoding efficiency and the simplicity of the imaging system structure, we re-discuss the mathematical model of STCI and adopt a stepwise strategy for spatiotemporal encoding. Through the preprocessing of the mask, we finally only need to introduce a single spatial light modulator (SLM) in the optical system to achieve effective spatiotemporal encoding.
  • By adopting a special structural spatial mask and introducing a hyperparameter module for multi-CR reconstruction in the network, we implemented an STCI reconstruction method that can handle reconstruction with multiple spatiotemporal CRs.
  • Through simulation and optical experiments, we verify that our proposed method can accurately reconstruct the motion and spatial details of high-speed motion scenes when the temporal and spatial compression ratio is 128:1. Similarly, we verify the performance of the reconstruction algorithm under nine spatiotemporal compression ratios.

2. Related Work

2.1. Compressive Imaging with Different Optical Modulation Devices

Based on the sparsity in different domains, CI has been successfully applied to improve spatial, temporal, or spectral resolution in an imaging system [19,20,21,22]. The CI system mainly consists of a hardware system for data acquisition and back-end algorithms for high-resolution image reconstruction. The high-resolution light modulation device directly affects the complexity and performance of CI systems. In previous research, mainstream high-spatial-resolution light modulation devices include digital micromirror device (DMD) [14], liquid crystal on silicon (LCoS) [15], etc. Although the high resolution SLM increases the complexity of the optical system to some extent, it can realize high-resolution spatial light modulation in different dimensions and can flexibly load modulation masks, so it has been widely used.
With the development of materials and micronano-manufacturing technology, new light modulation devices are used in CI, such as DoE [23] and metasurfaces [24]. These devices can further reduce the complexity of the system and improve the performance of the system, but their application in practical systems is not mature. Single-dimensional CI usually only needs to introduce a single modulation element for single-dimensional encoding. For multidimensional CI, previous research introduces different modulation elements to the system to realize multidimensional encoding [25]. This method can ensure the effectiveness of encoding in different dimensions, but it also increases the complexity of a system.

2.2. Typical Reconstruction Algorithms for Spatial or Temporal CI

For the traditional model-based reconstruction algorithm, high-resolution image reconstruction is usually solved as an optimization problem, in which prior information is required. For example, typical model-based algorithms, GAP-TV [26] and TwIST [16], take the total variation (TV) as the prior, while DeSCI [27] takes the low-rank prior into account. The model-based algorithm has strong interpretability and high reconstruction accuracy. However, most traditional model-based reconstruction algorithms adopt iterative methods, which is time-consuming and struggles to meet the real-time requirements of applications. In addition, the model-based algorithm has a strong dependence on prior information, which also limits the performance of the algorithm. DL-based algorithms can be further divided into data-driven and model-driven algorithms. Most data-driven methods are based on the convolution neural network (CNN) [28] or recurrent neural network (RNN) [29]. Model-driven methods fuse the optimization strategy of model-based algorithms into end-to-end networks [18,30], which is more efficient and interpretable. Compared with model-based algorithms, the DL-based method has advantages in reconstruction quality and speed.

3. Multiple CR Spatiotemporal Compressive Imaging (Multi-CR STCI)

Figure 1 shows the experimental optical configuration for the proposed STCI method. As shown in the figure, the optical system for STCI is similar to an SCI or TCI system. The imaging system is composed of five main parts: the target, the light source, the imaging lens, the SLM, and the detector array. In the STCI system, the SLM is a critical component that significantly affects system performance. Among the several types of SLM, the DMD offers a compelling combination of spatial resolution, frame rate, and cost-effectiveness. Consequently, we have selected DMD as our spatial light modulator. Light from an object is modulated by patterns displayed on a DMD. The modulated light is then refocused onto a detector array to take measurements. However, the modulation patterns displayed on a DMD and the compressive imaging model are different from the other two types of methods.

3.1. Imaging Model and Spatiotemporal Mask Design

In this work, we aim to develop an STCI system that can handle multiple CRs. Thus, we design the modulation patterns separable in the spatial and temporal domains. We call it a stepwise STCI strategy.
As shown in Figure 2a, the target scenes are initially modulated by spatial masks and compressed in the spatial domain. Then, temporal masks subsequently interact with the spatial compressed frames for temporal modulation. After that, continuous multiple spatiotemporal encoded images are integrated into a single compressed frame within a single exposure period of the camera. The mathematical expression of the STCI imaging model can be expressed as:
Y = i = 1 n t T i B a d d S i X i ,
in which X R n h × n w × n t represents target frames with high spatiotemporal resolution. S R n h × n w × n t indicates the spatial masks. ⊙ is the Hadamard product. We define B a d d as a block addition operation, which first divides the two-dimensional matrix into non-overlapping blocks with size s × s , and then adds the data in each block to obtain a new matrix with size reduced by s times. B a d d represents the compressed sampling in the spatial domain. T R n w / s × n h / s × n t is the temporal mask, whose size is the same as that of the spatial compressive images. The subscript i is the temporal coordinate of the image sequence. i = 1 n t represents the temporal compression during one exposure period of the camera. The spatiotemporal CR is composed of temporal CR and spatial CR, which can be expressed as C R = s 2 × n t .
Stepwise spatial–temporal encoding and compression can achieve the decoupling of spatiotemporal compression to a certain extent, alleviating the difficulty of reconstruction with a large CR. At the same time, considering the complexity of the imaging system, the stepwise strategy requires independent modulation and compression of the spatial and temporal domains, which will inevitably increase the complexity of the system. Therefore, to balance the system complexity and the efficiency of spatiotemporal modulation, we realize equivalent spatiotemporal modulation by preprocessing the spatial and temporal modulation mask. This means that we only need to load a set of spatiotemporal modulation masks to achieve equivalent spatiotemporal modulation as a stepwise method. Observing Equation (1), we can see that the operations contained in the equation are all linear. So we can obtain the following expression by transformation:
Y = i = 1 n t B a d d C T i S i X i ,
in which we define C · as a matrix expansion operation that enlarges the matrix s times by repeating each element in the matrix s × s times. Let M i = C T i S i , then Equation (2) can be expressed as Equation (3), which means that we can obtain a set of equivalent spatiotemporal modulation masks by preprocessing the temporal and spatial masks, as shown in Figure 2b.
Y = i = 1 n t B a d d M i X i ,
Another issue worth noting is that we use a special structural spatial mask to adapt to the multi-CR reconstruction task. As shown in Figure 2c, we chose the quasi-diagonal matrix, in which the upper left and lower right corners of the matrix are 1, and the rest of the matrix is 0, as the spatial sub-kernel. Repeating the spatial sub-kernel in space obtains a spatial mask with the same spatial resolution as the target. This structure can maintain stable original data characteristics with different spatial CRs. Our experimental results also support this point. The temporal masks consist of n t distinct random binary patterns. As discussed previously, we obtain a set of spatiotemporal masks, M i , through pre-processing. When we use these masks to modulate the target scene, the DMD sequentially loads the masks. Then, after the modulation, all modulated frames are summed into one measurement frame. Thus, one set of masks is used to make one measurement frame.
For further discussion of the imaging model, we rewrite the matrix form in Equation (3) into a vector form:
y = H x = H t H s x ,
in which y R ( n h n w / s 2 ) × 1 and x R ( n t n h n w ) × 1 represent spatiotemporal compressive measurements and original high-resolution frames, respectively. x is composed of n t column vectors { x i , i = 1 , , n t } that are concatenated in the column direction, and x i is obtained by reshaping a single high-resolution frame by rows into a column vector. H R ( n h n w / s 2 ) × ( n t n h n w ) is the vector form of spatiotemporal masks M in Equation (3). H t and H s represent temporal and spatial masks, respectively. As shown in Figure 3, the i t h spatial mask is divided into non-overlapping blocks with a size of s × s , and the block is placed in the corresponding position of a zero matrix of the same size as the spatial mask. The matrix is then expanded into a row vector in a row-first way. Go through all the blocks and concatenate the vectors into a spatial matrix. Finally, H s is obtained by splicing t spatial matrices in the row direction. The matrix H t R n × n n t is
H t = D 1 , D 2 , , D n t ,
with n = n h n w / s 2 and D i R n × n is obtained by placing the values in the i t h temporal mask T i sequentially on the diagonal of the diagonal matrix.

3.2. GAP for STCI

In our work, the goal of the algorithm design is to achieve spatiotemporal high-resolution image reconstruction with multiple CRs. We design an unfolding network, STCINet, based on the generalized alternating projection (GAP) [31] algorithm. The optimization problem solved using GAP for STCI is defined as
( x ^ , v ^ ) = a r g m i n x , v 1 2 x v 2 2 + λ g ( v ) s . t . y = H t ( H s x )
Let k denote the number of iterations in the GAP. Given v k , x ( k + 1 ) can be updated via a Euclidean projection of v k onto the linear manifold M : y = H x = H t ( H s x ) :
x k + 1 = v k + H T ( H H T ) 1 ( y H v k ) ) ,
in which H = H t H s . Then, a denoiser is used to update v k ,
v k + 1 = D δ ( x k + 1 ) ,
where δ = λ is the standard deviation of the assumed additive white Gaussian noise.

3.3. Reconstruction Algorithm

The proposed algorithm is an end-to-end unfolding structure. The core of the unfolding method lies in the combination of the optimization ideas of iterative algorithms with the strong learning capabilities of deep neural networks. By replacing iterations in an iterative method with a network module consisting of several layers, an end-to-end trainable model is constructed. This combination not only preserves the effectiveness of iterative algorithms in solving optimization problems and improves the interpretability of the network, but also takes advantage of deep learning to achieve faster convergence and better generalization ability. As shown in Figure 4, the proposed STCI reconstruction unfolding network is composed of multiple phases stacked together. There are two basic modules in a single phase, L and D , which are designed according to Equation (7) and Equation (10), respectively. Taking into account the balance of computational resources and reconstruction effects, in this work, we set up 15 phases in the reconstruction network. L is a linear operation. According to Equation (7), L updates the target reconstruction x k + 1 based on the compressed measurement y , the v k output of the previous phase, and the spatial and temporal mask, H t , H s :
x k + 1 = L ( v k , y , H t , H s ) .
In module L , H t and H s are fully utilized, making the whole network more explainable.
v k + 1 = D δ ( x k + 1 ) ,
The general structure of module D is shown in Figure 5. We design D based on ResUNet [32], in which residual blocks are integrated into a U-Net architecture. In module D , the input is the concatenation of x k + 1 and a denoise guide map that is filled with the value of δ . Note that in different phases of STCINet, the denoiser parameters, except δ , are the same. More discussion of δ will be presented later. To enhance the performance of the denoiser, we use a dynamic convolution unit (DCU) [33] to generate a kernel for input pre-processing. Therefore, we named module D Dy-ResUNet. The detailed structure of the DCU is shown in the red dotted box in Figure 5. The DCU includes two parts, the self-attention module and a set of convolution layers. The weights of the first B convolution layers are calculated from the attention module. Then, the weight of the final convolution layer is the sum of the weights of the first B kernels. In DCU, we set B = 4 instead of a larger value to avoid hard convergence in network training. The DCU module is followed by a standard encoder–decoder architecture with three down-sample and up-sample stages. Each of them is made up of three residual blocks with 16, 32, and 64 channels. The down-sample operations can reduce the requirement for GPU memory. It is worth mentioning that the module D is designed for single frame input to realize CI reconstruction with multiple CRs, which results in a slightly time-consuming increase for reconstruction.
For multi-CR reconstruction tasks, we introduce the module H into the network. Similarly to [32], the module H is designed to generate a parameter δ for each phase depending on the value of n t , s , in which s represents the spatial CR, and n t represents the temporal CR. There are three fully connected layers in the module H . During the training process, multiple n t and s are used to generate the δ values for all K phases of the network.
δ = H ( n t , s ) .

4. Simulation Experiments

4.1. Network Training

As mentioned in Section 3.3, we have designed a reconstruction network, STCINet, to reconstruct high-spatiotemporal-resolution images from compressive measurements. In this section, we mainly describe the training process, the loss function, and the training settings. We chose DAVIS2017 [34] as the dataset for network training, which contains more than ten thousand continuous images from ninety different high-motion scenes. We cropped these 1080p videos into non-overlapping patches with a size of 8 × 120 × 120 for the training process.
We have trained two versions of the reconstruction network for different reconstruction tasks. STCINet-v1 adopts the fixed spatiotemporal CR of 128:1, in which the spatial CR is fixed to s = 4 and the temporal CR n t = 8 . The spatiotemporal CR is equal to s 2 × n t . And STCINet-v2 is trained to solve the multiple CR problem, in which we define the array s-bar = [1,2,4], which includes all spatial CRs that can be selected, and the optional n t -bar = [2,4,8] as the temporal CR selection array. During every training step, the temporal and spatial CRs are randomly selected from n t -bar and s-bar. In other words, STCINet-v2 can handle nine spatiotemporal CRs. And it is a pure temporal reconstruction problem when s = 1 .
We adopt the Adam optimizer to train our model with batch size of 8 and the phase number of the network is set to 15. According to our experiments, more phases are beneficial to the final results and 15 is a proper number considering the trade-off between performance and speed. The initial learning rate is 1 × 10 4 and decays as the epochs increase until obvious loss drops cannot be observed. In this work, the trained model will be directly used for the reconstruction of compressive measurements captured by optical setup without further fine-tuning.

4.2. Results of STCINet_v1

As described in the above section, STCINet-v1 is trained to tackle the highest spatiotemporal CR, 128:1 ( n t = 8 , s = 4 ), which means the sampling rate is lower than 1 % in the data acquisition process. We selected 6 videos with a resolution of 1080 × 1920 to evaluate the model, and the reconstructions of three different scenes are shown in Figure 6. Frames 1, 5 and 8 of the reconstructions are shown to illustrate the performance of the proposed method in the spatial and temporal reconstruction. We can see that STCINet-v1 can recover clear and clean images, whether relatively stationary or high-motion targets, in different motion scenes. We take PSNR (peak signal-to-noise ratio) as the evaluation metric, as shown in Table 1.

4.3. Results of STCINet-v2

STCINet-v2 is designed to reconstruct high-spatiotemporal-resolution images with multiple spatiotemporal CRs. Sequential images with a resolution of n t × 1080 × 1920 are compressed into a single frame with a resolution of ( 1080 / s ) × ( 1920 / s ) . As mentioned above, s and n t represent the spatial and temporal CRs, respectively, whose value ranges are [ 1 , 2 , 4 ] and [ 2 , 4 , 8 ] , respectively. Figure 7 and Figure 8 show the reconstructions when n t = 4 and n t = 8 , and the spatial CR, s, is 1, 2 and 4, respectively.
The first and last frames of the reconstructed images from three different scenes are shown, together with magnified views of the compressed measurements and reconstructions of targets with varying scales and motion patterns in the target scene. For different objects in the moving scene, such as the text in the background with small motion in the subway scene, the tracked helicopter pilot, and the fast sliding plane with large-scale change, STCINet-v2 successfully reconstructs temporal and spatial details from compressed measurements at different CRs. Meanwhile, we can see that with the increase in spatiotemporal CR, the quality of reconstructions decreases to a certain extent, but the temporal and spatial details of the overall image are still well reconstructed. We take PSNR as an evaluation metric, and the PSNR of nine different CRs is calculated and shown in Table 1.
According to the results shown in Table 1, the mean PSNR is 28.99 dB for STCIet-v2 when the spatiotemporal CR is 128:1. Compared with STCINet-v1, there is only a slight decline in the quality of reconstructions, which proves the effectiveness of our method for multi-CR tasks. It can be seen from the results of STCINet-v2 that the quality of reconstructions decreases as the CR increases. Furthermore, by analyzing the influence of increasing temporal CR and increasing spatial CR on reconstructions, we can see that increasing spatial CR will lead to a more obvious decline in the quality of the reconstructions than that of temporal CR. We think the reason is that for high-motion scenes, the sparsity of image sequences in the temporal dimension is higher than the sparsity of the spatial dimension. According to the imaging model, target scenes are modulated by independent temporal and spatial masks, which may fit the redundancy difference in different domains. Moreover, a flexible and efficient network provides the possibility of handling reconstruction tasks with different spatiotemporal CRs.
To further evaluate the performance of STCINet, we conducted VMAF evaluations on 1-sec video sequences with a resolution of 1920 × 1080 and a frame rate of 120 fps [35]. By downsampling the original video to 480 × 270 at 15 fps, we obtained compressed measurements with spatiotemporal CR of 128:1. Our reconstruction algorithm was then used to recover the high-resolution, high-frame-rate video. We evaluated the quality of the reconstructed videos using the VMAF metric under nine different compression ratios for three distinct scenes.
According to Table 2, as the CR increases, the VMAF score of the reconstructed video decreases. When the CR is less than 64:1, the VMAF scores are higher than 80. When the CR is 128:1, the VMAF score of the reconstructed video decreases significantly. We analyze the reasons for this result mainly from two aspects. On the one hand, when we train the network, we use objective evaluation indicators to calculate the loss. Therefore, the reconstruction results will be more inclined to the standards of objective evaluation indicators. In subsequent research, we can consider further improving the algorithm performance by adding subjective evaluation indicators in the training process. On the other hand, there are essential differences between the image degradation factors in compressive imaging and the image degradation factors in video coding, which will also affect the VMAF score.

4.4. Temporal Reconstruction

In order to verify the temporal reconstruction capacity of the spatiotemporal reconstruction network, we compared the temporal reconstruction results of STCINet-v2 and other representative temporal compressive imaging reconstruction algorithms, such as GAP-TV, PnP-FFDNet, BIRNAT. The results of other algorithms are obtained by the released code and model weights. In this section, original high-motion scenes are compressed in the temporal domain only, which means s is equal to 1 for STCINet-v2. We chose 13 videos with size 8 × 1080 × 1920 for testing, which is the same as mentioned in previous simulation experiments. Table 3 shows the average PSNR values of 13 testing scenes, the mean running time, and the maximum GPU memory occupation of different algorithms. GAP-TV and PnP-FFDNet are flexible enough to deal with any temporal CR problem, but they are time-consuming, as those two algorithms adopt the iteration strategy. BIRNAT, which is an end-to-end network based on RNN, produces the best reconstruction result; however, it is memory-consuming that we have to chunk videos into blocks for testing. STCINet-v2 can reconstruct high-resolution videos in 3 s with a lower memory cost, and the quality of the reconstruction frames is close to BIRNAT. Moreover, STCINet-v2 can handle multi-CR tasks, while BIRNAT can only handle the specific CR.
Figure 9 shows one of the temporal reconstruction frames from two different scenes with temporal CR of n t = 8 . Comparing the enlarged details, we can see that GAP-TV and PnP-FFDNet become noisy or over-smooth results. BIRNAT recovers the richest details, and STCINet-v2 performs a high-quality reconstruction that is closer to BIRNAT with a smaller computing source.

4.5. Ablation

In this section, two sets of ablation experiments are designed to demonstrate the effectiveness of the modules H and DCU in STCINet-v2. We design an experimental paradigm, in which we remove the two modules from STCINet-v2, respectively, and then the same training strategy is used to train the model. The results of the ablation experiment are compared with the results of STCINet-v2. Table 4 shows the PSNR comparison of the three groups of experiments with nine CRs. w / o H means that the module H is removed from the network. The module H is a multi-CR reconstruction controller, which generates different hyperparameters according to different CRs to adjust the reconstruction process. When module H is removed, the higher CR contributes more to the loss during network training due to the increased difficulty of reconstruction, thus guiding network optimization toward reconstruction tasks of higher CR. On the other hand, at a CR of 32:1, the reconstruction results without module H are better than those with module H , while the reconstruction results at other CRs are inferior to those with module H . We believe this is because the network is unable to adapt to multi-CR tasks and becomes trapped in the optimization of a specific CR. The existence of module H can disperse the attention of the network to different CRs. The final average PSNR shows obvious advantages of H . It can be seen from the table that the introduction of the DCU module has a significant impact on improving the quality of reconstructions with multiple CRs.

5. Optical Experiment

5.1. Optical Setup

To further verify the effectiveness of the proposed method in practical applications, we built an optical system for verification. Based on the previous analysis, by preprocessing the spatial and temporal masks, we can use only one set of masks to achieve spatiotemporal modulation. Therefore, in the optical system, we only need to introduce one SLM. We choose DMD as the spatial light modulator due to its high frame rate and spatial resolution. The general structure of the optical system is shown in Figure 1. We use a rotating disk driven by an electric motor as the target. Roman numerals and patterns, such as stripes, are printed in circles on the disk. The light source is a customized ring light. We choose DLP6500 as the spatiotemporal modulator. The front imaging system consists of a single lens to image the target onto the DMD, while the back-end imaging system is a 4 f system consisting of two lenses to image the DMD onto the detector. We use Basler ACA720-520 to collect spatiotemporal compressive frames.
Based on the optical system, we have verified the spatiotemporal reconstruction with two CRs. Since the spatial CR is determined by the optical structure, the spatial CR is fixed to s = 4 in the optical setup. That is, in the back-end imaging path, the 4 × 4 pixels in the DMD will converge to a single pixel of the detector. The temporal CR can be adjusted by controlling the frame rate of the DMD and the detector. In the experiment, to guarantee the matching of the compressed measurement and the corresponding masks, we control the detector exposure by the DMD trigger signal. Specifically, DMD sends the trigger signal to the detector when t masks are loaded. The detector receives the trigger signal to complete an exposure.
With spatial CR of 16:1 ( s = 4 ), the compressed measurement with temporal CR of 4:1 ( n t = 4) and 8:1 ( n t = 8) were collected, respectively. So the spatiotemporal CRs are 64:1 and 128:1, respectively. In the experiment, we set the frame rate of DMD to 500 fps, while the camera integration time is set to 16 ms ( 62.5 fps) for spatiotemporal CR of 128:1 and 8 ms (125 fps) for spatiotemporal CR of 64:1. We capture three different objects, the horizontal stripe pattern, the number 9, and the vertical stripe pattern. The spatial resolution of the measurements is 30 × 30 . For two different spatiotemporal CRs, the resolution of reconstructed high-resolution images is 4 × 120 × 120 and 8 × 120 × 120 . We directly use STCINet-v2 trained on the public dataset to reconstruct the high-resolution images from compressed measurements collected by the optical system without additional network fine-tuning.

5.2. Optical Results

Reconstructions with spatiotemporal CR of 64:1 are shown in Figure 10. From the reconstruction results, we can see that spatial details that cannot be distinguished in the measurements, such as vertical stripes, have been successfully reconstructed. The motion state of the target has also been well reconstructed. Figure 11 shows the reconstructions with spatiotemporal CR of 128:1. Compared with the low spatiotemporal CR of 64:1, the reconstruction quality of the target is still satisfactory even with a higher CR.
To further demonstrate the ability of the algorithm to reconstruct fine details, we performed additional experiments. We reconstruct at different spatial resolutions for compressed measurement with a CR of 128:1, where n t = 8 and s = 4 . The fourth and fifth rows of the Figure 11 show the results of reconstructions with 2 times improvement in spatial resolution and without an improvement in spatial resolution, respectively. In the temporal dimension, we reconstructed 8 frames. Specifically, a single measurement frame with the resolution of 30 × 30 is reconstructed into images with resolutions of 60 × 60 × 8 and 30 × 30 × 8 , respectively. For reconstructions without improvement in spatial resolution, the fine spatial details of the target are indistinguishable. The 2 times reconstruction shows a marked improvement in spatial resolution, but the boundaries of the stripes remain somewhat fuzzy. However, 4 times reconstructions yield a much more complete recovery of the target’s spatial details. These results suggest that spatiotemporal reconstruction is capable of effectively recovering detailed information of moving targets.

5.3. Discussions

The optical experiments validated the effectiveness and feasibility of the STCI method. Experimental results demonstrate that the imaging system can achieve up to a 4 times improvement in spatial resolution and an 8 times enhancement in temporal resolution simultaneously using the STCI. Note that the spatial and temporal resolutions of DMD restrict the system’s resolutions. However, it is not uncommon for a DMD to have spatial resolution greater than 1024 × 1024 and frame rate greater than 10 K fps, which can meet the requirements in most applications.
In addition, dynamic range is a critical performance metric for imaging systems. We discuss the dynamic range in STCI from two perspectives. The first is about compressive measurements. If we assume that a compressive measurement frame accumulates temporal and spatial original scene pixels, then the measurements can easily exceed the dynamic range of a detector. To address the issue, in the simulated experiments, we multiply the template by a factor 1 / s 2 B , where B is the number of frames compressed into one measurement frame, and s represents that s × s pixels are spatially compressed into one pixel of a measurement frame. This ensures that the measurement frame is still in the dynamic range of a detector. The second is about the dynamic range in terms of each detectable object frame. Although the STCI method improves the resolution of the system, it introduces a reduction in the dynamic range of the original object frames. In STCI, we optically compress the light from the s × s × B object pixels into one STCI measurement. Thus, the dynamic range of a detector pixel is dispersed into B object frames. The dynamic range of the detectable object frames is reduced. This is consistent with what appears in high-speed cameras. As the speed increases, the object light for each frame is reduced. This makes the noise more noticeable. However, in STCI, the reconstruction network has been designed and trained to deal with noise. Hence, the back-end algorithm can compensate for the loss of dynamic range during reconstruction, ensuring that no significant dynamic range degradation is observed in the final results.

6. Conclusions

In this paper, we present a spatiotemporal compressive imaging (STCI) method. To balance the encoding efficiency and system simplicity in STCI, we revisited the mathematical model of STCI and introduced a stepwise strategy for spatiotemporal encoding. By carefully preprocessing the mask, we demonstrate that effective spatiotemporal encoding can be achieved with a single spatial light modulator (SLM), significantly simplifying the optical system. Furthermore, we developed a novel deep learning-based reconstruction method that incorporates a hyperparameter module to enable reconstruction across a range of spatiotemporal compression ratio (CRs). Through extensive simulations and experimental validations, we verified the efficacy of our approach in accurately reconstructing high-speed motion scenes with a spatiotemporal compression ratio of 128:1. Importantly, our method demonstrated robust performance across nine different spatiotemporal CRs, highlighting its versatility and adaptability. This work explores the development of more efficient and practical STCI systems with enhanced capabilities for capturing dynamic scenes. In our future work, we will further explore the improvement of image quality in spatiotemporal compressive imaging from both the perspectives of imaging models and imaging algorithms.

Author Contributions

Conceptualization, J.K.; Methodology, X.H. and D.Z.; Software, D.Z.; Validation, X.H.; Writing—original draft, X.H. and D.Z.; Writing—review & editing, J.K. All authors have read and agreed to the published version of the manuscript.

Funding

National Natural Science Foundation of China (U2241275).

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

Data underlying the results presented in this paper are not publicly available at this time but may be obtained from the authors upon reasonable request.

Conflicts of Interest

The authors declare no conflicts of interest.

References

  1. Gao, L.; Liang, J.; Li, C.; Wang, L.V. Single-shot compressed ultrafast photography at one hundred billion frames per second. Nature 2014, 516, 74–77. [Google Scholar] [CrossRef]
  2. El-Desouki, M.; Deen, M.J.; Fang, Q.; Liu, L.; Tse, F.; Armstrong, D. CMOS image sensors for high speed applications. Sensors 2009, 9, 430–444. [Google Scholar] [CrossRef] [PubMed]
  3. Donoho, D.L. Compressed sensing. IEEE Trans. Inf. Theory 2006, 52, 1289–1306. [Google Scholar] [CrossRef]
  4. Ke, J.; Lam, E.Y. Object reconstruction in block-based compressive imaging. Opt. Express 2012, 20, 22102–22117. [Google Scholar] [CrossRef] [PubMed]
  5. Zhou, Q.; Ke, J.; Lam, E.Y. Dual-waveband temporal compressive imaging. In Proceedings of the Computational Optical Sensing and Imaging, Munich, Germany, 23–27 June 2019; Optica Publishing Group: Washington, DC, USA, 2019; p. CTu2A.8. [Google Scholar]
  6. Cao, X.; Yue, T.; Lin, X.; Lin, S.; Yuan, X.; Dai, Q.; Carin, L.; Brady, D.J. Computational snapshot multispectral cameras: Toward dynamic capture of the spectral world. IEEE Signal Process. Mag. 2016, 33, 95–108. [Google Scholar] [CrossRef]
  7. Dong, D.W.; Atick, J.J. Statistics of natural time-varying images. Netw. Comput. Neural Syst. 1995, 6, 345. [Google Scholar] [CrossRef]
  8. Treeaporn, V.; Ashok, A.; Neifeld, M.A. Space–time compressive imaging. Appl. Opt. 2012, 51, A67–A79. [Google Scholar] [CrossRef]
  9. Harmany, Z.T.; Marcia, R.F.; Willett, R.M. Spatio-temporal compressed sensing with coded apertures and keyed exposures. arXiv 2011, arXiv:1111.7247. [Google Scholar]
  10. Ke, J.; Zhang, L.; Lam, E.Y. Temporal compressed measurements for block-wise compressive imaging. In Proceedings of the Mathematics in Imaging, San Francisco, CA, USA, 2–7 February 2019; Optical Society of America: Washington, DC, USA, 2019; p. JW4B.1. [Google Scholar]
  11. Gan, L. Block compressed sensing of natural images. In Proceedings of the 2007 15th International Conference on Digital Signal Processing, Cardiff, UK, 1–4 July 2007; pp. 403–406. [Google Scholar]
  12. Tsai, T.H.; Llull, P.; Yuan, X.; Carin, L.; Brady, D.J. Spectral-temporal compressive imaging. Opt. Lett. 2015, 40, 4054–4057. [Google Scholar] [CrossRef]
  13. Zhang, B.; Yuan, X.; Deng, C.; Zhang, Z.; Suo, J.; Dai, Q. End-to-end snapshot compressed super-resolution imaging with deep optics. Optica 2022, 9, 451–454. [Google Scholar] [CrossRef]
  14. Llull, P.; Liao, X.; Yuan, X.; Yang, J.; Kittle, D.; Carin, L.; Sapiro, G.; Brady, D.J. Coded aperture compressive temporal imaging. Opt. Express 2013, 21, 10526–10545. [Google Scholar] [CrossRef] [PubMed]
  15. Reddy, D.; Veeraraghavan, A.; Chellappa, R. P2C2: Programmable pixel compressive camera for high speed imaging. In Proceedings of the CVPR 2011, Colorado Springs, CO, USA, 20–25 June 2011; pp. 329–336. [Google Scholar]
  16. Bioucas-Dias, J.M.; Figueiredo, M.A. A new TwIST: Two-step iterative shrinkage/thresholding algorithms for image restoration. IEEE Trans. Image Process. 2007, 16, 2992–3004. [Google Scholar] [CrossRef] [PubMed]
  17. Ma, J.; Liu, X.Y.; Shou, Z.; Yuan, X. Deep tensor admm-net for snapshot compressive imaging. In Proceedings of the IEEE/CVF International Conference on Computer Vision, Seoul, Republic of Korea, 27 October–2 November 2019; pp. 10223–10232. [Google Scholar]
  18. Zhang, J.; Ghanem, B. ISTA-Net: Interpretable optimization-inspired deep network for image compressive sensing. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Salt Lake City, UT, USA, 18–23 June 2018; pp. 1828–1837. [Google Scholar]
  19. Cui, C.; Ke, J. Spatial compressive imaging deep learning framework using joint input of multi-frame measurements and degraded maps. Opt. Express 2022, 30, 1235–1248. [Google Scholar] [CrossRef]
  20. Zhang, L.; Lam, E.Y.; Ke, J. Temporal compressive imaging reconstruction based on a 3D-CNN network. Opt. Express 2022, 30, 3577–3591. [Google Scholar] [CrossRef]
  21. Arce, G.R.; Brady, D.J.; Carin, L.; Arguello, H.; Kittle, D.S. Compressive coded aperture spectral imaging: An introduction. IEEE Signal Process. Mag. 2013, 31, 105–115. [Google Scholar] [CrossRef]
  22. Babbitt, W.R.; Barber, Z.W.; Renner, C. Compressive laser ranging. Opt. Lett. 2011, 36, 4794–4796. [Google Scholar] [CrossRef]
  23. Mengu, D.; Tabassum, A.; Jarrahi, M.; Ozcan, A. Snapshot multispectral imaging using a diffractive optical network. Light Sci. Appl. 2023, 12, 86. [Google Scholar] [CrossRef]
  24. Hu, X.; Xu, W.; Fan, Q.; Yue, T.; Yan, F.; Lu, Y.; Xu, T. Metasurface-based computational imaging: A review. Adv. Photonics 2024, 6, 014002. [Google Scholar] [CrossRef]
  25. Wang, X.; Xu, T.; Zhang, Y.; Xu, C.; Fan, A.; Yu, Y. A multi-channel spectral coding method for the coded aperture tunable filter spectral imager. In Proceedings of the Optics, Photonics and Digital Technologies for Imaging Applications VI. SPIE, Online, 6–10 April 2020; Volume 11353, pp. 271–279. [Google Scholar]
  26. Yuan, X. Generalized alternating projection based total variation minimization for compressive sensing. In Proceedings of the 2016 IEEE International Conference on Image Processing (ICIP), Phoenix, AZ, USA, 25–28 September 2016; pp. 2539–2543. [Google Scholar] [CrossRef]
  27. Liu, Y.; Yuan, X.; Suo, J.; Brady, D.J.; Dai, Q. Rank Minimization for Snapshot Compressive Imaging. IEEE Trans. Pattern Anal. Mach. Intell. 2019, 41, 2990–3006. [Google Scholar] [CrossRef]
  28. Kulkarni, K.; Lohit, S.; Turaga, P.; Kerviche, R.; Ashok, A. Reconnet: Non-iterative reconstruction of images from compressively sensed measurements. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Las Vegas, NV, USA, 27–30 June 2016; pp. 449–458. [Google Scholar]
  29. Cheng, Z.; Lu, R.; Wang, Z.; Zhang, H.; Chen, B.; Meng, Z.; Yuan, X. BIRNAT: Bidirectional Recurrent Neural Networks with Adversarial Training for Video Snapshot Compressive Imaging. In Proceedings of the 16th European Conference, Glasgow, UK, 23–28 August 2020; pp. 258–275. [Google Scholar] [CrossRef]
  30. Yuan, X.; Liu, Y.; Suo, J.; Dai, Q. Plug-and-Play Algorithms for Large-Scale Snapshot Compressive Imaging. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), Seattle, WA, USA, 14–19 June 2020. [Google Scholar]
  31. Liao, X.; Li, H.; Carin, L. Generalized alternating projection for weighted-2,1 minimization with applications to model-based compressive sensing. Siam J. Imaging Sci. 2014, 7, 797–823. [Google Scholar] [CrossRef]
  32. Zhang, K.; Gool, L.V.; Timofte, R. Deep Unfolding Network for Image Super-Resolution. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), Seattle, WA, USA, 13–19 June 2020. [Google Scholar]
  33. Chen, Y.; Dai, X.; Liu, M.; Chen, D.; Yuan, L.; Liu, Z. Dynamic convolution: Attention over convolution kernels. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Seattle, WA, USA, 13–19 June 2020; pp. 11030–11039. [Google Scholar]
  34. Pont-Tuset, J.; Perazzi, F.; Caelles, S.; Arbeláez, P.; Sorkine-Hornung, A.; Van Gool, L. The 2017 davis challenge on video object segmentation. arXiv 2017, arXiv:1704.00675. [Google Scholar]
  35. Mercat, A.; Viitanen, M.; Vanne, J. UVG dataset: 50/120fps 4K sequences for video codec analysis and development. In Proceedings of the 11th ACM Multimedia Systems Conference, Istanbul, Turkey, 8–11 June 2020; pp. 297–302. [Google Scholar]
Figure 1. Schematic illustration of the experimental optical configuration for the proposed STCI method.
Figure 1. Schematic illustration of the experimental optical configuration for the proposed STCI method.
Sensors 25 01334 g001
Figure 2. (a) Stepwise spatial–temporal encoding and compression model. (b) Spatial and temporal mask design and preprocessing for spatiotemporal mask. (c) Structure of spatial kernel; s 2 is the spatial sub-mask for s = 2 and s 4 is the spatial sub-mask for s = 4 . (d) One of spatiotemporal masks in this work.
Figure 2. (a) Stepwise spatial–temporal encoding and compression model. (b) Spatial and temporal mask design and preprocessing for spatiotemporal mask. (c) Structure of spatial kernel; s 2 is the spatial sub-mask for s = 2 and s 4 is the spatial sub-mask for s = 4 . (d) One of spatiotemporal masks in this work.
Sensors 25 01334 g002
Figure 3. Illustration of transformer for spatial mask.
Figure 3. Illustration of transformer for spatial mask.
Sensors 25 01334 g003
Figure 4. Overall structure of STCINet. It consists of three modules, H (blue box), L (green box), D (yellow box). Module L is for implementing Equation (7), module H is a multi-CR controller and module D is a CNN-based network, which is detailed in Figure 5.
Figure 4. Overall structure of STCINet. It consists of three modules, H (blue box), L (green box), D (yellow box). Module L is for implementing Equation (7), module H is a multi-CR controller and module D is a CNN-based network, which is detailed in Figure 5.
Sensors 25 01334 g004
Figure 5. Structure of D (Dy-ResUNet).
Figure 5. Structure of D (Dy-ResUNet).
Sensors 25 01334 g005
Figure 6. Reconstruction results of STCINet-v1 with CR = 128:1. The first column is the spatiotemporal measurement values and their enlarged images. The second, third and fourth columns are the first, fifth and eighth frames of the reconstructions, respectively.
Figure 6. Reconstruction results of STCINet-v1 with CR = 128:1. The first column is the spatiotemporal measurement values and their enlarged images. The second, third and fourth columns are the first, fifth and eighth frames of the reconstructions, respectively.
Sensors 25 01334 g006
Figure 7. Reconstructions with different spatiotemporal CRs. Temporal CR is fixed to n t = 4 and spatial CR, s, is 1 , 2 , 4 respectively. (a,d,g) are reconstructions of different moving targets, in which odd rows are the first frame of reconstructions and even rows are the last frame. (b,e,h) are compressive measurements corresponding to the green box in reconstruction results, and measurements with different spatial CRs are enlarged to the same size for display. (c,f,i) are enlarged images of the box selected area in reconstructions.
Figure 7. Reconstructions with different spatiotemporal CRs. Temporal CR is fixed to n t = 4 and spatial CR, s, is 1 , 2 , 4 respectively. (a,d,g) are reconstructions of different moving targets, in which odd rows are the first frame of reconstructions and even rows are the last frame. (b,e,h) are compressive measurements corresponding to the green box in reconstruction results, and measurements with different spatial CRs are enlarged to the same size for display. (c,f,i) are enlarged images of the box selected area in reconstructions.
Sensors 25 01334 g007
Figure 8. Reconstructions with different spatiotemporal CRs. Temporal CR is fixed to n t = 8 and spatial CR, s, is 1 , 2 , 4 respectively. (a,d,g) are reconstructions of different moving targets, in which odd rows are the first frame of reconstructions and even rows are the last frame. (b,e,h) are compressive measurements corresponding to the green box in reconstruction results, and measurements with different spatial CRs are enlarged to the same size for display. (c,f,i) are enlarged images of the box selected area in reconstructions.
Figure 8. Reconstructions with different spatiotemporal CRs. Temporal CR is fixed to n t = 8 and spatial CR, s, is 1 , 2 , 4 respectively. (a,d,g) are reconstructions of different moving targets, in which odd rows are the first frame of reconstructions and even rows are the last frame. (b,e,h) are compressive measurements corresponding to the green box in reconstruction results, and measurements with different spatial CRs are enlarged to the same size for display. (c,f,i) are enlarged images of the box selected area in reconstructions.
Sensors 25 01334 g008
Figure 9. Reconstructions of temporal compressed measurements using different algorithms.
Figure 9. Reconstructions of temporal compressed measurements using different algorithms.
Sensors 25 01334 g009
Figure 10. Reconstructions of STCINet-v2 with C R = 64:1, in which n t = 4 , s = 4 . (ac) are spatiotemporal compressive measurements of three different moving targets (green boxes) and their enlarged images. The right side shows the 4-frame spatiotemporal reconstruction image corresponding to the measurement.
Figure 10. Reconstructions of STCINet-v2 with C R = 64:1, in which n t = 4 , s = 4 . (ac) are spatiotemporal compressive measurements of three different moving targets (green boxes) and their enlarged images. The right side shows the 4-frame spatiotemporal reconstruction image corresponding to the measurement.
Sensors 25 01334 g010
Figure 11. Reconstructions of STCINet-v2 with C R = 128:1, in which n t = 8 , s = 4 . (ac) are spatiotemporal compressive measurements of three different moving targets (green boxes) and their enlarged images. The right is the 8-frame spatiotemporal reconstruction image corresponding to the measurement. The last two rows in the figure are the results of 2 times spatial reconstruction and no spatial reconstruction on the measurement (c), respectively. The reconstruction results are interpolated to the same size as that of 4 times spatial reconstruction (row 4). (d,e) are enlarged images of the blue box in the reconstruction results.
Figure 11. Reconstructions of STCINet-v2 with C R = 128:1, in which n t = 8 , s = 4 . (ac) are spatiotemporal compressive measurements of three different moving targets (green boxes) and their enlarged images. The right is the 8-frame spatiotemporal reconstruction image corresponding to the measurement. The last two rows in the figure are the results of 2 times spatial reconstruction and no spatial reconstruction on the measurement (c), respectively. The reconstruction results are interpolated to the same size as that of 4 times spatial reconstruction (row 4). (d,e) are enlarged images of the blue box in the reconstruction results.
Sensors 25 01334 g011
Table 1. PSNR (dB) using STCINet-v2/STCINet-v1 with different compression ratios.
Table 1. PSNR (dB) using STCINet-v2/STCINet-v1 with different compression ratios.
n t / s / CR 2/1/2:12/2/8:12/4/32:14/1/4:14/2/16:14/4/64:18/1/8:18/2/32:18/4/128:1
Subway46.5438.7531.6343.6336.7430.6839.7533.5328.33/28.56
Aerobatics36.7934.2230.9735.2533.0430.2033.6931.3828.48/28.79
Plane42.2634.3829.1040.8234.3029.2938.8934.1529.98/30.16
Chameleon42.7234.4528.7739.2031.9927.9635.0529.1925.82/26.09
Car-race41.1534.5427.8739.7233.7228.7338.5633.3727.91/28.40
skiing44.5439.0935.2541.9937.5739.1235.8429.8933.44/33.69
Avg.42.3335.9130.6040.1034.5631.0036.9631.9228.99/29.28
Table 2. VMAF with different compression ratios.
Table 2. VMAF with different compression ratios.
n t / s / CR 2/1/2:12/2/8:12/4/32:14/1/4:14/2/16:14/4/64:18/1/8:18/2/32:18/4/128:1
Beauty91.9489.3481.6390.1586.4177.8886.8180.9671.56
Bosphours99.9399.9089.3399.9297.6979.3199.4189.4366.15
Honeybee91.9489.3481.6390.1586.4177.8886.8180.9678.41
Avg.94.6092.8684.1993.4090.1778.3591.0183.7872.04
Table 3. Comparison of different TCI algorithms (PSNR: dB, running time: s, memory: MB).
Table 3. Comparison of different TCI algorithms (PSNR: dB, running time: s, memory: MB).
AlgorithmsAverage PSNRRunning TimeMemory
GAP-TV32.0139.5-
PnP-FFDNet35.9329.89017
BIRNAT37.93.8>24,000
Ours37.142.93199
Table 4. PSNR (dB) in ablation experiments.
Table 4. PSNR (dB) in ablation experiments.
n t / s / CR 2/1/2:12/2/8:12/4/32:14/1/4:14/2/16:14/4/64:18/1/8:18/2/32:18/4/128:1Avg.
STCINet42.3335.9130.6040.1034.5631.0036.9631.9228.9934.68
w/o H 38.6734.7230.6637.4133.5330.0735.3932.0328.8633.48
w/o DCU38.4134.2830.5636.8933.1929.8734.9431.7528.7033.29
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Hao, X.; Zhao, D.; Ke, J. Multiple CR Spatiotemporal Compressive Imaging System. Sensors 2025, 25, 1334. https://doi.org/10.3390/s25051334

AMA Style

Hao X, Zhao D, Ke J. Multiple CR Spatiotemporal Compressive Imaging System. Sensors. 2025; 25(5):1334. https://doi.org/10.3390/s25051334

Chicago/Turabian Style

Hao, Xiaowen, Dingaoyu Zhao, and Jun Ke. 2025. "Multiple CR Spatiotemporal Compressive Imaging System" Sensors 25, no. 5: 1334. https://doi.org/10.3390/s25051334

APA Style

Hao, X., Zhao, D., & Ke, J. (2025). Multiple CR Spatiotemporal Compressive Imaging System. Sensors, 25(5), 1334. https://doi.org/10.3390/s25051334

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop