You are currently on the new version of our website. Access the old version .
BiomimeticsBiomimetics
  • Article
  • Open Access

8 January 2026

Bio-Inspired Ghost Imaging: A Self-Attention Approach for Scattering-Robust Remote Sensing

,
,
,
,
,
,
and
1
Key Laboratory of Biomimetic Robots and Systems, Ministry of Education, Beijing 100081, China
2
School of Optics and Photonics, Beijing Institute of Technology, Beijing 100081, China
3
Bionic Robots and Photonics Key Laboratory, Ministry of Education, School of Optics and Photonics, Beijing Institute of Technology, Beijing 100081, China
4
Xi’an Moder Control Technology Research Institute, Xi’an 710018, China
This article belongs to the Special Issue Bionic Vision Applications and Validation

Abstract

Ghost imaging (GI) offers a robust framework for remote sensing under degraded visibility conditions. However, atmospheric scattering in phenomena such as fog introduces significant noise and signal attenuation, thereby limiting its efficacy. Inspired by the selective attention mechanisms of biological visual systems, this study introduces a novel deep learning (DL) architecture that embeds a self-attention mechanism to enhance GI reconstruction in foggy environments. The proposed approach mimics neural processes by modeling both local and global dependencies within one-dimensional bucket measurements, enabling superior recovery of image details and structural coherence even at reduced sampling rates. Extensive simulations on the Modified National Institute of Standards and Technology (MNIST) and a custom Human-Horse dataset demonstrate that our bio-inspired model outperforms conventional GI and convolutional neural network-based methods. Specifically, it achieves Peak Signal-to-Noise Ratio (PSNR) values between 24.5–25.5   dB / m and Structural Similarity Index Measure (SSIM) values of approximately 0.8 under high scattering conditions ( β     3.0   dB / m ) and moderate sampling ratios ( N     50 % ). A comparative analysis confirms the critical role of the self-attention module, providing high-quality image reconstruction over baseline techniques. The model also maintains computational efficiency, with inference times under 0.12   s , supporting real-time applications. This work establishes a new benchmark for bio-inspired computational imaging, with significant potential for environmental monitoring, autonomous navigation and defense systems operating in adverse weather.

1. Introduction

Ghost imaging (GI) has emerged as a powerful computational imaging technique with significant potential for remote sensing applications, offering innovative solutions to constraints inherent in conventional imaging systems [1]. By leveraging the correlation between spatially structured illumination and a single-pixel detector (SPD), GI provides notable robustness to noise and reduced hardware complexity. However, its performance is critically degraded in scattering media, such as fog, haze, and turbulence, where image reconstruction suffers from severe contrast loss, blurred edges, and diminished detail fidelity [2]. These challenges are particularly acute in real-time imaging, where stringent constraints on measurement time and sampling rates further impair practical utility.
To confront the challenge of imaging through obscurants, we can look to biological systems that have evolved advanced strategies for perception in degraded visual environments. From the neural circuitry in the mammalian visual cortex that performs selective noise suppression to the ability of certain marine species to discern structures in turbid water, biology demonstrates the power of attention-based processing to prioritize salient information from noisy sensory inputs [3,4,5,6]. This bio-inspired principle, which focuses computational resources on the most informative features, offers a compelling paradigm for advancing computational imaging under adverse conditions.
Driven by this insight, our work introduces a bio-inspired deep learning (DL) framework for GI that embeds a self-attention mechanism to mimic selective processing. Traditional techniques for improving GI, such as iterative denoising [7], scalar matrix methods [8], differential GI [9], and Hadamard-based approaches [10], are fundamentally based on linear correlation. Under high scattering conditions, the bucket signal becomes dominated by noise, causing these methods to fail as the underlying correlation assumption breaks down. While DL, particularly Convolutional Neural Networks (CNNs), has improved GI [11,12], their localized receptive fields struggle to model the long-range dependencies necessary to distinguish a coherent global structure from globally pervasive scattering noise. Our bio-inspired transformer architecture addresses this directly by using a self-attention mechanism to integrate global context, learning a nonlinear mapping from degraded bucket signals to clear images that is inherently robust to this noise [13].
The advent of Transformer-based architectures, with their core Multi-Head Self-Attention (MHSA) mechanism [14], presents a transformative opportunity to capture these global contextual relationships, as initially demonstrated in vision tasks by the Vision Transformer (ViT) [14]. However, the direct application of standard Transformers to GI in low-visibility conditions is hampered by their prohibitive computational cost. Although efficient variants like SNR-Net [15] and Uformer [16] have been proposed to mitigate this through architectural optimizations, the potential of tailored attention mechanisms to simultaneously reduce sampling requirements, improve reconstruction fidelity, and maintain efficiency for GI in fog remains largely untapped.
The main objective of this study are:
  • To develop a novel, bio-inspired DL architecture incorporating a self-attention mechanism for robust GI reconstruction technique in foggy environments for remote sensing applications.
  • To enable high-fidelity image recovery from highly compressed and severely degraded bucket signals.
  • To rigorously evaluate the model performance against state-of-the-art methods and through an ablation study to quantify the contribution of the self-attention mechanism.
  • To demonstrate the hybrid model computational efficiency, ensuring its viability for real-time remote sensing applications.
To overcome this gap by developing a bio-inspired self-attention architecture for GI, explicitly designed for a spectrum of fog density. This work introduces a novel, bio-inspired architecture that adapts self-attention to the unique challenge of reconstructing images from globally degraded, one-dimensional bucket signals. Our model is specifically designed to capture the long-range dependencies necessary to overcome scattering noise, a capability that localized convolutional filters lack. We rigorously train and validate our model on the MNIST and a remote sensing Human-Horse dataset under varying scattering coefficients ( β { 0.1 0.5 1.0 2.0 3.0 5.0 } dB/m) and sampling ratios ( N { 10 % 25 % 50 % 75 % 100 % } ). The results demonstrate that our method achieves superior reconstruction quality, with PSNR values up to 24.5–25.5 dB and SSIM values of approximately 0.8–0.85 under high-fog conditions ( β     3.0 dB/m) and moderate sampling ratios ( N     50 % ), underscoring its significant potential for applications in environmental monitoring, security surveillance, and defense.

2. Proposed Model and Prior

2.1. Ghost Imaging in Foggy Environments

Building upon the bio-inspired premise of our approach, this section details the computational imaging framework designed to emulate the robustness of biological vision in scattering media. The core challenge in foggy environments is the severe degradation of optical signals, a problem that natural systems overcome through efficient neural processing. Our method addresses this by augmenting the physical GI setup with a neural-inspired reconstruction network that mimics this selective, attention-driven processing.
The standard GI process begins with a series of structured illumination patterns P i ( x , y ) preloaded onto a Digital Micromirror Device (DMD) and speckle patterns were generated according to Equation (3) with N = 64 for the 64 × 64 Human-Horse dataset and N = 28 for the MNIST dataset. The distance between the DMD and the target object was simulated as d = 50   m. As illustrated in Figure 1a, a laser illuminates the DMD, and the modulated light is directed through a beam splitter (BS) and a beam expander (BE) to ensure sufficient spatial coverage for long-range propagation. This structured light then traverses the foggy medium S ( x , y ) , undergoing significant scattering and distortion before illuminating the target object F ( x , y ) . A SPD collects the resulting light, which has interacted with both the object and the scattering medium, producing a one-dimensional bucket measurement vector H m . In foggy environments, the scattering medium S ( x , y ) severely degrades the optical signal through two primary mechanisms: attenuation and additive scattering noise. Attenuation reduces the intensity of the light incident on the target F ( x , y ) , weakening the object’s reflected signal and leading to a lower signal-to-noise ratio (SNR) in the bucket measurement H m . Additive noise arises from photons being scattered by fog particles directly into the bucket detector without interacting with the target. This introduces uncorrelated noise into the bucket signal. In the conventional correlation reconstruction Equation (1), this combined effect manifests as a significant loss of contrast, blurred edges, and a noisy, low-fidelity image O G I as the correlated signal from the object is effected by the uncorrelated scattering noise.
Figure 1. (a) Schematic of the GI system operating in a foggy medium. A laser beam is modulated by a DMD to generate structured patterns, which are directed through a BS and BE for long-range propagation. The light scattered by the object is captured by a bucket detector (BD). (b) The computed bucket signals are processed by a self-attention based DL model to generate a high-quality image, emulating a neural-inspired reconstruction pathway.
The conventional correlation-based reconstruction in GI is expressed as:
O G I = 1 n i = 1 n ( H H ) P i ( x , y ) ,
where H is the bucket measurement vector. The value of each bucket measurement H i is given by:
H i = P i ( x , y ) × F ( x , y ) × S i ( x , y ) d x d y ,
Here, P i ( x , y ) represents the i-th illumination pattern as illustrated in Figure 2, with i = 1 , 2 , , M , and M being the total number of measurements. In this work, we employ speckle patterns formulated as:
P k , l ( x , y ) = α k α l cos π k ( 2 x + 1 ) 2 N cos π l ( 2 y + 1 ) 2 N ,
where,
α k = 1 N if k = 0 2 N if k 0 , α l = 1 N if l = 0 2 N if l 0 .
Figure 2. Illustration of illumination patterns utilized in GI under scattering environment.
While this physical setup provides the raw data, the conventional reconstruction in Equation (1) falters under strong scattering, much like a rudimentary sensory system without higher-level processing. To overcome this, we introduce a neural-inspired reconstruction model that learns a direct, intelligent mapping from the degraded 1D bucket signals to a clear 2D image. This process, which replaces the linear correlation of standard GI, is described as:
I ^ ( x , y ) = R ( H ) ,
where H = [ H 1 , H 2 , , H M ] T is the vector of bucket measurements (the correlation of reflected light and known illumination patterns), and  R ( · ) is the complex, nonlinear reconstruction function learned by our self-attention model. The output I ^ ( x , y ) approximates the true reflectivity distribution of the object.
The model is trained and evaluated using a synthetic pairs of foggy bucket signals and their corresponding ground truth clear images C ( x , y ) . The architecture of our proposed bio-inspired attention-based Mechanism for Foggy-condition GI Reconstruction (AMFR-GI) is shown in Figure 3. The main part is a self-attention mechanism, bio-inspired by the brain’s ability to focus on salient features and suppress noise. This allows the model to capture long-range dependencies within the compressed measurement data, effectively “attending” to the most informative components to achieve robust image recovery under various fog density and sampling rates, thereby realizing a computational form of selective perceptual attention.
Figure 3. Schematic illustration of the proposed encoder-decoder network for reconstructing high-quality 2D images from compressed 1D bucket measurements under foggy conditions. The model integrates a specialized self-attention mechanism within the decoder to capture long-range spatial dependencies, emulating the selective attention processes of biological visual systems.

2.2. Self-Attention Mechanism

The self-attention mechanism is bio-inspired by the brain’s capacity for selective attention, which allows biological visual systems, feedback connections modulate neural activity to prioritize salient features while suppressing noise, a process essential for perceiving objects in fog. Our artificial module functionally mimics this by dynamically calculating attention scores (Equations (6) and (7)), effectively working as a top-down attentional signal that highlights informative features in the bucket signal and suppresses noise, thereby enabling the network to ‘see through’ scattering media such as fog. In our model, this is implemented computationally. The mechanism allows the network to dynamically weight the importance of all elements in the compressed measurement sequence. When processing a fog degraded bucket signal, the model learns to ‘attend to’ or prioritize components that correspond to the true object’s structure while suppressing components correlated with scattering noise. This global, content-aware feature extraction is enables the accurate recovery of structural coherence and fine details that are lost in conventional methods. The architecture comprises two components: the self-attention module, which performs feature refinement through neural-inspired attention, and the self-attention reconstructor, which progressively builds the image from compressed measurements.

2.2.1. Self-Attention Module

The self-attention module functions as an artificial neural circuit that learns complex spatial relationships across different regions of the input feature map. Bio-inspired by the top-down modulation observed in biological visual systems, this module dynamically weights the importance of different features, effectively focusing computational resources on the most informative elements while suppressing noise, a critical capability for imaging in scattering environments.
The module transforms an input feature map x R B × C × H × W (where B denotes batch size, C channels, and  H × W spatial dimensions) through three parallel 1 × 1 convolutional layers to extract query (Q), key (K), and value (V) projections:
proj query = Conv query ( x ) proj key = Conv key ( x ) proj value = Conv value ( x )
These projections undergo spatial reshaping to facilitate the computation of attention scores, which determine how much focus to place on different parts of the input:
E = proj query proj key C
where E R H W × H W represents the attention energy matrix, and  C provides scaling stabilization in high-dimensional spaces.
The attention probabilities are obtained through softmax normalization:
attention = Softmax ( E )
The final attended features are computed as a weighted combination of value projections:
out = proj value · attention
A residual connection with learnable scaling preserves gradient flow and enables adaptive feature integration:
out final = γ · out + x
where γ R is a trainable parameter initialized to zero, allowing the network to progressively learn the optimal contribution of attention. This design enables selective focus on relevant spatial locations while maintaining original feature information, effectively mimicking the brain’s ability to enhance salient features in noisy sensory input.

2.2.2. Self-Attention Reconstructor

The SelfAttentionReconstructor forms the image formation pathway of our bio-inspired architecture, transforming compressed GI measurements into high-fidelity images through a hierarchical decoding process. Following an encoder-decoder paradigm, this component maps the input measurement vector x R B × N to a compact latent representation:
x latent = FC ( x )
The latent representation is reshaped into a 4D tensor ( B , 1024 , 8 , 8 ) and progressively decoded through transposed convolutional layers that systematically increase spatial resolution while integrating self-attention modules at multiple scales:
x reconstructed = Decoder ( x latent )
The decoder architecture employs:
(a)
A cascade of transposed convolutional layers with batch normalization and Leaky ReLU activations, interspersed with self-attention modules to capture multi-scale contextual relationships;
(b)
A final convolutional layer with hyperbolic tangent activation producing normalized output in [ 1 , 1 ] .
The complete reconstruction process is defined as:
I ^ = SelfAttentionReconstructor ( M )
where I ^ R B × 1 × H × W represents the reconstructed image and M R B × N the input measurements. This formulation enables the transformation of low-dimensional compressed signals into high-resolution images through the synergistic integration of attention-guided feature processing and progressive spatial upsampling.

2.3. Training and Evaluation Framework

2.3.1. Dataset Preparation

To effectively train the bio-inspired AMFR-GI model, we generated a comprehensive dataset simulating realistic foggy conditions. Since acquiring paired clean and foggy ground-truth images is challenging, we developed a physical-based scattering model to synthesize fog effects on clear images I ( x , y ) . The fog synthesis follows atmospheric scattering theory:
I f ( x , y ) = I ( x , y ) · ( 1 D ) + F ( x , y ) · D ,
where I f ( x , y ) is the fogged image, D represents fog density which was linearly scaled with the scattering coefficient β , and  F ( x ) = e β d ( x ) defines the fog transmission map with a constant distance d ( x ) = 50 m across the scene to simulate uniform fog. The atmospheric scattering model is illustrated in Figure 4.
Figure 4. (a) Schematic diagram of atmospheric scattering model for simulating foggy environments (b) Visual impact of fog with varying scattering coefficient β .
To enhance realism, we incorporated spatially varying fog with Gaussian noise:
F ( x , y ) = D · N ( 0 ,   1 ) ,
followed by smoothing to introduce natural spatial coherence:
F s = 1 K 2 i = K 2 K 2 j = K 2 K 2 F ( x + i , y + j ) ,
The final fogged image is computed as:
I f = clamp I · ( 1 D ) + F s , 1 , 1 ,
preserving both scene content and realistic fog characteristics for training.

2.3.2. Model Evaluation

To ensure a careful evaluation and prevent overfitting, the dataset was partitioned into three distinct subsets: training, validation, and testing. The training set was used to update the model parameters, the validation set was used for hyperparameter tuning and monitoring generalization performance during training, and the final performance metrics reported in Section 3 were computed on the held-out test set, which was never used in the training process.
The proposed model was implemented in PyTorch 2.1.1 and trained in a Python 3.12 environment on an NVIDIA RTX 3060 GPU. We employed Mean Squared Error (MSE) as the primary optimization objective:
L MSE = 1 B · H · W i = 1 B j = 1 H k = 1 W I i j k I ^ i j k 2 .
The training procedure follows Algorithm 1. During each epoch, the model processes batches from the training set, computes the loss, and updates parameters via the Adam optimizer (learning rate η = 10 4 , weight decay 10 5 . Crucially, after processing the training batch, the model enters a validation phase where it evaluates the current state on the separate validation set without gradient computation. The Peak Signal-to-Noise Ratio (PSNR) and Structural Similarity Index (SSIM) are computed on this validation set for monitoring:
PSNR ( I , I ^ ) = 10 · log 10 MAX I 2 MSE ( I , I ^ ) ,
and perceptual quality was assessed via Structural Similarity Index (SSIM):
SSIM ( I , I ^ ) = l ( I , I ^ ) · c ( I , I ^ ) · s ( I , I ^ )
with component functions defined as:
l ( I , I ^ ) = 2 μ I μ I ^ + C 1 μ I 2 + μ I ^ 2 + C 1
c ( I , I ^ ) = 2 σ I σ I ^ + C 2 σ I 2 + σ I ^ 2 + C 2
s ( I , I ^ ) = σ I I ^ + C 3 σ I σ I ^ + C 3 ,
Training and validation progress for both datasets is shown in Figure 5.
Algorithm 1 Self-Attention Reconstruction Training with Validation
 Require: 
Training dataset D train with sampling ratio N [ 0 , 1 ] , scattering coefficient
β [ 1 dB / m , 5 dB / m ] and fog density d [ 0 , 1 ]
 Require: 
Validation dataset D val for monitoring
 Require: 
Model f θ with parameters θ
 Require: 
Learning rate η , batch size B, epochs T
 Ensure: 
Trained model f θ
  1:
Initialize θ randomly
  2:
for epoch = 1 to T do
  3:
    Training Phase
  4:
    for batch ( M , I ) D train  do
  5:
         I ^ f θ ( M )                                                                                                ▹ Forward pass
  6:
        Compute L MSE (Equation (17))                                                          ▹ Loss calculation
  7:
        Update θ via Adam optimizer                                                     ▹ η , weight decay 10 5
  8:
    end for
  9:
    Validation Phase
10:
    Initialize accumulators: val _ loss = 0 , val _ psnr = 0 , val _ ssim = 0 , count = 0
11:
    for batch ( M , I ) D val  do
12:
         I ^ f θ ( M )                                                                             ▹ No gradient computation
13:
        Compute batch L MSE , PSNR (Equation (18)), SSIM (Equation (19))
14:
        Accumulate: val _ loss val _ loss + L MSE
15:
        Accumulate: val _ psnr val _ psnr + PSNR
16:
        Accumulate: val _ ssim val _ ssim + SSIM
17:
         count count + 1
18:
    end for
19:
    Compute averages: val _ loss = val _ loss / count , val _ psnr = val _ psnr / count ,
val _ ssim = val _ ssim / count
20:
    Log validation metrics: val _ loss , val _ psnr , val _ ssim for epoch monitoring
21:
end for
22:
return  f θ
Figure 5. Training evaluation metrics for (a) MNIST (28 × 28 pixels) and (b) Human & Horse (64 × 64 pixels) datasets, demonstrating convergence and reconstruction performance across epochs.
With the bio-inspired AMFR-GI architecture and comprehensive foggy dataset established, the model was systematically trained and validated across varying sampling ratios and fog density. The following section presents a detailed evaluation of the model’s performance under these challenging conditions.

3. Results

To validate the efficacy of our bio-inspired computational imaging framework, we conducted comprehensive numerical simulations under varying system parameters. The following analysis demonstrates how the self-attention mechanism modeled after neural selective attention enables robust image reconstruction in challenging foggy conditions across different sampling regimes and scattering intensities.

3.1. Reconstruction Performance Under Varying Sampling Ratios

This section investigates the performance of our bio-inspired GI model under varying sampling ratios, evaluating its ability to reconstruct images from sparse measurements in foggy environments. The scattering coefficient was fixed at β = 1.0 dB / m to simulate moderate fog conditions, while the sampling ratio N was varied across { 10 % , 25 % , 50 % , 75 % , 100 % } to assess data efficiency.
Figure 6 presents comprehensive qualitative and quantitative results for both the high-resolution Human and Horse dataset (64 × 64 pixels) and the low-resolution MNIST dataset (28 × 28 pixels). The bio-inspired architecture demonstrates remarkable capability in leveraging global contextual information emulating the brain’s ability to infer complete structures from partial data. At a sampling ratio of 25%, the Human and Horse dataset achieves near-complete reconstruction despite the scattering environment, with the self-attention mechanism effectively filling in missing spatial information. This mirrors biological vision systems that excel at perceiving coherent objects from fragmented visual cues in poor visibility.
Figure 6. Performance evaluation under varying sampling ratios demonstrates the bio-inspired model’s data efficiency: (a) Human and Horse reconstructions show robust recovery even at 25% sampling, (b) MNIST reconstructions require higher sampling for complete recovery, (c) PSNR improvement with increased sampling, and (d) SSIM trends reflecting perceptual quality maintenance.
Conversely, the MNIST dataset exhibits incomplete recovery at the same sampling ratio, reflecting the fundamental limitations of low-resolution data under scattering conditions. However, as sampling increases to 50%, both datasets achieve complete reconstruction, with the Human and Horse dataset maintaining superior visibility due to its richer structural information. The significant degradation observed at 10% sampling highlights the challenge of extreme undersampling in scattering media, where even sophisticated attention mechanisms struggle with insufficient signal content.
Quantitative metrics in Figure 6c,d reveal asymptotic improvement in both PSNR and SSIM with increasing sampling ratios, validating the model’s data-efficient learning capability. The Human and Horse dataset achieves PSNR ≈ 24.5 dB and SSIM ≈ 0.8 at higher sampling ratios, demonstrating robust reconstruction quality. The MNIST dataset shows lower baseline performance but significant improvement with increased sampling, indicating the model’s adaptive capability across resolution domains.
These findings underscore the advantage of our bio-inspired approach: by emulating the brain’s contextual processing, the model achieves substantial reconstruction quality at lower sampling ratios, enabling energy-efficient and time-constrained sensing systems for practical deployment in challenging environmental conditions.

3.2. Robustness to Progressive Scattering Effects

To evaluate the model’s resilience to increasingly adverse conditions, we systematically tested its performance under varying scattering coefficients while maintaining full sampling ( N = 100 % ). The scattering coefficient β was varied across { 0.1 dB / m , 0.5 dB / m , 1.0 dB / m , 2.0 dB / m , 3.0 dB / m , 5.0 dB / m } to emulate progressive fog density encountered in remote sensing applications.
Figure 7 presents the qualitative and quantitative results, revealing the model’s remarkable robustness to scattering-induced degradation. The bio-inspired attention mechanism demonstrates a capacity akin to biological vision in maintaining perceptual coherence despite signal deterioration. For the Human and Horse dataset (Figure 7a), reconstructed images maintain structural integrity and key features even at extreme scattering ( β = 5.0 dB / m ), with only gradual quality degradation. This resilience stems from the attention mechanism’s ability to prioritize and preserve semantically important features while suppressing noise, directly mirroring the neural processes that enable biological systems to see through fog and haze.
Figure 7. Bio-inspired model demonstrates scattering resilience: (a) Human and Horse reconstructions maintain structural coherence even at β = 5.0 dB / m , (b) MNIST shows progressive degradation with increasing scattering, (c) PSNR stability across scattering levels, and (d) SSIM preservation highlighting perceptual robustness.
The MNIST dataset (Figure 7b) exhibits more pronounced degradation with increasing β , reflecting the inherent vulnerability of low-resolution, structurally simple images to scattering effects. The limited feature diversity in MNIST digits provides fewer robust cues for the attention mechanism to leverage under severe degradation, highlighting the interplay between image complexity and reconstruction robustness.
Quantitative analysis in Figure 7c,d confirms these observations: PSNR values show a general decline with increasing β , reflecting the fundamental signal-to-noise ratio challenge in dense fog. However, the Human and Horse dataset maintains significantly higher PSNR across all scattering levels, demonstrating the model’s superior handling of complex scenes. Similarly, the SSIM values show a progressive decline, but the high-resolution dataset maintains a relatively stable perceptual quality, whereas MNIST experiences a sharp drop, particularly beyond β = 2.0 dB / m .
This scattering resilience demonstrates the practical value of our bio-inspired approach for real-world remote sensing, where atmospheric conditions can vary dramatically. The model’s ability to maintain reconstruction quality across a wide scattering range positions it as a robust solution for environmental monitoring, security surveillance, and defense applications operating in degraded visual environments.

3.3. Comparative Analysis

To objectively assess the performance of our proposed bio-inspired model (AMFR-GI), we conducted a qualitative and quantitative comparison against several established benchmark methods as illustrated in Figure 8 and Table 1, respectively. These included conventional computational GI (CGI), compressive sensing-based CGI (CSCGI), convolutional neural network-based CGI (CNN-CGI), and generative adversarial network-based CGI (GAN-CGI). The evaluation was performed under consistent scattering conditions β = 1.0 dB / m at full sampling ( N = 100 % ) using the MNIST dataset. As illustrated in Figure 8, visual inspection reveals that the proposed AMFR-GI method more effectively recovers structural details, enhances image contrast, and mitigates scattering-induced noise, demonstrating superior reconstruction quality compared to both traditional and DL-based GI approaches.
Figure 8. Comparative reconstruction performance under scattering conditions. The top row displays the target object from the (left) MNIST and (right) scattering effect object. The second row shows the reconstructions using CGI, CSCGI, CNN-CGI, and GAN-CGI methods. Visual comparison demonstrates that the proposed method effectively recovers structural details, enhances contrast, and suppresses scattering-induced noise, outperforming the traditional GI approach.
Table 1. Quantitative Comparative analysis under scattering conditions.

3.4. Computational Efficiency and Performance Analysis

The proposed self-attention-based framework for GI demonstrate significant advancements in reconstruction quality, particularly under adverse foggy conditions. However, its computational efficiency and performance on low-resolution datasets, such as MNIST (28 × 28 pixels), warrant detailed examination to ensure scalability and applicability across diverse imaging scenarios.

3.4.1. Computational Efficiency of the Bio-Inspired Architecture

The self-attention mechanism, core to our bio-inspired architecture, introduces a computational overhead that scales with input resolution, a known trade-off for its superior capacity to model global context. To quantify this, we evaluated inference time and memory usage on an NVIDIA RTX 3060 GPU. As summarized in Table 2, the model requires approximately 0.12 s per inference for the high-resolution Human and Horse dataset (64 × 64 pixels) and only 0.08 s for the MNIST dataset (28 × 28 pixels), with same GPU memory usage. These metrics confirm that the model’s computational cost is well within the bounds for real-time remote sensing applications, as the inference time remains orders of magnitude below the typical 1-s threshold for such tasks [10]. The efficiency of this bio-inspired approach demonstrates that the benefits of global contextual processing can be achieved without prohibitive computational cost.
Table 2. Computational performance and reconstruction quality across different datasets.

3.4.2. Performance Analysis Across Resolutions

The model’s performance exhibits a strong dependency on input complexity, a characteristic shared with biological visual systems that excel at parsing rich scenes. On low-resolution, structurally simple datasets like MNIST, performance at very low sampling ratios ( N = 10 % ) is limited (PSNR ≈ 13.2 dB, SSIM ≈ 0.5), as the sparse measurements provide insufficient features for the attention mechanism to resolve. In stark contrast, the high-resolution Human and Horse dataset achieves near-complete reconstruction at N = 25 % (PSNR ≈ 20 dB, SSIM ≈ 0.57), underscoring the model’s strength in leveraging complex structural information. This dichotomy mirrors the principle that biological perception is most robust when sensory input is information-rich, highlighting that our bio-inspired model’s superiority is most pronounced in complex, real-world scenarios.

4. Discussion

The comprehensive results presented in this study demonstrate a clear performance difference between simple and complex datasets. For the structurally rich Human-Horse dataset, the self-attention mechanism effectively leverages the abundant global context, achieving robust reconstruction (PSNR ≥ 23.5 dB, SSIM ≥ 0.7) even at a low sampling ratio ( N 50 % ) and under high scattering. Conversely, the MNIST dataset, with its low resolution and simple shapes, provides fewer robust features for the attention mechanism to resolve, leading to poorer performance under aggressive undersampling ( N 50 % ) or extreme scattering. This indicates that the advantage of our bio-inspired approach is most pronounced for complex, real-world scenes, mirroring biological vision which excels in information-rich environments. The methodology positively affects sampling ratios by enabling high-fidelity reconstruction with fewer measurements, as the global context allows it to infer missing information. Self-attention-based GI model achieves robust high-fidelity reconstruction across a wide spectrum of foggy conditions, characterized by scattering coefficients from 0.1 dB/m to 5.0 dB/m and sampling ratios from 10% to 100%. This discussion contextualizes these findings by comparing our model’s performance against state-of-the-art methodologies, elucidating its unique contributions and inherent limitations for computational imaging in remote sensing. However, current limitation of our work is, the model trained and evaluated using a fixed basis of speckle illumination patterns and synthetic fog model based on atmospheric physics (Equations (13)–(16)). While this model is widely used in computational imaging literature [17,18] and provides a controlled environment for testing, its absolute fidelity to real-world fog conditions is a subject for further validation.
Comparison with Traditional and CNN-based GI: The proposed model’s ability to maintain reconstruction quality under high scattering and low sampling distinguishes it fundamentally from traditional GI approaches [19]. For instance, compressive sensing GI [20] achieves reasonable reconstruction with reduced measurements but relies on sparse representations that fail under significant scattering-induced noise. Similarly, while CNN-based methods [21] improve upon traditional techniques, their localized receptive fields limit their ability to integrate global context, leading to pronounced degradation at low sampling ratios ( N 25 % ). In contrast, the self-attention mechanism of our model, inspired by the brain’s capacity for global feature integration, leverages long-range dependencies within the measurement data. This enables superior performance in high-scattering environments, as evidenced by the high PSNR (≈24.5 dB) and SSIM (≈0.8) achieved for the Human and Horse datasets at β = 3.0 dB/m (Figure 7). The primary trade-off in our methodology is the increased computational complexity of the self-attention mechanism compared to simpler CNN architectures. This is a known cost for gaining the ability to model global context. However, as quantified in Table 2 and Figure 9, this cost is manageable and does not preclude real-time application.
Figure 9. Average reconstruction time (sec) for MNIST ( 28 × 28 ) and Human and Horse ( 64 × 64 ) resolutions under varying sampling ratios at fog density β = 21.71 dB and Gaussian noise level 10 dB . Higher resolution and increased sampling ratios result in longer reconstruction times.
Advancements Over Contemporary GI in Scattering Media: Recent work by Jiang et al. [22] on GI in atmospheric turbulence achieved improved reconstruction through adaptive correlation but reported lower PSNR values (≈15 dB) under high scattering compared to our model. Our approach demonstrates that a dedicated, learnable attention mechanism is more effective than post-hoc correlation adjustments in mitigating the severe effects of fog. Furthermore, while transformer-based models such as TransUNet [23] excel in domains such as medical image segmentation, their direct application to GI reconstruction in fog remains unexplored and is likely suboptimal because of their task-agnostic design. Our architecture is specifically tailored for the GI reconstruction task, optimizing the self-attention mechanism to directly address scattering-induced degradation, which is confirmed by its higher PSNR and SSIM across varying scattering coefficients (Figure 6c and Figure 7c). The limitation of existing GI methods in scattering media are shown in Table 3.
Table 3. Limitations of existing GI reconstruction methods in scattering media.
The Bio-Inspired Advantage: The core of the robustness of our model stems from its bio-inspired design. The self-attention mechanism functionally mimics the selective context integration properties of biological visual systems. This allows the model to prioritize salient features and suppress noise, which is particularly advantageous when reconstructing complex scenes from highly degraded or undersampled data. This principled approach moves beyond mere architectural choice and establishes a new paradigm for bio-inspired computational imaging, offering a significant step toward robust remote sensing under real-world adverse conditions.

5. Conclusions

This study has presented a transformative, bio-inspired computational imaging framework that significantly advances the robustness of GI for remote sensing in foggy conditions. By integrating a self-attention mechanism explicitly modeled after the brain’s capacity for selective perceptual processing, the proposed model effectively captures both local details and long-range contextual dependencies within highly compressed, one-dimensional bucket signals. This neural-inspired approach enables high-fidelity image reconstruction across a wide spectrum of scattering coefficients ( β [ 0.1 dB / m , 5.0 dB / m ] ) and sampling ratios ( N [ 10 % , 100 % ] ). The main results confirm the following: (1) Our model achieves high-fidelity reconstruction (PSNR 25 dB, SSIM 0.8) under severe scattering ( β 3.0 dB / m ) and moderate sampling ( N = 50%). (2) It outperforms conventional and CNN-based GI methods, particularly in global structure recovery. (3) An ablation study confirms the self-attention module achieved high quality reconstruction. (4) The framework is computationally efficient for real-time use.
The proposed bio-inspired GI framework can be directly integrated with active remote sensing systems such as LIDAR or airborne surveillance platforms. By replacing their conventional reconstruction algorithms with our AMFR-GI model, these systems could achieve significantly enhanced perception in fog, haze, and other obscurants. For real-world deployment, the next step involves training the model on large-scale, real paired data collected from these active systems operating in diverse weather conditions. These advancements position our bio-inspired computational imaging framework as a powerful tool for critical remote sensing applications, including environmental monitoring, security surveillance, autonomous navigation and defense operations under adverse atmospheric conditions.
Current limitations, such as the model’s reliance on a fixed illumination basis and the use of synthetic training data, highlight clear avenues for future work. Exploring hybrid architectures that combine the global context of self-attention with the spatial efficiency of lightweight convolutional layers promises to further optimize this paradigm. Extending the model to be pattern-agnostic and validating it with real-world GI data will be essential to bridge the simulation-to-reality gap. Such developments will continue to harness the principles of biological vision, ultimately creating computational imaging systems that see through obscurants with unprecedented resilience and efficiency.

Author Contributions

R.I.: Writing—review & editing, Writing—original draft, Validation, Software, Project administration, Methodology, Investigation, Formal analysis, Data curation, Conceptualization, Y.S.: Writing—review & editing, Visualization, Validation, K.Z.: review, L.D., D.T., Y.W., P.W.: Writing—review & editing, Validation, Formal analysis, project administration. J.C.: Writing—review & editing, Visualization, Validation, Supervision, Resources, Funding acquisition, Formal analysis. All authors have read and agreed to the published version of the manuscript.

Funding

This work was supported by the Strategic Priority Research Program of the Chinese Academy of Sciences, Grant No. XDB1050000, National Natural Science Foundation of China (Grant No. 62275022) and Beijing Natural Science Foundation (Grant No. 4232014).

Institutional Review Board Statement

Not applicable.

Data Availability Statement

Data underlying the results presented in this paper are not publicly available at this time but may be obtained from the authors upon reasonable request.

Conflicts of Interest

The author Peng Wang is employed by Beijing Aerospace Measurement and Control Technology Co., Ltd., Beijing, 100041, China., which is registered as a limited company. The other authors declare no conflicts of interest related to this work.

References

  1. Pittman, T.B.; Shih, Y.; Strekalov, D.; Sergienko, A.V. Optical imaging by means of two-photon quantum entanglement. Phys. Rev. A 1995, 52, R3429. [Google Scholar] [CrossRef]
  2. Soleimani, P.; Kheradmand, R.; Kandjani, S.A. The Study of Computational Ghost Imaging in Presence of Atmospheric Turbulence. Optica 2024. [Google Scholar] [CrossRef]
  3. Land, M.F.; Osorio, D. Extraordinary color vision. Science 2014, 343, 381–382. [Google Scholar] [CrossRef]
  4. Barber, J.R.; Conner, W.E. Acoustic mimicry in a predator–prey interaction. Proc. Natl. Acad. Sci. USA 2007, 104, 9331–9334. [Google Scholar] [CrossRef]
  5. Reynolds, J.H.; Heeger, D.J. The normalization model of attention. Neuron 2009, 61, 168–185. [Google Scholar] [CrossRef] [PubMed]
  6. He, L.; Wei, B.; Hao, K.; Gao, L.; Peng, C. Bio-inspired deep neural local acuity and focus learning for visual image recognition. Neural Netw. 2025, 181, 106712. [Google Scholar] [CrossRef] [PubMed]
  7. Li, G.; Yang, Z.; Zhao, Y.; Yan, R.; Liu, X.; Liu, B. Normalized iterative denoising ghost imaging based on the adaptive threshold. Laser Phys. Lett. 2017, 14, 025207. [Google Scholar] [CrossRef]
  8. Yang, C.; Wang, C.; Guan, J.; Zhang, C.; Guo, S.; Gong, W.; Gao, F. Scalar-matrix-structured ghost imaging. Photonics Res. 2016, 4, 281–285. [Google Scholar]
  9. O-oka, Y.; Fukatsu, S. Differential ghost imaging in time domain. Appl. Phys. Lett. 2017, 111, 061106. [Google Scholar] [CrossRef]
  10. Wu, H.; Wang, R.; Zhao, G.; Xiao, H.; Wang, D.; Liang, J.; Tian, X.; Cheng, L.; Zhang, X. Sub-Nyquist computational ghost imaging with deep learning. Opt. Express 2020, 28, 3846–3853. [Google Scholar] [CrossRef]
  11. LeCun, Y.; Bengio, Y.; Hinton, G. Deep learning. Nature 2015, 521, 436–444. [Google Scholar] [CrossRef] [PubMed]
  12. Krizhevsky, A.; Sutskever, I.; Hinton, G.E. Imagenet classification with deep convolutional neural networks. In Advances in Neural Information Processing Systems; Curran Associates, Inc.: Red Hook, NY, USA, 2012; Volume 25. [Google Scholar]
  13. Wu, Y.; Ma, Y.; Liu, J.; Du, J.; Xing, L. Self-attention convolutional neural network for improved MR image reconstruction. Inf. Sci. 2019, 490, 317–328. [Google Scholar] [CrossRef]
  14. Cui, Z.; Li, K.; Gu, L.; Su, S.; Gao, P.; Jiang, Z.; Qiao, Y.; Harada, T. You only need 90k parameters to adapt light: A light weight transformer for image enhancement and exposure correction. arXiv 2022, arXiv:2205.14871. [Google Scholar] [CrossRef]
  15. Xu, X.; Wang, R.; Fu, C.W.; Jia, J. Snr-aware low-light image enhancement. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, New Orleans, LA, USA, 18–24 June 2022; pp. 17714–17724. [Google Scholar]
  16. Liang, J.; Jin, Y.; Chen, X.; Huang, H.; Deng, Y. SVTSR: Image super-resolution using scattering vision transformer. Sci. Rep. 2024, 14, 31770. [Google Scholar] [CrossRef]
  17. Lin, H.; Luo, C. Demonstration of computational ghost imaging through fog. Opt. Laser Technol. 2025, 182, 112075. [Google Scholar]
  18. Iqbal, R.; Jie, C.; Kai, L.; Hao, Q. Ghost imaging for remote sensing: Experimental demonstration utilizing humidifier-simulated fog. Opt. Commun. 2025, 596, 132508. [Google Scholar] [CrossRef]
  19. Iqbal, R.; Jie, C.; Abbas, A.; Lin, K.; Hao, Q. Fourier domain ghost imaging with adaptive enhancement for scattering media. Appl. Opt. 2025, 64, 9553–9563. [Google Scholar] [CrossRef]
  20. Li, F.; Zhao, M.; Tian, Z.; Willomitzer, F.; Cossairt, O. Compressive ghost imaging through scattering media with deep learning. Opt. Express 2020, 28, 17395–17408. [Google Scholar] [CrossRef] [PubMed]
  21. Lyu, M.; Wang, W.; Wang, H.; Wang, H.; Li, G.; Chen, N.; Situ, G. Deep-learning-based ghost imaging. Sci. Rep. 2017, 7, 17865. [Google Scholar] [CrossRef]
  22. Jiang, C.; Xu, B.; Zhang, L.; Zhang, D. Restoration of Ghost Imaging in Atmospheric Turbulence Based on Deep Learning. Curr. Opt. Photonics 2023, 7, 655–664. [Google Scholar]
  23. Chen, J.; Mei, J.; Li, X.; Lu, Y.; Yu, Q.; Wei, Q.; Luo, X.; Xie, Y.; Adeli, E.; Wang, Y.; et al. TransUNet: Rethinking the U-Net architecture design for medical image segmentation through the lens of transformers. Med. Image Anal. 2024, 97, 103280. [Google Scholar] [PubMed]
  24. Bashkansky, M.; Park, S.D.; Reintjes, J. Single pixel structured imaging through fog. Appl. Opt. 2021, 60, 4793–4797. [Google Scholar] [CrossRef]
  25. Yu, Z.; Li, X.; Bai, X.; Wang, Y.; Chen, X.; Liu, Y.; Sun, M.; Zhou, X. Computational ghost imaging through a dynamic scattering medium based on a convolutional neural network from simulation. Laser Phys. Lett. 2023, 20, 055204. [Google Scholar] [CrossRef]
  26. Chen, P.; Li, Z.; Bing, P.; Song, Z.; Zhou, X.; Yuan, S. Improved transformer-based ghost imaging in underwater scattering medium. Opt. Lasers Eng. 2026, 196, 109427. [Google Scholar]
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Article Metrics

Citations

Article Access Statistics

Multiple requests from the same IP address are counted as one view.