Next Article in Journal
Research on Open-Set Recognition Methods for Rolling Bearing Fault Diagnosis
Previous Article in Journal
Recent Advancement in Fluorescent Probes for Peroxynitrite (ONOO)
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Detection and Recognition of Visual Geons Based on Specific Object-of-Interest Imaging Technology

School of Electronic Science and Engineering (School of Microelectronics), South China Normal University, Foshan 528000, China
*
Author to whom correspondence should be addressed.
Sensors 2025, 25(10), 3022; https://doi.org/10.3390/s25103022
Submission received: 14 March 2025 / Revised: 6 May 2025 / Accepted: 8 May 2025 / Published: 10 May 2025
(This article belongs to the Section Sensing and Imaging)

Abstract

:
Across domains such as visual processing, computer graphics, neuroscience, and biological sciences, geons are recognized as fundamental components of complex shapes. Their theoretical significance has been extensively acknowledged in scientific research. However, accurately identifying and extracting these structural components remains a persistent challenge. This study integrates theoretical foundations from signal processing, computer graphics, neuroscience, and biological sciences. We employ specific object-of-interest imaging and neural networks to mathematically operationalize visual geon characterization, thereby elucidating their intrinsic properties. Experiments validate the core hypothesis of geon theory, namely that geons are foundational components for the visual system to recognize complex objects. Through training, neural networks are capable of identifying distinct basic geons and, on this basis, performing target recognition in more complex scenarios. These findings provide empirical confirmation of geons’ existence and their critical role in visual recognition, establishing novel computational paradigms and theoretical foundations for interdisciplinary research.

1. Introduction

The generation and interpretation of complex geometric structures follow hierarchical construction principles, in which a finite set of core topological elements is combined to form higher-order visual representations through specific combinatorial rules. Geons are widely recognized by researchers as fundamental units underlying all visual–graphic representations, also referred to as geometric primitives or simply primitives. Since the establishment of their theoretical foundation in Marr’s computational vision framework, formal representations of geons have progressed significantly in computer graphics, cognitive neuroscience, and biological sciences, highlighting the pivotal role of geon theory in explainable artificial intelligence and bioinspired computing.

1.1. Recognition-by-Components (RBCs) Framework Based on Marr

In the 1980s, Biederman et al. proposed the RBCs theory [1,2]. This theory systematically explains the conceptual framework of geons by proposing that the human visual system recognizes objects via a limited set of basic volumetric geons and their spatial relationships. These geons are characterized by distinct volumetric properties. The parallel work of Marr’s visual processing theory posited that image perception begins with pixel-level analysis, followed by hierarchical extraction of low-level features—such as edges, line orientations, and luminance variations—to generate primitives analogous to geons [3,4,5,6,7,8,9,10,11,12,13,14,15]. Marr’s multi-level visual representation model detects and combines these geons to construct complex geometric forms, enabling a comprehensive understanding of the scene. This model emphasizes the optimized information flow at each processing level under specific rules. It generates a “raw primal sketch”, which serves as the foundation for higher-level visual tasks such as stereo matching, depth estimation, and motion detection. Marr also proposed the use of mathematical approaches and computational models to extract these basic geons from images. This work simulated the mechanisms of the human visual system for 3D object recognition and led to the development of a three-level framework: computational theory–algorithm–implementation:
  • Computational level: Geons as the minimal units of visual representation;
  • Algorithmic level: Deep learning simulates hierarchical feature extraction;
  • Implementation level: The optimization of weight in neural networks ( θ , ξ ) corresponds to synaptic plasticity.
Later, Tomaso Poggio et al. regarded “geons” as the theoretical foundation of visual processing and object recognition. They investigated how the visual cortex recognizes shapes and objects by progressively abstracting image features [16,17,18,19,20,21,22]. In Perceptrons, Marvin Minsky et al. explored pattern recognition and the idea of decomposing complex shapes into simple elements. They emphasized the importance of identifying visual components necessary for successful object recognition in artificial intelligence and later developed a hierarchical processing framework that treats geons as fundamental cognitive units [23,24,25,26,27,28]. The core idea of RBCs theory is that object recognition is achieved by decomposing objects into basic components and analyzing their spatial relationships. This process is known as componential analysis. RBCs is based on three key principles:
  • Constructing complex objects: By combining and arranging different basic geons, various complex visual objects can be created, offering high flexibility.
  • Providing recognition cues: Basic geons not only form the shape and structure of objects, but also provide key cues for object recognition. When an observer sees an object, they first identify the basic geons of the object and then recognize the object based on the combination and arrangement of these elements.
  • Supporting rapid recognition: Due to the limited number and types of basic geons and the relatively fixed ways they can be combined, the visual system can quickly recognize objects, making this process highly automated.
However, scholars within the RBCs framework have retrospectively identified several core limitations in the framework: First, the representational system of visual geons suffers from conceptual ambiguity. Key aspects such as feature detection mechanisms and topological structure analysis remain poorly defined. Second, the model does not specify how the visual system dynamically balances bottom-up and top-down processing. Most critically, the theory lacks a unified computational framework for multi-modal perceptual integration, which restricts the hierarchical integration of perceptual information. This absence of a system-level architecture ultimately constrains the emergence of integrated visual perception representations. Such critical reflection reveals fundamental limitations in early computational vision theories, particularly regarding neural interpretability and cognitive integration mechanisms.

1.2. Deep Learning-Based CV Scholars

Since the early 21st century, research in visual computing has shifted from traditional theoretical exploration to the development of practical deep learning-based techniques. Yan et al. proposed the HPNet model, which integrates pixel-level semantic segmentation with geon shape decomposition for end-to-end, high-precision geon segmentation [29]. For 3D shape generation, Zou et al. developed 3D-PRNN—an encoder–decoder architecture augmented with a recurrent neural network—effectively addressing traditional methods’ limitations in capturing shape diversity and structural detail [30]. Lin et al. combined synthetic data and deep neural networks to construct a parametric representation of object structures, successfully applying this approach to robotic grasping tasks [31]. In graphics generation, Aliev et al. introduced a neural point-based model representing geometric shapes through point clouds, reducing computational cost while maintaining high output quality [32]. Zhang et al. developed H3DNet, which combines hybrid geon representations with a multi-branch network architecture to enable precise parsing and reconstruction of LiDAR point cloud data [33]. Li et al. demonstrated that geons are central to sketch segmentation, with their model identifying, segmenting, and automatically annotating these primitives [34]. Huang et al. proposed DeepPrimitive, a method that decomposes images into geons for component recognition and segmentation [35]. Sharma et al. introduced a neural shape parser framework for automatic parsing and reconstruction of complex 3D shapes based on geon elements [36]. Collectively, these studies advance geon-based methods in computer vision and graphics, offering novel solutions for complex scene understanding and shape reconstruction. Recent U-Net variants demonstrate potential for improved geon processing—Sharp U-Net [37] employs depthwise separable convolutions to enhance edge preservation through learned sharpening filters, achieving 12% better boundary recall on medical shapes compared with baseline U-Net. While promising for geon extraction, we retain the original U-Net architecture to (1) maintain direct comparability with Marr’s biological vision analogies, (2) avoid task-specific optimizations that might obscure fundamental geon detection principles, and (3) ensure consistency with established neurocomputational interpretations.
Despite significant advances achieved by deep learning-based methods in the evolution of visual geon theory, several fundamental problems remain unresolved and warrant further investigation. Marr, in his hierarchical processing framework, warned that neglecting the organizational principles of the biological visual system could mislead researchers into focusing on pseudo-problems—those arising from sensor limitations, hardware constraints, or computational bottlenecks—rather than the essence of visual cognition. Despite decades of research, both theory and practice still face major challenges rooted in two fundamental issues: First, defining the segmentation target precisely—either in images or in the physical world—is intrinsically difficult. The philosophical ambiguity surrounding “objecthood” complicates the semantic separation between entity and background. Second, current segmentation methods typically rely on specific prior assumptions and constraint conditions. Their inductive biases differ fundamentally from the adaptive representational mechanisms of biological vision, which limits the generalizability of these technologies.

1.3. Psychology, Neuroscience, and Other Fields

In psychology and neuroscience, significant advancements have elucidated the role of geons in visual cognition. According to Feature Integration Theory (FIT), visual processing begins with rapid, parallel extraction of basic features followed by their binding into coherent objects via spatial attention, with neurons in areas V1–V4 mediating low-level feature extraction and higher-order integration [38,39,40,41,42,43,44,45,46,47,48,49]. These findings establish geons as fundamental units in early visual processing and support the object-imaging methods employed in this study. Building on FIT, Itti et al. developed a computational saliency model incorporating center-surround difference algorithms, cross-channel normalization, and inhibition of return, demonstrating that geon-based saliency maps predict attentional shifts and reflect independent feature channel representations in V1 [50,51,52,53,54,55,56,57].
Recent work at South China Normal University has further advanced geon research. Specifically, Hang et al. identified visual geons as foundational units of visual encoding and processing. They emphasized their role in spatial information integration and early-stage visual perception [58]. Yuxuan et al. demonstrated their role in object recognition, spatial localization, and motion perception [59]. Wen et al. reported increased task-related network connectivity in geon-processing regions [60]. Zhou et al. and Zhang et al. revealed geons’ involvement in facial perception and vergence planning [58,61]. Zhu et al. demonstrated their necessity for stereoscopic depth perception [62]. Collectively, these studies underscore geons’ central role in spatial integration, early visual processing, and higher-level cognition.
As shown in Table 1, although existing studies have confirmed the cross-disciplinary significance of visual geons, three critical limitations persist: First, theoretical frameworks fail to establish mathematical models that reconcile cognitive plausibility with computational feasibility. Second, methodological approaches excessively depend on hypothetical deduction and simulated experiments without empirical substantiation. Third, current CV segmentation techniques remain constrained by scene specificity and preset conditions. Significantly, while RBCs theory’s engineering applications are limited by inadequate dynamic adaptation mechanisms, deep learning—despite its breakthroughs in representation learning—struggles with the inductive bias dilemma identified by the Yan School. Ultimately, these contradictions compel us to revisit the neural computational essence of Treisman’s feature integration theory.

1.4. Research Objectives and Contributions

This study addresses the challenges of detecting geons in noisy, multi-object environments by integrating RBCs theory with deep learning. In practical scenarios, environmental noise and multiple-object scenes often degrade geon detection accuracy. For example, Gaussian and salt-and-pepper noise can disrupt feature extraction, and multiple-object scenes amplify detection complexity. Consequently, developing a robust geon detection method for these conditions holds both theoretical and practical significance. Specifically, we address three interrelated challenges:
  • Constructing a dynamically adaptive geon representation system to overcome RBCs theory’s rigid classification constraints.
  • Leveraging neural network plasticity to model hierarchical biological vision processing.
  • Establishing a noise-robust evaluation framework to assess the model’s biological plausibility.
To tackle these challenges, we propose a task-specific object imaging approach that combines object-specific imaging with a U-Net architecture to form a novel computational framework. Unlike conventional segmentation, this method retains input–output dimensional consistency, emulating the biological vision system. Specifically, the network adaptively learns to detect geons without relying on explicit segmentation steps. This approach improves detection efficiency and enhances performance in noisy, multi-object scenarios, aligning more closely with biological visual processing. The innovations of this study include the following:
  • Theoretical innovation: Introduction of an adaptive geon detection framework integrating RBCs theory with neural networks, bridging theory and practice for novel detection paradigms.
  • Experimental validation: Empirical demonstration of geons as foundational units in multi-object scenes, confirming their essential role in complex visual tasks.
  • Robustness analysis: Evaluation under Gaussian and salt-and-pepper noise demonstrates the method’s stability and reliability, supporting real-world applications.

2. Materials and Methods

2.1. Principles of the Experiment

FIT demonstrates the biological plausibility of contextual suppression and geon extraction, providing a theoretical basis for feature-binding mechanisms in perceptual systems. This mechanism enables the human visual system to isolate and extract target geons from complex backgrounds, enhancing perceptual efficiency and supporting contextual suppression and object representation. RBCs theory posits that geons, as independent visual units, must be distinctly segregated from complex backgrounds to ensure accurate object recognition. These theories underpin the conceptual framework of this study. Building on these foundations, this study incorporates attention mechanisms to implement the “specific object-of-interest imaging” methodology introduced by Li et al. [63,64]. This approach aims to suppress irrelevant background information in images. Using target-specific labels, the neural network learns exclusively target information, ultimately achieving complete target reconstruction. This end-to-end neural network framework learns from paired training samples. Each training pair consists of a panoramic image M h and its corresponding label, N h , where h = 1 , 2 , The training process is similar to an optimization process and can be formulated as
{ ξ , θ } = argmin ξ , θ 1 H h = 1 H N h N ˜ h
N ˜ h = P θ ( O ξ ( M h ) )
Here, input M h is processed by the compression network O ξ and reconstruction network P θ to yield N h , where ξ and θ denote their respective weights. The notation | | . | | denotes the loss function measuring the approximation error between N h and N ˜ h . This method offers a novel approach to salient object detection and demonstrates broad applicability to other computer vision tasks. By extracting key target features and comparing them to hypothesized geons, this approach enables rapid validation of theoretical assumptions and model stability. This object-imaging-based geon framework enhances the robustness and adaptability of object representation, integrating biologically inspired principles with computational models.
The “specific object-of-interest imaging” methodology is now implemented to experimentally realize geon recognition and detection. Given a photographic input containing N hypothesized geons, each object is formally represented as
K i , i = 0 , 1 . . . . , N
For the graphical objects of the entire scene, it can be represented as
K = i = 0 N 1 K i
Each graphical object K i can be represented as a linear combination in an M-dimensional sparse space and projected onto this subspace. The sparse representation is achieved using the sparse basis ψ and sparse coefficients x within the sparse space, formulated as
K i = j = 1 M ψ i j x i j
The panoramic image K can then be formally expressed as
K = i = 0 N 1 K i = i = 0 N 1 j = 1 M ψ i j x i j
A simple network is employed to train on sample images and ground truth of target K to capture its complete feature profile. After training, the neural network’s parameters encapsulate the feature information of target K , termed the feature filter Φ . Φ applies sparse transformation to panoramic image K, yielding estimated salient objects formulated as
Φ K = i = 0 N 1 j = 1 M Φ ψ i j x i j = j = 1 M ψ 0 j x 0 j = K ^ 0
As shown in Figure 1, applying the feature filter Φ to other scene images allows for filtering and ultimately recognizing the assumed basic geons.
Considering that the mathematical operator of “geons” is difficult to express directly with mathematical functions, this study uses a simple universal neural network to describe this operator. By setting different input-output data pairs to train, optimize, and approximate this mathematical operator. The resulting neural network will be regarded as the desired “geons”. U-Net is selected as the foundational network architecture [65]. U-Net exhibits unique advantages in multi-scale feature fusion and processing complex spatial relationships. Preliminary experiments in our laboratory demonstrated that compared with alternative architectures (e.g., MLP [66] and KAN [67]), U-Net’s skip connections effectively integrate hierarchical features, emulating the layered processing mechanism of the human visual system. Skip connections enable direct utilization of encoder-stage low-level features during decoding, facilitating complementary integration of detailed low-level features with semantic high-level features. This mechanism achieves multi-scale feature fusion and cross-layer information transfer, proving critical for processing complex geon topologies and geometric reconstruction. Comparatively, Multi-Layer Perceptron (MLP) lacks spatial information capture capability due to its fully connected nature; and Kernel Attention Network (KAN) shows limitations in multi-scale feature representation despite its attention mechanism focus; Half-U-Net’s unilateral architecture undermines decomposition–recomposition capacity by underutilizing cross-layer connectivity. The symmetric encoder–decoder architecture of U-Net aligns perfectly with the decomposition–recomposition geon theory: The encoder pathway comprises four downsampling stages that progressively extract high-level semantic features while synchronously processing geon characteristics from pixel-level details to global structures through hierarchical feature extraction, corresponding to geon decomposition. The decoder pathway achieves precise localization via four upsampling stages, employing cross-layer fusion strategies to integrate low-level details with high-level semantics, maintaining spatial accuracy during geometric reconstruction—this pathway mirrors the recomposition process that restores spatial relationships and geometric configurations, establishing U-Net as the optimal choice for this task.
As shown in Figure 2, Corresponding to Marr’s trilevel framework, this study also constructs a “computational–algorithmic–implementational” three-level paradigm:
  • Computational level: Consistent with RBCs theory, geons serve as minimal units of visual representation.
  • Algorithmic level: Deep learning simulates hierarchical feature extraction, where U-Net’s encoder→decoder architecture corresponds to human V1→IT cortical pathways.
  • Implementational level: Neural network weight optimization ( θ , ξ ) mirrors synaptic plasticity, partially validating Zhang et al.’s visual cortical activity modulation [58].
Itti’s saliency model, grounded in biological visual attentional mechanisms, achieves salient region extraction through feature contrast computation. Compared with conventional fixed filters, this study introduces a dynamic feature filter ( ϕ ) design that exhibits theoretical congruence with Itti’s saliency framework. Distinct from the static nature of fixed filters, ϕ adaptively extracts saliency features across scales via end-to-end learning of geon characteristics, enabling precise target geon capture with concurrent background suppression, thereby resolving the “feature detection ambiguity” inherent in the RBCs framework. Furthermore, our theoretical framework incorporates biological visual noise suppression mechanisms [50,51,52,53,54,55,56,57], providing explanatory power for model performance variance under different noise types, thus achieving closer alignment with Itti’s saliency principles. Experimental comparisons between Gaussian and salt-and-pepper noise conditions revealed superior robustness under Gaussian noise, attributable to its spectral continuity better approximating natural environmental interference patterns—a characteristic congruent with biological visual perception. Subsequent experimental validation demonstrates that our U-Net-based object imaging model effectively emulates V1 cortical contrast adaptation mechanisms through cross-channel normalization, thereby enhancing noise suppression capabilities.

2.2. Datasheet

Building on Biederman’s 36-category taxonomy and Zhou et al.’s findings on non-conscious geon perception [61], we select triangles, circles, and squares as target geon categories to ensure broad applicability. We construct a custom geon dataset comprising three geometric categories with random perturbations, yielding 1100 synthetic images: 1000 for training and 100 for validation.
We establish a multi-level evaluation framework based on cognitive interpretability, organizing nine test sets hierarchically for single-geon detection, geon separation, and multi-geon detection under geometric disturbances, each containing 100 images. We then apply two noise conditions—additive Gaussian noise and salt-and-pepper noise—to each disturbance scenario, resulting in a total of 900 test images. These test sets enable systematic evaluation of single-geon recognition, multiple-geon separation, multi-target detection, and noise-robust reconstruction across varied task requirements.

2.3. Evaluation Indicators

We employ three quantitative metrics to systematically evaluate algorithm performance:
  • Structural Similarity Index Measure (SSIM) [68]
    SSIM quantifies perceptual consistency by modeling the human visual system’s multi-channel characteristics. It is defined as
    S S I M ( X , Y ) = ( 2 μ x μ y + C 1 ) ( 2 δ x y + C 2 ) ( μ x 2 μ y 2 + C 1 ) ( δ x 2 δ y 2 + C 2 )
    By comparing brightness, contrast, and structural similarity, visual perceptual consistency is quantified within the range of [0, 1].
  • Peak Signal-to-Noise Ratio (PSNR) [69]
    PSNR measures fidelity in the frequency domain and is computed as
    P S N R = 20 l o g 10 ( M A X i M S E )
    Here, M A X i represents the maximum pixel value of the image (e.g., 255 for an 8-bit image). PSNR reflects the ratio of signal power to distortion.
  • Mean Squared Error (MSE) [70]
    MSE quantifies the average squared difference between corresponding pixels:
    M S E = 1 m n i = 0 m 1 j = 0 n 1 [ I ( i , j ) K ( i , j ) ] 2
    MSE indicates optimization convergence.
Together, SSIM, PSNR, and MSE form a complementary evaluation system, quantifying perceptual similarity, distortion fidelity, and convergence. The experimental section combines these quantitative metrics with visual comparisons to ensure comprehensive performance evaluation.

2.4. Experimental Setup Implementation Details

We conducted all experiments on an AMD Ryzen 7 6800H CPU (AMD, Santa Clara, CA, USA) paired with an NVIDIA GeForce RTX 3050 Ti GPU (AMD, Santa Clara, USA). We strictly adhered to the original U-Net–based architecture to ensure objective methodological comparisons. Training used Mean Squared Error (MSE) as the loss function. Models were trained for 60 epochs with a batch size of 4, adapting the training duration to the geon recognition complexity. Optimization was performed using Adam with an initial learning rate of 0.002, selected based on geon recognition performance. All input images were standardized to 64 × 64 pixel grayscale format to ensure consistent resolution. All experiments were repeated 10 times with different random seeds. We report mean ± standard deviation unless specified. Confidence intervals were calculated using non-parametric bootstrap (1000 resamples). The following significance testing was employed:
  • Paired t-tests for within-condition comparisons.
  • ANOVA with Tukey HSD for cross-condition analysis.
  • Wilcoxon tests for non-normal distributions.
  • Type I error control used Bonferroni correction ( α = 0.05/9 = 0.0056 for 9 test sets).

3. Experiments and Results

3.1. Fundamental Detection

The training process employing the aforementioned methodology yielded diverse geon imaging recognition outcomes, with representative results partially displayed below. All metrics were calculated from 10 independent training runs. Confidence intervals (95% CI) were derived using bootstrap resampling (Sample size: 1000).
As shown in Figure 3 and Table 2, all (* p < 0.05, ** p < 0.01 (paired t-test vs. random baseline; n = 100 images per category)) geon classes achieved SSIM values above 0.93, with circles reaching 0.99, indicating precise structural reconstruction. PSNR values between 54.64 dB and 59.14 dB indicate low reconstruction noise, and MSE values below 0.23 confirm minimal reconstruction error. These complementary metrics substantiate the framework’s high precision and stability in geon detection. Furthermore, the results corroborate Marr’s theoretical proposition that basic geons facilitate efficient visual recognition at neural–computational levels.
These tripartite metrics collectively validate the high precision of our geon extraction approach. The experimental findings resonate with Marr’s geon theory at the neural–computational level, where the feature extraction and combinatorial capacity of basic geons enable efficient visual object recognition.

3.2. Geon Separation Detection

We extended the investigation to geon separation tasks, with representative results shown below. Statistical significance was evaluated through repeated-measures ANOVA with Bonferroni correction. Error bounds represent the standard error of the mean (SEM).
As shown in Figure 4 and Table 3, all (*** p < 0.001, ** p < 0.01 (vs. single-object detection baseline) *) geon classes achieved SSIM values above 0.80, demonstrating accurate separation in complex scenes. In circular separation, SSIM reached 0.99 and MSE was 0.06, indicating near-perfect structural preservation and noise resilience. Triangle and square separation yielded PSNR above 58 dB and MSE below 0.10, indicating minimal feature degradation. These metrics confirm the framework’s ability to accurately extract fundamental geon features in composite images. Moreover, these findings align with Marr’s geon theory, validating that deconstructing objects into basic geons supports effective visual recognition.

3.3. Multi-Geon Detection

To assess the network’s generalization to multi-geon scenarios, we evaluated its performance on multi-object detection tasks. Representative results are presented (*** p < 0.001, ** p < 0.01 (vs. single-object detection baseline) *) below.
As shown in Figure 5 and Table 4, all geon classes achieved robust SSIM values—0.79 for triangles, 0.98 for circles, and 0.77 for squares—demonstrating stable feature extraction in complex scenes. Circular geons achieved the highest SSIM (0.98) and PSNR (48.58 dB), indicating superior generalization and noise resilience. Triangle and square geons obtained PSNR values of 46.17 dB and 51.55 dB with moderate MSE, indicating maintained contour integrity despite performance differences. These observations align with Treisman’s attentional bottleneck theory, suggesting that integration limits affect the detection of less regular geons.

3.4. Robustness Evaluation

We evaluated robustness by systematically testing geon detection under Gaussian and salt-and-pepper noise conditions. We introduced Gaussian noise ( σ = 30–35) for triangle and circle detection tasks and σ = 17 for square detection and separation tasks. Representative results under Gaussian noise are shown below. We performed Wilcoxon signed-rank tests to compare noise conditions. Confidence intervals (99%) were computed via the percentile method.
As shown in Figure 6, Figure 7 and Figure 8 and Table 5, these results demonstrate that the framework’s high-dimensional feature representations effectively separate geon structures from noise, maintaining low sensitivity to Gaussian disturbances. Notably, circles retain an SSIM of 0.93 under strong Gaussian noise, confirming the superior noise resilience of circular geon encoding.
We applied salt-and-pepper noise at a density of 0.05 for the best-performing tasks and 0.02 for all other scenarios. Representative results under salt-and-pepper noise are shown below.
As shown in Figure 9, Figure 10 and Figure 11 and Table 6, overall, salt-and-pepper noise results confirm robust performance, with circles exhibiting the highest SSIM and benefiting from their symmetry and smooth boundaries to mitigate noise effects.

3.5. Discussion of Experimental Results

The comprehensive experiments demonstrate that our U-Net–based geon detection framework achieves high accuracy and reliability across recognition, separation, multi-object detection, and noisy scenarios, evidencing strong generalization capabilities and robustness. It should be noted that while this study primarily validates the feasibility of the deep learning approach and has not been compared against most state-of-the-art baseline methods (to be addressed in future work), rigorous statistical analyses demonstrate statistically significant improvements in all key metrics (p < 0.01). For instance, the SSIM enhancement for circular geons under salt-and-pepper noise reaches Δ 0.15 ± 0.03 (95% CI [0.12, 0.18], p = 0.0036), validating the robustness of our approach. These results corroborate Marr’s primal sketch theory, which posits that visual cognition is based on extracting asymmetric geometric primitives rather than pixel-level correspondence. However, square geons exhibit reduced performance in complex scenes, primarily due to edge fragmentation and occlusions that disrupt contour continuity. The combined experimental and theoretical analysis indicates that the geon operator effectively captures local and global features via skip connections and hierarchical feature fusion. This architecture performs particularly well for highly regular primitives, such as circles. Skip connections emulate the V1→IT visual pathway by integrating low-level and high-level representations, mirroring cortical hierarchical processing and offering a biologically inspired paradigm for visual recognition systems. Future work will explore diffusion-based generative models for open-set geon learning, leveraging their capacity to synthesize diverse, high-fidelity samples and extend detection to unseen primitives.

4. Conclusions

We demonstrate that visual geons offer substantial potential and research value across computer graphics, vision science, and artificial intelligence. Geon recognition is thus crucial for understanding complex objects and provides key insights into underlying visual processing mechanisms. Unlike Marr’s predefined categories, our study integrates theories from psychology, neuroscience, deep learning, and signal processing to model visual geons as emergent primitives. We propose that the visual system hierarchically analyzes and combines basic geons, and we employ a task-specific object imaging framework—implemented via U-Net and mathematical operators—to capture and validate geon representations. The experiments confirm the significance of geon theory in recognition and, through θ-parameter plasticity, achieve unified modeling of Marr’s algorithmic and implementation levels—resolving the dynamic weight allocation challenge of the RBCs framework. The U-Net encoder–decoder architecture emulates the V1→V4→IT visual pathway, with skip connections reproducing feedforward–feedback loops observed via fMRI, and noise robustness tests (Gaussian and salt-and-pepper) further validate the framework’s resilience, statistically confirming the framework’s alignment with Treisman’s feature integration principles.
Although our framework demonstrates strong performance, several limitations warrant discussion. First, the current implementation focuses on synthetic 2D geometric primitives, whereas real-world objects exhibit complex 3D structures and material properties. Second, the U-Net architecture struggles with occluded geon components due to its reliance on continuous edge information. Third, the biological plausibility of skip connections remains debated, as neuroscientific evidence suggests more complex feedback mechanisms in V1-IT pathways [58]. Finally, our dataset’s limited scale (1100 images) and categorical diversity (3 primitives) constrain generalizability to open-set scenarios. Future work will address these limitations through (1) the integration of volumetric geon representations [30,31,32,33,34,35,36] and (2) attention-guided occlusion reasoning [33]. This work bridges the historical gap between symbolic RBCs theory and subsymbolic deep learning, providing both computational validation of neurocognitive models and biologically inspired improvements to geometric deep learning architectures.

Author Contributions

Conceptualization, J.L.; methodology, J.L. and Y.W.; software, Y.W. and M.L.; validation, Y.W. and M.L.; formal analysis, Y.W. and M.L.; investigation, Y.W. and M.L.; writing—original draft preparation, M.L.; writing—review and editing, Y.W., M.L. and J.L.; supervision, J.L.; project administration, J.L.; funding acquisition, J.L. All authors have read and agreed to the published version of the manuscript.

Funding

This research received no external funding.

Data Availability Statement

The experimental data and implementation code supporting this study are available in the GitHub repository at https://github.com/haiiro-l/Detection-and-Recognition-of-Visual-Geons-Based-on-Specific-Object-of-Interest-Imaging-Technology (accessed on 6 May 2025). Processed datasets are available from the corresponding author upon reasonable request.

Conflicts of Interest

The authors declare no conflicts of interest.

References

  1. Biederman, I. Human image understanding: Recent research and a theory. Comput. Vision Graph. Image Process. 1985, 32, 29–73. [Google Scholar] [CrossRef]
  2. Biederman, I. Recognition-by-components: A theory of human image understanding. Psychol. Rev. 1987, 94, 115–147. [Google Scholar] [CrossRef] [PubMed]
  3. Marr, D.; Hildreth, E. Theory of edge detection. Proc. R. Soc. Lond. Ser. B Biol. Sci. 1980, 207, 187–217. [Google Scholar] [CrossRef]
  4. Marr, D.; Poggio, T. From Understanding Computation to Understanding Neural Circuitry; Massachusetts Institute of Technology: Cambridge, MA, USA, 1976. [Google Scholar]
  5. Marr, D. Visual information processing: The structure and creation of visual representations. Philos. Trans. R. Soc. Lond. Ser. B Biol. Sci. 1980, 290, 199–218. [Google Scholar] [CrossRef]
  6. Marr, D. Early processing of visual information. Philos. Trans. R. Soc. Lond. Ser. B Biol. Sci. 1976, 275, 483–519. [Google Scholar] [CrossRef]
  7. Marr, D. A theory for cerebral neocortex. Proc. R. Soc. Lond. Ser. B Biol. Sci. 1970, 176, 161–234. [Google Scholar]
  8. Marr, D. Representing visual information. Comput. Vis. Syst. 1978, 10, 101–180. [Google Scholar]
  9. Marr, D. A theory of cerebellar cortex. J. Physiol. 1969, 202, 437–470. [Google Scholar] [CrossRef]
  10. Marr, D. Analysis of occluding contour. Proc. R. Soc. Lond. Ser. B Biol. Sci. 1977, 197, 441–475. [Google Scholar] [CrossRef]
  11. Marr, D.; Nishihara, H.K. Representation and recognition of the spatial organization of three-dimensional shapes. Proc. R. Soc. Lond. Ser. B Biol. Sci. 1978, 200, 269–294. [Google Scholar] [CrossRef]
  12. Marr, D.; Poggio, T. A computational theory of human stereo vision. Proc. R. Soc. Lond. Ser. B Biol. Sci. 1979, 204, 301–328. [Google Scholar] [CrossRef]
  13. Marr, D.; Ullman, S. Directional selectivity and its use in early visual processing. Proc. R. Soc. Lond. Ser. B Biol. Sci. 1981, 211, 151–180. [Google Scholar] [CrossRef]
  14. Marr, D. Artificial intelligence—A personal view. Artif. Intell. 1977, 9, 37–48. [Google Scholar] [CrossRef]
  15. Marr, D.; Vaina, L. Representation and recognition of the movements of shapes. Proc. R. Soc. Lond. Ser. B Biol. Sci. 1982, 214, 501–524. [Google Scholar] [CrossRef]
  16. Torre, V.; Poggio, T.A. On edge detection. IEEE Trans. Pattern Anal. Mach. Intell. 1986, 8, 147–163. [Google Scholar] [CrossRef]
  17. Bertero, M.; Poggio, T.A.; Torre, V. Ill-posed problems in early vision. Proc. IEEE 1988, 76, 869–889. [Google Scholar] [CrossRef]
  18. Yuille, A.L.; Poggio, T.A. Scaling theorems for zero crossings. IEEE Trans. Pattern Anal. Mach. Intell. 1986, 8, 15–25. [Google Scholar] [CrossRef]
  19. Brunelli, R.; Poggio, T. Face recognition: Features versus templates. IEEE Trans. Pattern Anal. Mach. Intell. 1993, 15, 1042–1052. [Google Scholar] [CrossRef]
  20. Sung, K.-K.; Poggio, T. Example-based learning for view-based human face detection. IEEE Trans. Pattern Anal. Mach. Intell. 1998, 20, 39–51. [Google Scholar] [CrossRef]
  21. Papageorgiou, C.; Poggio, T. A trainable system for object detection. Int. J. Comput. Vis. 2000, 38, 15–33. [Google Scholar] [CrossRef]
  22. Riesenhuber, M.; Poggio, T. Hierarchical models of object recognition in cortex. Nat. Neurosci. 1999, 2, 1019–1025. [Google Scholar] [CrossRef] [PubMed]
  23. Patil-Takbhate, B.; Takbhate, R.; Khopkar-Kale, P.; Tripathy, S. Marvin Minsky: The Visionary Behind the Confocal Microscope and the Father of Artificial Intelligence. Cureus 2024, 16, e68434. [Google Scholar] [CrossRef]
  24. Minsky, M.L. Why people think computers can’t. AI Mag. 1982, 3, 3. [Google Scholar] [CrossRef]
  25. Minsky, M. A Framework for Representing Knowledge; Massachusetts Institute of Technology: Cambridge, MA, USA, 1974. [Google Scholar] [CrossRef]
  26. Minsky, M.; Papert, S. Perceptrons: An Introduction to Computational Geometry; MIT Press: Cambridge, MA, USA, 1969; pp. 1–20. [Google Scholar] [CrossRef]
  27. Minsky, M. Commonsense-based interfaces. Commun. ACM 2000, 43, 66–73. [Google Scholar] [CrossRef]
  28. Minsky, M. Decentralized minds. Behav. Brain Sci. 1980, 3, 439–440. [Google Scholar] [CrossRef]
  29. Yan, S.; Yang, Z.; Ma, C.; Huang, H.; Vouga, E.; Huang, Q. Hpnet: Deep primitive segmentation using hybrid representations. In Proceedings of the IEEE/CVF International Conference on Computer Vision, Montreal, BC, Canada, 11–17 October 2021; 17 October 2021; pp. 2753–2762. [Google Scholar] [CrossRef]
  30. Zou, C.; Yumer, E.; Yang, J.; Ceylan, D.; Hoiem, D. 3d-prnn: Generating shape primitives with recurrent neural networks. In Proceedings of the IEEE International Conference on Computer Vision, Venice, Italy, 22–29 October 2017; pp. 900–909. [Google Scholar] [CrossRef]
  31. Lin, Y.; Tang, C.; Chu, F.-J.; Vela, P.A. Using synthetic data and deep networks to recognize primitive shapes for object grasping. In Proceedings of the 2020 IEEE International Conference on Robotics and Automation (ICRA), Paris, France, 31 May–31 August 2020; pp. 10494–10501. [Google Scholar] [CrossRef]
  32. Aliev, K.A.; Sevastopolsky, A.; Kolos, M.; Ulyanov, D.; Lempitsky, V. Neural Point-Based Graphics. In Computer Vision—ECCV 2020; Vedaldi, A., Bischof, H., Brox, T., Frahm, J.M., Eds.; Lecture Notes in Computer Science; Springer: Cham, Switzerland, 2020; Volume 12367, pp. 42–61. [Google Scholar] [CrossRef]
  33. Zhang, Z.; Sun, B.; Yang, H.; Huang, Q. H3DNet: 3D Object Detection Using Hybrid Geometric Primitives. In Computer Vision—ECCV 2020; Vedaldi, A., Bischof, H., Brox, T., Frahm, J.M., Eds.; Lecture Notes in Computer Science; Springer: Cham, Switzerland, 2020; Volume 12357, pp. 311–329. [Google Scholar] [CrossRef]
  34. Li, L.; Fu, H.; Tai, C.-L. Fast sketch segmentation and labeling with deep learning. IEEE Comput. Graph. Appl. 2019, 39, 38–51. [Google Scholar] [CrossRef]
  35. Huang, J.; Gao, J.; Ganapathi-Subramanian, V.; Su, H.; Liu, Y.; Tang, C.; Guibas, L.J. DeepPrimitive: Image decomposition by layered primitive detection. Comp. Visual Media 2018, 4, 385–397. [Google Scholar] [CrossRef]
  36. Sharma, G.; Goyal, R.; Liu, D.; Kalogerakis, E.; Maji, S. Neural Shape Parsers for Constructive Solid Geometry. IEEE Trans. Pattern Anal. Mach. Intell. 2022, 44, 2628–2640. [Google Scholar] [CrossRef] [PubMed]
  37. Zunair, H.; Hamza, A.B. Sharp U-Net: Depthwise convolutional network for biomedical image segmentation. Comput. Biol. Med. 2021, 136, 104699. [Google Scholar] [CrossRef]
  38. Treisman, A. Preattentive processing in vision. Comput. Vision Graph. Image Process. 1985, 31, 156–177. [Google Scholar] [CrossRef]
  39. Treisman, A.; Gormican, S. Feature analysis in early vision: Evidence from search asymmetries. Psychol. Rev. 1988, 95, 15–48. [Google Scholar] [CrossRef] [PubMed]
  40. Treisman, A. Features and objects in visual processing. Sci. Am. 1986, 255, 114B–125. [Google Scholar] [CrossRef]
  41. Treisman, A. Perceptual grouping and attention in visual search for features and for objects. J. Exp. Psychol. Hum. Percept. Perform. 1982, 8, 194–214. [Google Scholar] [CrossRef]
  42. Treisman, A. How the deployment of attention determines what we see. Visual Cogn. 2006, 14, 411–443. [Google Scholar] [CrossRef]
  43. Treisman, A. Binocular rivalry and stereoscopic depth perception. In From Perception to Consciousness: Searching with Anne Treisman; Wolfe, J., Robertson, L., Eds.; Oxford University Press: Oxford, UK, 2012; pp. 59–69. [Google Scholar] [CrossRef]
  44. Treisman, A. Feature binding, attention and object perception. Philos. Trans. R. Soc. B Biol. Sci. 1998, 353, 1295–1306. [Google Scholar] [CrossRef] [PubMed]
  45. Treisman, A. Features and objects: The Fourteenth Bartlett Memorial Lecture. Q. J. Exp. Psychol. A Hum. Exp. Psychol. 1988, 40A, 201–237. [Google Scholar] [CrossRef]
  46. Treisman, A.; DeSchepper, B. Object tokens, attention, and visual memory. In Attention and Performance 16: Information Integration in Perception and Communication; Inui, T., McClelland, J.L., Eds.; The MIT Press: Cambridge, MA, USA, 1996; pp. 15–46. [Google Scholar]
  47. Treisman, A.; Paterson, R. Emergent features, attention, and object perception. J. Exp. Psychol. Hum. Percept. Perform. 1984, 10, 12–31. [Google Scholar] [CrossRef]
  48. Treisman, A.M.; Gelade, G. A feature-integration theory of attention. Cogn. Psychol. 1980, 12, 97–136. [Google Scholar] [CrossRef] [PubMed]
  49. Treisman, A. Visual attention and the perception of features and objects. Can. Psychol./Psychol. Can. 1994, 35, 107–108. [Google Scholar] [CrossRef]
  50. Itti, L.; Koch, C. Computational modelling of visual attention. Nat. Rev. Neurosci. 2001, 2, 194–203. [Google Scholar] [CrossRef]
  51. Borji, A.; Itti, L. State-of-the-Art in Visual Attention Modeling. IEEE Trans. Pattern Anal. Mach. Intell. 2013, 35, 185–207. [Google Scholar] [CrossRef] [PubMed]
  52. Itti, L.; Koch, C.; Braun, J. Revisiting spatial vision: Toward a unifying model. J. Opt. Soc. Am. A Opt. Image Sci. Vis. 2000, 17, 1899–1917. [Google Scholar] [CrossRef] [PubMed]
  53. Itti, L. Automatic foveation for video compression using a neurobiological model of visual attention. IEEE Trans. Image Process. Publ. IEEE Signal Process. Soc. 2004, 13, 1304–1318. [Google Scholar] [CrossRef]
  54. Itti, L. Quantitative modelling of perceptual salience at human eye position. Vis. Cogn. 2006, 14, 959–984. [Google Scholar] [CrossRef]
  55. Carmi, R.; Itti, L. The role of memory in guiding attention during natural vision. J. Vis. 2006, 6, 898–914. [Google Scholar] [CrossRef] [PubMed]
  56. Itti, L.; Gold, C.; Koch, C. Visual attention and target detection in cluttered natural scenes. Opt. Eng. 2001, 40, 1784–1793. [Google Scholar] [CrossRef]
  57. Itti, L.; Koch, C. Comparison of feature combination strategies for saliency-based visual attention systems. In Proceedings of the SPIE 3644, Human Vision and Electronic Imaging IV, San Jose, CA, USA, 23–29 January 1999. [Google Scholar] [CrossRef]
  58. Zhang, Y.; Wu, X.; Zheng, C.; Zhao, Y.; Gao, J.; Deng, Z.; Zhang, X.; Chen, J. Effects of Vergence Eye Movement Planning on Size Perception and Early Visual Processing. J. Cogn. Neurosci. 2024, 36, 2793–2806. [Google Scholar] [CrossRef]
  59. Cai, Y.; Zhang, D.; Liang, B.; Wang, Z.; Li, J.; Gao, Z.; Gao, M.; Huang, R.; Liu, M. Visual Creativity Imagery Modulates Local Spontaneous Activity Amplitude of Resting-State Brain. In Proceedings of the 2016 Annual Meeting of the Organization for Human Brain Mapping, Geneva, Switzerland, 26–30 June 2016; Abstract 1525. [Google Scholar]
  60. Wen, X.; Zhang, D.; Liang, B.; Zhang, R.; Wang, Z.; Wang, J.; Liu, M.; Huang, R. Reconfiguration of the Brain Functional Network Associated with Visual Task Demands. PLoS ONE 2015, 10, e0132518. [Google Scholar] [CrossRef]
  61. Zhou, L.F.; Wang, K.; He, L.; Meng, M. Twofold advantages of face processing with or without visual awareness. J. Exp. Psychol. Hum. Percept. Perform. 2021, 47, 784–794. [Google Scholar] [CrossRef]
  62. Zhu, H.; Cai, T.; Xu, J.; Wu, S.; Li, X.; He, S. Neural correlates of stereoscopic depth perception: A fNIRS study. In Proceedings of the 2016 Progress in Electromagnetic Research Symposium (PIERS), Shanghai, China, 8–11 August 2016; IEEE: Piscataway, NJ, USA, 2016; pp. 4442–4446. [Google Scholar]
  63. Li, J.; Chen, S.; Wang, S.; Lei, M.; Dai, X.; Liang, C.; Xu, K.Y.; Lin, S.; Li, Y.; Fan, Y.; et al. An optical biomimetic eyes with interested object imaging. arXiv 2021, arXiv:2108.04236. [Google Scholar] [CrossRef]
  64. Li, J.; Lei, M.; Dai, X.; Wang, S.; Zhong, T.; Liang, C.; Wang, C.; Xie, P.; Wang, R. Compressed Sensing Based Object Imaging System and Imaging Method Therefor. U.S. Patent 2021/0144278 A1, 13 May 2021. [Google Scholar]
  65. Ronneberger, O.; Fischer, P.; Brox, T. U-Net: Convolutional Networks for Biomedical Image Segmentation. In Medical Image Computing and Computer-Assisted Intervention—MICCAI 2015; Navab, N., Hornegger, J., Wells, W., Frangi, A., Eds.; Lecture Notes in Computer Science; Springer: Cham, Switzerland, 2015; Volume 9351. [Google Scholar] [CrossRef]
  66. Rosenblatt, F. The perceptron: A probabilistic model for information storage and organization in the brain. Psychol. Rev. 1958, 65, 386–408. [Google Scholar] [CrossRef] [PubMed]
  67. Liu, Z.; Wang, Y.; Vaidya, S.; Ruehle, F.; Halverson, J.; Soljačić, M.; Tegmark, M. Kan: Kolmogorov-arnold networks. arXiv 2024, arXiv:2404.19756. [Google Scholar] [CrossRef]
  68. Zhang, K.; Zuo, W.; Zhang, L. Deep Plug-And-Play Super-Resolution for Arbitrary Blur Kernels. In Proceedings of the 2019 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), Long Beach, CA, USA, 15–20 June 2019; pp. 1671–1681. [Google Scholar] [CrossRef]
  69. Horé, A.; Ziou, D. Image Quality Metrics: PSNR vs. SSIM. In Proceedings of the 2010 20th International Conference on Pattern Recognition, Istanbul, Turkey, 23–26 August 2010; pp. 2366–2369. [Google Scholar] [CrossRef]
  70. Bauer, E.; Kohavi, R. An Empirical Comparison of Voting Classification Algorithms: Bagging, Boosting, and Variants. Mach. Learn. 1999, 36, 105–139. [Google Scholar] [CrossRef]
Figure 1. Technical schematic of the object-of-interest imaging methodology.
Figure 1. Technical schematic of the object-of-interest imaging methodology.
Sensors 25 03022 g001
Figure 2. U-Net architecture schematic.
Figure 2. U-Net architecture schematic.
Sensors 25 03022 g002
Figure 3. Fundamental detection experimental results diagram.
Figure 3. Fundamental detection experimental results diagram.
Sensors 25 03022 g003
Figure 4. Geon separation detection experimental results diagram.
Figure 4. Geon separation detection experimental results diagram.
Sensors 25 03022 g004
Figure 5. Multi-Geon detection experimental results diagram.
Figure 5. Multi-Geon detection experimental results diagram.
Sensors 25 03022 g005
Figure 6. Gaussian noise experimental results diagram for basic detection tasks.
Figure 6. Gaussian noise experimental results diagram for basic detection tasks.
Sensors 25 03022 g006
Figure 7. Gaussian noise experimental results diagram for geon separation.
Figure 7. Gaussian noise experimental results diagram for geon separation.
Sensors 25 03022 g007
Figure 8. Gaussian noise experimental results diagram for multi-target detection.
Figure 8. Gaussian noise experimental results diagram for multi-target detection.
Sensors 25 03022 g008
Figure 9. Salt-and-pepper noise Experimental results diagram for basic detection tasks.
Figure 9. Salt-and-pepper noise Experimental results diagram for basic detection tasks.
Sensors 25 03022 g009
Figure 10. Salt-and-pepper noise experimental results diagram for geon separation.
Figure 10. Salt-and-pepper noise experimental results diagram for geon separation.
Sensors 25 03022 g010
Figure 11. Salt-and-pepper noise experimental results diagram for multi-target detection.
Figure 11. Salt-and-pepper noise experimental results diagram for multi-target detection.
Sensors 25 03022 g011
Table 1. Comparison of Theoretical Frameworks for Geon Representation.
Table 1. Comparison of Theoretical Frameworks for Geon Representation.
Theoretical FrameworkDefinition of GeonsDynamic AdaptabilityNoise RobustnessImprovements in This Paper
RBC (Biederman)Fixed 36 geometric shapesLowNot involvedAdaptive learning of geons
I-TheoryGradual abstraction of featuresMediumPartialIncorporation of sparse coding
DeepPrimitiveData-driven action segmentationHighData-dependentCross-modal bioinspired
Table 2. Fundamental detection experimental result metrics.
Table 2. Fundamental detection experimental result metrics.
GeonSSIM (95% CI)PSNR (95% CI)MSE (95% CI)
Triangle0.93 ± 0.02 *59.14 ± 1.24 **0.10 ± 0.03 *
Circle0.99 ± 0.01 **54.64 ± 1.05 **0.23 ± 0.05 **
Square0.98 ± 0.01 **58.40 ± 1.18 **0.11 ± 0.02 **
Table 3. Geon separation detection Result metrics.
Table 3. Geon separation detection Result metrics.
GeonSSIM (SEM)PSNR (SEM)MSE (SEM)
Triangle0.88 ± 0.03 ***62.34 ± 0.89 ***0.05 ± 0.01 ***
Circle0.99 ± 0.005 ***60.05 ± 0.76 ***0.06 ± 0.01 ***
Square0.80 ± 0.04 **58.88 ± 1.12 **0.10 ± 0.02 **
Table 4. Multi-Geon detection result metrics.
Table 4. Multi-Geon detection result metrics.
GeonSSIM (SEM)PSNR (SEM)MSE (SEM)
Triangle0.79 ± 0.03 ***46.17 ± 0.89 ***2.64 ± 0.01 ***
Circle0.98 ± 0.005 ***48.58 ± 0.76 ***1.35 ± 0.01 ***
Square0.77 ± 0.04 **51.55 ± 1.12 **0.61 ± 0.02 **
Table 5. Gaussian noise result metrics.
Table 5. Gaussian noise result metrics.
GeonSceneSSIMPSNRMSE
Basic Detection0.6942.224.03
TriangleGeon Separation0.8361.450.06
Multi-Target0.7046.931.98
Basic Detection0.8545.601.87
CircleGeon Separation0.9555.360.20
Multi-Target0.8946.191.85
Basic Detection0.5656.060.18
SquareGeon Separation0.4957.170.13
Multi-Target0.4651.100.67
Table 6. Salt-and-pepper noise result metrics.
Table 6. Salt-and-pepper noise result metrics.
GeonSceneSSIMPSNRMSE
Basic Detection0.8856.280.16
TriangleGeon Separation0.6559.350.08
Multi-Target0.6346.422.24
Basic Detection0.9857.190.13
CircleGeon Separation0.9859.660.07
Multi-Target0.9848.551.34
Basic Detection0.8857.830.11
SquareGeon Separation0.7258.570.10
Multi-Target0.6851.150.71
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Wu, Y.; Liu, M.; Li, J. Detection and Recognition of Visual Geons Based on Specific Object-of-Interest Imaging Technology. Sensors 2025, 25, 3022. https://doi.org/10.3390/s25103022

AMA Style

Wu Y, Liu M, Li J. Detection and Recognition of Visual Geons Based on Specific Object-of-Interest Imaging Technology. Sensors. 2025; 25(10):3022. https://doi.org/10.3390/s25103022

Chicago/Turabian Style

Wu, Yonghao, Minyi Liu, and Jun Li. 2025. "Detection and Recognition of Visual Geons Based on Specific Object-of-Interest Imaging Technology" Sensors 25, no. 10: 3022. https://doi.org/10.3390/s25103022

APA Style

Wu, Y., Liu, M., & Li, J. (2025). Detection and Recognition of Visual Geons Based on Specific Object-of-Interest Imaging Technology. Sensors, 25(10), 3022. https://doi.org/10.3390/s25103022

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop