Sign in to use this feature.

Years

Between: -

Article Types

Countries / Regions

remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline

Search Results (2,282)

Search Parameters:
Journal = J. Imaging

Order results
Result details
Results per page
Select all
Export citation of selected articles as:
24 pages, 28764 KB  
Article
Restoration of Non-Uniform Motion-Blurred Star Images Based on Dynamic Strip Attention
by Jixin Han, Zhaodong Niu and Jun He
J. Imaging 2026, 12(3), 103; https://doi.org/10.3390/jimaging12030103 (registering DOI) - 27 Feb 2026
Abstract
When capturing star images in long-exposure mode, due to the relative motion between stars and space objects and the observation camera, strip tailings with different directions and lengths will be formed, resulting in a serious decline in image quality and inaccurate centroid positioning. [...] Read more.
When capturing star images in long-exposure mode, due to the relative motion between stars and space objects and the observation camera, strip tailings with different directions and lengths will be formed, resulting in a serious decline in image quality and inaccurate centroid positioning. Traditional methods for restoring star images are prone to ringing effects and cannot restore the non-uniformly blurred star images. Aiming at this problem, this paper proposes a star image restoration network based on a dynamic strip attention mechanism. Firstly, a Multi-scale Dynamic Strip Pooling Module is designed to adaptively extract blurred features of different lengths and directions by dynamically adjusting the strip convolution. After that, a Multi-scale Feature Fusion Module is designed to fuse multi-level features to reduce the loss of image details of stars and space objects in the image. Experimental results demonstrate that the proposed method achieves a PSNR of 84.08 and an SSIM of 0.9928 on the 16-bit simulated dataset, outperforming both traditional methods and other deep learning-based approaches. Specifically, the recognition accuracy of star points is increased by 174% in comparison with unprocessed images. Furthermore, this paper validates the network using the real-world dataset spotGEO, and the results indicate that the average number of successfully recognized star points is increased by 57% compared to direct processing of the original images. Full article
(This article belongs to the Section Image and Video Processing)
Show Figures

Figure 1

30 pages, 18122 KB  
Article
Fine-Grained Age-Class Identification of Moso Bamboo Using an Improved Lightweight YOLO11 Model
by Yingbin Zhang, Xinhuang Zhang, Zhichao Cai, Xi He, Shuwei Chen, Zhengxuan Lai, Kunyong Yu and Riwen Lai
J. Imaging 2026, 12(3), 102; https://doi.org/10.3390/jimaging12030102 - 27 Feb 2026
Abstract
Accurate identification of moso bamboo (Phyllostachys edulis) age classes is essential for effective forestry resource management, yet existing methods often struggle to achieve a satisfactory balance between accuracy and computational efficiency under complex field conditions. To address this challenge, this study [...] Read more.
Accurate identification of moso bamboo (Phyllostachys edulis) age classes is essential for effective forestry resource management, yet existing methods often struggle to achieve a satisfactory balance between accuracy and computational efficiency under complex field conditions. To address this challenge, this study proposes a lightweight object detection model, termed YOLO11-GCR, for fine-grained moso bamboo age-class classification based on close-range imagery. The proposed approach builds upon the YOLO11 framework and incorporates Ghost convolution, the Convolutional Block Attention Module (CBAM), and a Receptive Field Block (RFB) to reduce model complexity, enhance discriminative feature representation, and improve sensitivity to subtle texture variations among age classes. A dataset consisting of 9538 annotated bamboo culm images covering four age classes (I-du to IV-du) was constructed and divided into training, validation, and independent test sets with strict spatiotemporal separation. Experimental results indicate that YOLO11-GCR achieves robust detection performance with a lightweight architecture of 2.62 × 106 parameters and 6.2 GFLOPs, yielding an mAP@0.5 of 0.913 and an mAP@0.5–0.95 of 0.895 on the independent test set. Notably, the model demonstrates improved classification stability for visually similar age classes, such as II-du and III-du. Overall, this study presents an efficient and practical imaging-based solution for automated moso bamboo age-class recognition in complex natural environments. Full article
(This article belongs to the Section Computer Vision and Pattern Recognition)
Show Figures

Figure 1

12 pages, 6299 KB  
Communication
Lensless Quantitative Phase Imaging with Bayer-Filtered Color Sensors Under Sequential RGB-LED Illumination
by Jiajia Wu, Yining Li, Yuheng Luo, Leiting Pan, Pengming Song and Qiang Xu
J. Imaging 2026, 12(3), 101; https://doi.org/10.3390/jimaging12030101 - 26 Feb 2026
Abstract
Lensless on-chip microscopy enables high-throughput, wide-FOV imaging; however, the Bayer color filter array (CFA) in standard color sensors spatially multiplexes spectral channels, introducing sub-sampling and spectral crosstalk that degrade phase retrieval. We propose a Wirtinger Poly-Gradient Solver (WPGS) for quantitative phase reconstruction with [...] Read more.
Lensless on-chip microscopy enables high-throughput, wide-FOV imaging; however, the Bayer color filter array (CFA) in standard color sensors spatially multiplexes spectral channels, introducing sub-sampling and spectral crosstalk that degrade phase retrieval. We propose a Wirtinger Poly-Gradient Solver (WPGS) for quantitative phase reconstruction with Bayer-filtered color sensors under sequential Red–Green–Blue Light-Emitting Diode (RGB-LED) illumination. The method combines Transport of Intensity Equation (TIE)-based initialization with polychromatic Wirtinger optimization to suppress CFA-induced artifacts and enable pixel super-resolution (PSR). Experiments resolve a 2.76 μm linewidth using a 1.85 μm pixel-pitch sensor, exceeding the nominal Nyquist limit imposed by pixel sampling. We further demonstrate label-free imaging of HeLa cells and unstained tissue sections, supporting high-throughput digital pathology and offering potential for longitudinal biological observation. Full article
(This article belongs to the Section Computational Imaging and Computational Photography)
Show Figures

Figure 1

39 pages, 1849 KB  
Review
The Augmented Cytopathologist: A Conceptual Exploratory Narrative Review on Immersive and Vision–Language Models Tools in Digital Pathology
by Enrico Giarnieri, Andrea Lastrucci, Alberto Ricci, Pierdonato Bruno and Daniele Giansanti
J. Imaging 2026, 12(3), 100; https://doi.org/10.3390/jimaging12030100 - 26 Feb 2026
Abstract
Emerging digital technologies, including immersive environments (VR/AR/XR) and Vision–Language Models (VLMs), have the potential to reshape digital pathology and medical imaging. While immersive tools can enhance spatial visualization and procedural training, VLM-based copilots offer cognitive and workflow support. Their combined impact on cytopathology [...] Read more.
Emerging digital technologies, including immersive environments (VR/AR/XR) and Vision–Language Models (VLMs), have the potential to reshape digital pathology and medical imaging. While immersive tools can enhance spatial visualization and procedural training, VLM-based copilots offer cognitive and workflow support. Their combined impact on cytopathology remains largely conceptual and preclinical. This Conceptual Exploratory Narrative Review (CENR) examines how immersive technologies and VLM-based copilots may jointly influence cytopathologists’ professional workflow, training, and diagnostic processes, introducing the notion of the “augmented cytopathologist.” A structured exploratory approach integrated peer-reviewed literature, position papers, preprints, gray literature (technical reports, white papers, conference abstracts, blogs), and cross-disciplinary perspectives. Database searches (PubMed, Web of Science, Scopus) confirmed a limited number of studies addressing immersive or AI-assisted cytopathology imaging. Thematic analysis focused on four conceptual dimensions: (1) technological capabilities and maturity; (2) workflow and educational applications; (3) professional implications and cytopathologist role; and (4) responsible use of LLMs and VLMs as supportive tools. This approach emphasizes interpretation of emerging trends over aggregation of empirical data, enabling conceptual synthesis of early-stage implementations and perspectives in the field. Immersive technologies facilitate three-dimensional visualization, procedural skill development, and collaborative engagement, whereas VLMs support report generation, literature retrieval, and decision guidance. Together, they offer a synergistic model for perceptual and cognitive augmentation. Key challenges include technical maturity, interoperability, workflow integration, regulatory compliance, and ethical oversight. Figures illustrate representative examples of (1) remote collaborative immersive evaluation and (2) integration of immersive visualization with VLM-based copilots, highlighting potential applications in training and workflow support. The CENR underscores the potential of combining immersive tools and AI copilots to support cytopathology, particularly for education, workflow efficiency, and cognitive augmentation. Adoption should be incremental and carefully governed, emphasizing augmentative rather than transformative use. Future research should focus on clinical validation, scalable integration, and regulatory and ethical frameworks to realize the concept of the augmented cytopathologist in practice. Full article
Show Figures

Figure 1

21 pages, 31860 KB  
Article
Design and Development of an Automated Pipeline for Medical Hyperspectral Image Acquisition, Processing, and Fusion
by Felix Wühler, Tim Markus Häußermann, Alessa Rache, Björn van Marwick, Carmen Wängler, Julian Reichwald and Matthias Rädle
J. Imaging 2026, 12(3), 99; https://doi.org/10.3390/jimaging12030099 - 25 Feb 2026
Viewed by 25
Abstract
Automated and comprehensive processing of hyperspectral image data is increasingly important in academic research and medical technology. This study presents an automated processing pipeline that integrates hyperspectral image acquisition, analysis, multimodal fusion, and centralized data management to improve the interpretability of spectral information [...] Read more.
Automated and comprehensive processing of hyperspectral image data is increasingly important in academic research and medical technology. This study presents an automated processing pipeline that integrates hyperspectral image acquisition, analysis, multimodal fusion, and centralized data management to improve the interpretability of spectral information for biological tissue analysis. The pipeline supports modular hyperspectral data processing, fusion of complementary wavelength ranges, and scalable data storage, and was implemented in Python 3.13.3. The pipeline was evaluated using hyperspectral imaging data acquired from a coronal mouse brain section. Clustering-based analysis and spectral correlation metrics were applied to assess the impact of multimodal data fusion on spectral representation. Clustering of individual modalities yielded silhouette coefficients of 0.5879 for near-infrared data, 0.6020 for mid-infrared data, and 0.6715 for RGB data. Multimodal fusion reduced the silhouette coefficient to 0.5420 and enabled the identification of anatomical structures that were not distinguishable in any single modality. High spectral correlation coefficients exceeding 0.98 confirmed that spectral fidelity was preserved during fusion. These results demonstrate that automated multimodal hyperspectral data fusion can enhance the interpretability of biological tissue despite reduced clustering compactness. The proposed pipeline provides a structured framework for preclinical hyperspectral imaging workflows and supports exploratory biological analysis in medical imaging contexts. Full article
(This article belongs to the Special Issue Multispectral and Hyperspectral Imaging: Progress and Challenges)
Show Figures

Figure 1

20 pages, 1379 KB  
Article
Hybrid Vision Transformer–CNN Framework for Alzheimer’s Disease Cell Type Classification: A Comparative Study with Vision–Language Models
by Md Easin Hasan, Md Tahmid Hasan Fuad, Omar Sharif and Amy Wagler
J. Imaging 2026, 12(3), 98; https://doi.org/10.3390/jimaging12030098 - 25 Feb 2026
Viewed by 34
Abstract
Accurate identification of Alzheimer’s disease (AD)-related cellular characteristics from microscopy images is essential for understanding neurodegenerative mechanisms at the cellular level. While most computational approaches focus on macroscopic neuroimaging modalities, cell type classification from microscopy remains relatively underexplored. In this study, we propose [...] Read more.
Accurate identification of Alzheimer’s disease (AD)-related cellular characteristics from microscopy images is essential for understanding neurodegenerative mechanisms at the cellular level. While most computational approaches focus on macroscopic neuroimaging modalities, cell type classification from microscopy remains relatively underexplored. In this study, we propose a hybrid vision transformer–convolutional neural network (ViT–CNN) framework that integrates DeiT-Small and EfficientNet-B7 to classify three AD-related cell types—astrocytes, cortical neurons, and SH-SY5Y neuroblastoma cells—from phase-contrast microscopy images. We perform a comparative evaluation against conventional CNN architectures (DenseNet, ResNet, InceptionNet, and MobileNet) and prompt-based multimodal vision–language models (GPT-5, GPT-4o, and Gemini 2.5-Flash) using zero-shot, few-shot, and chain-of-thought prompting. Experiments conducted with stratified fivefold cross-validation show that the proposed hybrid model achieves a test accuracy of 61.03% and a macro F1 score of 61.85, outperforming standalone CNN baselines and prompt-only LLM approaches under data-limited conditions. These results suggest that combining convolutional inductive biases with transformer-based global context modeling can improve generalization for cellular microscopy classification. While constrained by dataset size and scope, this work serves as a proof of concept and highlights promising directions for future research in domain-specific pretraining, multimodal data integration, and explainable AI for AD-related cellular analysis. Full article
Show Figures

Figure 1

18 pages, 14317 KB  
Article
A Deep Learning-Based Correction for Scanning Radius Errors in Circular-Scan Photoacoustic Tomography
by Jie Yin, Yingjie Feng, Junjun He, Min Xie and Chao Tao
J. Imaging 2026, 12(3), 97; https://doi.org/10.3390/jimaging12030097 - 25 Feb 2026
Viewed by 51
Abstract
Circular-Scan photoacoustic tomography (PAT) can provide high-resolution images of optical absorption, but its analytical reconstructions, such as delay-and-sum (DAS), are highly sensitive to scanning radius (SR) inaccuracies, which cause severe geometric distortions and artifacts. In this work, we propose a deep learning framework, [...] Read more.
Circular-Scan photoacoustic tomography (PAT) can provide high-resolution images of optical absorption, but its analytical reconstructions, such as delay-and-sum (DAS), are highly sensitive to scanning radius (SR) inaccuracies, which cause severe geometric distortions and artifacts. In this work, we propose a deep learning framework, termed smooth deconvolution ResNet (SD-ResNet), to correct DAS reconstruction degradation induced by SR errors. SD-ResNet uses an ImageNet-pretrained ResNet-50 encoder and a lightweight deconvolutional decoder with additional smoothing convolutions to suppress checkerboard artifacts and restore fine structural details. A paired training dataset is generated using k-Wave simulations driven by human thoracic computed tomography (CT) slices: for each phantom, radiofrequency data are simulated once, and DAS images reconstructed with the true SR serve as ground truth, whereas images reconstructed with biased SR values serve as inputs. This design provides structurally diverse training samples and enhances generalization. In silico experiments show that SD-ResNet effectively recovers image quality across a range of SR deviations. Phantom experiments with polyethylene microspheres further confirm that the proposed method can substantially reduce artifacts and recover correct source shapes under practical SR mismatches, offering a robust tool for SR-error-resilient PAT imaging. Full article
(This article belongs to the Section AI in Imaging)
Show Figures

Figure 1

31 pages, 11349 KB  
Article
Recognition, Localization and 3D Geometric Morphology Calculation of Microblind Holes in Complex Backgrounds Based on the Improved YOLOv11 Network and AVC Algorithm
by Chengfen Zhang, Dong Xia, Ruizhao Chen, Qunfeng Niu, Tao Wang and Li Wang
J. Imaging 2026, 12(3), 96; https://doi.org/10.3390/jimaging12030096 - 24 Feb 2026
Viewed by 162
Abstract
Microblind hole processing quality inspection, especially accurately identifying microblind hole contour features and precisely detecting 3D and morphological parameters, has always been challenging, especially for accurately identifying those of different sizes, depths, and contour features simultaneously. This poses a great challenge for identifying [...] Read more.
Microblind hole processing quality inspection, especially accurately identifying microblind hole contour features and precisely detecting 3D and morphological parameters, has always been challenging, especially for accurately identifying those of different sizes, depths, and contour features simultaneously. This poses a great challenge for identifying and localizing microblind hole contours based on machine vision and accurately calculating three-dimensional parameters. This study takes cigarette microblind holes (diameter of 0.1–0.2 mm, depth of approximately 35 µm) as the research object. It focuses on solving two major challenges: recognizing and localizing microblind hole contours in complex texture backgrounds and accurately calculating their 3D geometric morphology. An improved YOLOv11s model is proposed for microblind hole image multiobject detection with complex texture backgrounds to extract their features completely. An Area–Volume Computation (AVC) algorithm, which utilizes discrete integral estimation and curve-fitting principles, is also proposed for computing their surface area and volume. The experimental results show that the precision, recall, mAP@0.5, mAP@0.5:0.95, and prediction time of the improved YOLOv11 network are 0.915, 0.948, 0.925, 0.615, and 1.27 ms, respectively. The relative errors (REs) of the surface area and volume calculation of the microblind holes are 5.236% and 3.964%, respectively. The proposed method achieves microblind hole recognition, localization and 3D morphology calculation accuracy, meeting cigarette on-site inspection criteria. Additionally, a reference for detecting other similar objects in complex texture backgrounds and accurately calculating 3D tasks is provided. Full article
(This article belongs to the Topic Computer Vision and Image Processing, 3rd Edition)
Show Figures

Figure 1

33 pages, 2043 KB  
Article
Hybrid MICO-LAC Segmentation with Panoptic Tumor Instance Analysis for Dense Breast Mammograms
by Razia Jamil, Min Dong, Orken Mamyrbayev and Ainur Akhmediyarova
J. Imaging 2026, 12(3), 95; https://doi.org/10.3390/jimaging12030095 - 24 Feb 2026
Viewed by 77
Abstract
This study proposes a clinically driven hybrid segmentation framework for dense breast tissue analysis in mammographic images, addressing persistent challenges associated with intensity inhomogeneity, low-contrast, and complex tumor morphology. The framework integrates Multiplicative Intrinsic Component Optimization (MICO_2D) for bias field correction, followed by [...] Read more.
This study proposes a clinically driven hybrid segmentation framework for dense breast tissue analysis in mammographic images, addressing persistent challenges associated with intensity inhomogeneity, low-contrast, and complex tumor morphology. The framework integrates Multiplicative Intrinsic Component Optimization (MICO_2D) for bias field correction, followed by a distance-regularized multiphase Vese–Chan level-set model for coarse global tumor segmentation. To achieve precise boundary delineation, a localized refinement stage is employed using Localized Active Contours (LAC) with Local Image Fitting (LIF) energy, supported by Gaussian regularization to ensure smooth and coherent boundaries in regions with ambiguous tissue transitions. Building upon the refined semantic tumor mask, the framework further incorporates a panoptic-style tumor instance segmentation stage, enabling the decomposition of connected tumor regions into distinct anatomical instances, which were evaluated on both MIAS and INBreast mammography datasets to demonstrate generalizability. This extension facilitates detailed structural analysis of tumor multiplicity and spatial organization, enhancing interpretability beyond conventional pixel wise segmentation. Experiments conducted on Cranio-Caudal (CC) and Medio-Lateral Oblique (MLO) mammographic views demonstrate competitive performance relative to baseline U-Net and advanced deep learning fusion architectures, including multi-scale and multi-view networks, while offering improved interpretability and robustness. Quantitative evaluation using overlap-related metrics shows strong spatial agreement between predicted and reference segmentations, with per-image Dice Similarity Coefficient (DSC) and Intersection over Union (IoU) distributions reported to ensure reproducibility. Descriptive per-image analysis, supported by bootstrap-based confidence intervals and paired comparisons, indicates consistent performance improvements across images. Robustness analysis under realistic perturbations, including noise, contrast degradation, blur, and rotation, demonstrates stable performance across varying imaging conditions. Furthermore, feature space visualizations using t-SNE and UMAP reveal clear separability between cancerous and non-cancerous tissue regions, highlighting the discriminative capability of the proposed framework. Overall, the results demonstrate the effectiveness, robustness, and clinical motivation of this hybrid panoptic framework for comprehensive dense breast tumor analysis in mammography, while emphasizing reproducibility and conservative statistical assessment. Full article
(This article belongs to the Special Issue Current Progress in Medical Image Segmentation)
33 pages, 5215 KB  
Article
Towards Lightweight and Multi-Scale Scene Classification: A Lie Group-Guided Deep Learning Network with Collaborative Attention
by Xuefei Xu and Chengjun Xu
J. Imaging 2026, 12(3), 94; https://doi.org/10.3390/jimaging12030094 - 24 Feb 2026
Viewed by 117
Abstract
Remote sensing scene classification (RSSC) plays a crucial role in Earth observation. Current deep learning methods, while accurate, tend to focus on high-level semantic features and overlook complementary shallow details such as edges and textures. Moreover, conventional CNNs are limited by fixed receptive [...] Read more.
Remote sensing scene classification (RSSC) plays a crucial role in Earth observation. Current deep learning methods, while accurate, tend to focus on high-level semantic features and overlook complementary shallow details such as edges and textures. Moreover, conventional CNNs are limited by fixed receptive fields, whereas transformers incur high computational costs. To address these limitations, we propose the Lie Group lightweight multi-scale network (LGLMNet), a lightweight multi-scale network that integrates Lie Group covariance features. It employs a dual-branch architecture combining Lie Group machine learning (LGML) for shallow feature extraction and a deep learning branch for high-level semantics. In the deep branch, we design a parallel depthwise separable convolution block (PDSCB) for multi-scale perception and a spatial-channel collaborative attention mechanism (SCCA) for efficient global–local modeling. A cross-layer feature fusion block (CLFFB) effectively merges the two branches. Compared with state-of-the-art methods, the proposed LGLMNet achieves accuracy improvements of 2.14%, 2.32%, and 1.12% on UCM-21, AID, and NWPU-45 datasets, respectively, while maintaining a lightweight structure with only 2.6 M parameters. Full article
(This article belongs to the Section AI in Imaging)
Show Figures

Figure 1

27 pages, 4807 KB  
Article
LTPNet: Lesion-Aware Triple-Path Feature Fusion Network for Skin Lesion Segmentation
by Yange Sun, Sen Chen, Huaping Guo, Li Zhang, Hongzhou Yue and Yan Feng
J. Imaging 2026, 12(3), 93; https://doi.org/10.3390/jimaging12030093 - 24 Feb 2026
Viewed by 132
Abstract
Skin lesion segmentation has achieved notable progress in recent years; however, accurate delineation remains challenging due to complex backgrounds, ambiguous boundaries, and low lesion-to-skin contrast. To address these issues, we propose the lesion-aware triple-path feature fusion network (LTPNet), an end-to-end framework that progressively [...] Read more.
Skin lesion segmentation has achieved notable progress in recent years; however, accurate delineation remains challenging due to complex backgrounds, ambiguous boundaries, and low lesion-to-skin contrast. To address these issues, we propose the lesion-aware triple-path feature fusion network (LTPNet), an end-to-end framework that progressively processes features through extraction, refinement, and aggregation stages. In the extraction stage, we incorporate a general foreground–background attention to suppress background interference and accelerate model convergence. In the refinement stage, we introduce an attentive spatial modulator (ASM) to jointly exploit local structural cues and global semantic context for precise spatial modulation. We further develop a lesion-aware lite-gate attention (LALGA) module that performs local spatial feature modulation and global channel recalibration tailored to lesion characteristics. In the aggregation stage, we propose a triple-path feature fusion (TPFF) module that explicitly models feature relationships across scales via three complementary pathways: a common path (CP) for semantic consistency, a saliency path (SP) for highlighting co-activated regions, and a difference path (DP) for accentuating structural discrepancies. Extensive experiments on in-domain and cross-domain datasets show that LTPNet achieves superior segmentation accuracy with reasonable inference efficiency and model complexity, demonstrating its potential for efficient and reliable clinical decision support. Full article
(This article belongs to the Special Issue Computer Vision for Medical Image Analysis)
Show Figures

Figure 1

19 pages, 2747 KB  
Article
UVSegNet: Semantic Boundary-Aware Neural UV Parameterization for Man-Made Objects
by Hairun Zhang and Ying Song
J. Imaging 2026, 12(3), 92; https://doi.org/10.3390/jimaging12030092 - 24 Feb 2026
Viewed by 121
Abstract
UV parameterization is a fundamental step in building textured 3D models, but minimizing texture distortion and ensuring seams are placed along meaningful boundaries remains a challenge. This paper proposes UVSegNet, a novel semantic boundary-aware UV parameterization framework that combines part-level segmentation with geometry-aware [...] Read more.
UV parameterization is a fundamental step in building textured 3D models, but minimizing texture distortion and ensuring seams are placed along meaningful boundaries remains a challenge. This paper proposes UVSegNet, a novel semantic boundary-aware UV parameterization framework that combines part-level segmentation with geometry-aware parameterization. To address the common seam placement issues in parameterization, we introduce a boundary-aware guided UV mapping module that jointly optimizes geometric accuracy and seam layout. Furthermore, to better handle the cylindrical structures common in man-made objects, we introduce a cylindrical supervision strategy to reduce misalignment and unfolding distortion. Experiments on representative object categories show that UVSegNet outperforms other excellent baseline models in both texture quality and seam quality. Compared to Nuvo, UVSegNet improves the angular distortion (conformality) metric by 24.1% and seam compactness by 60.5% by generating a more compact seam layout. Experimental results demonstrate that UVSegNet outperforms baseline methods in both mapping quality and seam quality, thanks to the complementary mechanism of boundary constraints and geometry-driven modeling. Full article
(This article belongs to the Section Visualization and Computer Graphics)
Show Figures

Figure 1

22 pages, 4410 KB  
Article
Accelerating Point Cloud Computation via Memory in Embedded Structured Light Cameras
by Yanan Zhang, Shikang Meng, Shijie Wang and Yaheng Ren
J. Imaging 2026, 12(2), 91; https://doi.org/10.3390/jimaging12020091 - 21 Feb 2026
Viewed by 168
Abstract
Embedded structured light cameras have been widely applied in various fields. However, due to constraints such as insufficient computing resources, it remains difficult to achieve high-speed structured light point cloud computation. To address this issue, this study proposes a memory-driven computational framework for [...] Read more.
Embedded structured light cameras have been widely applied in various fields. However, due to constraints such as insufficient computing resources, it remains difficult to achieve high-speed structured light point cloud computation. To address this issue, this study proposes a memory-driven computational framework for accelerating point cloud computation. Specifically, the point cloud computation process is precomputed as much as possible and stored in memory in the form of parameters, thereby significantly reducing the computational load during actual point cloud computation. The framework is instantiated in two forms: a low-memory method that minimizes memory footprint at the expense of point cloud stability, and a high-memory method that preserves the nonlinear phase–distance relation via an extensive lookup table. Experimental evaluations demonstrate that the proposed methods achieve comparable accuracy to the conventional method while delivering substantial speedups, and data-format optimizations further reduce required bandwidth. This framework offers a generalizable paradigm for optimizing structured light pipelines, paving the way for enhanced real-time 3D sensing in embedded applications. Full article
Show Figures

Figure 1

25 pages, 3654 KB  
Article
MDF2Former: Multi-Scale Dual-Domain Feature Fusion Transformer for Hyperspectral Image Classification of Bacteria in Murine Wounds
by Decheng Wu, Wendan Liu, Rui Li, Xudong Fu, Lin Tao, Yinli Tian, Anqiang Zhang, Zhen Wang and Hao Tang
J. Imaging 2026, 12(2), 90; https://doi.org/10.3390/jimaging12020090 - 19 Feb 2026
Viewed by 129
Abstract
Bacterial wound infection poses a major challenge in trauma care and can lead to severe complications such as sepsis and organ failure. Therefore, rapid and accurate identification of the pathogen, along with targeted intervention, is of vital importance for improving treatment outcomes and [...] Read more.
Bacterial wound infection poses a major challenge in trauma care and can lead to severe complications such as sepsis and organ failure. Therefore, rapid and accurate identification of the pathogen, along with targeted intervention, is of vital importance for improving treatment outcomes and reducing risks. However, current detection methods are still constrained by procedural complexity and long processing times. In this study, a hyperspectral imaging (HSI) acquisition system for bacterial analysis and a multi-scale dual-domain feature fusion transformer (MDF2Former) were developed for classifying wound bacteria. MDF2Former integrates three modules: a multi-scale feature enhancement and fusion module that generates tokens with multi-scale discriminative representations, a spatial–spectral dual-branch attention module that strengthens joint feature modeling, and a frequency and spatial–spectral domain encoding module that captures global and local interactions among tokens through a hierarchical stacking structure, thereby enabling more efficient feature learning. Extensive experiments on our self-constructed HSI dataset of typical wound bacteria demonstrate that MDF2Former achieved outstanding performance across five metrics: Accuracy (91.94%), Precision (92.26%), Recall (91.94%), F1-score (92.01%), and Kappa coefficient (90.73%), surpassing all comparative models. These results have verified the effectiveness of combining HSI with deep learning for bacterial identification, and have highlighted its potential in assisting in the identification of bacterial species and making personalized treatment decisions for wound infections. Full article
(This article belongs to the Section Color, Multi-spectral, and Hyperspectral Imaging)
Show Figures

Figure 1

26 pages, 12208 KB  
Article
Classification of the Surrounding Rock Based on Image Processing Analysis and Transfer Learning
by Yanyun Fan, Jiaqi Zhu, Hua Luo, Yaxi Shen, Shuanglong Wang, Xiaoning Liu, Dong Li and Chuhan Deng
J. Imaging 2026, 12(2), 89; https://doi.org/10.3390/jimaging12020089 - 19 Feb 2026
Viewed by 262
Abstract
Currently, standardized classification methods of surrounding rock are relatively insufficient. The classification of surrounding rock mainly relies on the subjective judgment of technicians, leading to diverse evaluation results. This study focuses on the feature extraction and classification methods of surrounding rock images in [...] Read more.
Currently, standardized classification methods of surrounding rock are relatively insufficient. The classification of surrounding rock mainly relies on the subjective judgment of technicians, leading to diverse evaluation results. This study focuses on the feature extraction and classification methods of surrounding rock images in a certain tunnel of the Central Yunnan Water Diversion Project by using image processing analysis and transfer learning. Rich surrounding rock images and the water conservancy tunnel data are collected, and then the surrounding rock is classified relatively accurately according to the code and expert guidance. By introducing the fractal theory, the complexity and irregularity of the spatial distribution of weak layers and joints on the surrounding rock surface are revealed effectively. Based on the analysis of changes in fractal dimension characteristic values, a classification method for surrounding rock based on the fractal theory is proposed. Combined with the quantified parameters of surrounding rock images and the strength data collected by rebound meters, a method for correcting the surrounding rock strength based on image analysis is proposed, which can effectively solve the error caused by the uneven distribution of rock masses in the traditional rebound meter strength values. After correction, more accurate strength characteristics can be obtained, which is conducive to the standardized classification of the surrounding rock. After studying the recognition of tunnel surrounding rock images with transfer learning, a model is constructed to achieve rapid classification of tunnel surrounding rock. This research provides support for the standardized classification of tunnel surrounding rock. Full article
(This article belongs to the Section Image and Video Processing)
Show Figures

Figure 1

Back to TopTop