Sign in to use this feature.

Years

Between: -

Subjects

remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline

Journals

Article Types

Countries / Regions

Search Results (129)

Search Parameters:
Keywords = Gabor filter image

Order results
Result details
Results per page
Select all
Export citation of selected articles as:
21 pages, 1405 KB  
Article
Bionic Corner Detection Based on Cooperative Processing of Simple Cells and End-Stopped Cells
by Shuo Sun and Haiyang Yu
Algorithms 2026, 19(5), 343; https://doi.org/10.3390/a19050343 - 30 Apr 2026
Abstract
Corner detection is a fundamental task in computer vision that plays a critical role in applications such as image registration, 3D reconstruction, and object tracking. In biological visual systems, simple cells in the primary visual cortex exhibit high selectivity to edge stimuli of [...] Read more.
Corner detection is a fundamental task in computer vision that plays a critical role in applications such as image registration, 3D reconstruction, and object tracking. In biological visual systems, simple cells in the primary visual cortex exhibit high selectivity to edge stimuli of specific orientations, while end-stopped cells can detect geometric singular structures such as line segment endpoints and corners. Existing corner detection methods based on visual neural computation typically employ a strategy of densely distributed end-stopped cells for corner localization, which suffers from significant localization deviation under small angle conditions due to mutual interference between responses of adjacent neurons. To address this problem, this paper proposes a bionic corner detection method based on cooperative processing of simple cells and end-stopped cells. The method constructs a two-stage cooperative processing framework: the edge filtering stage employs a Gabor filter bank to simulate the orientation selectivity of simple cells, extracting edge positions and orientation information; the dynamic construction stage builds unilateral end-stopped cells only at filtered edge positions based on local orientation information, fundamentally avoiding computational redundancy and response interference caused by global dense distribution; the corner localization stage determines precise corner coordinates through hierarchical clustering and dual-cluster centroid fusion strategies. Experimental results demonstrate that, in the 15° acute-angle regime where dense end-stopped schemes are most severely affected by response interference, the proposed method reduces the mean localization error from 8.76 to 2.34 pixels, corresponding to a 73.3% improvement; averaged across the eight tested angle levels from 15° to 165°, the improvement is approximately 40.9%, and all per-angle differences are statistically significant (paired t-test, p < 0.01 or below, N = 10 independent runs). On standard test images, the method attains the lowest mean localization error among the eight compared detectors (1.58 pixels, versus 1.68–3.42 pixels for Harris, FAST, COSFIRE, KAZE, SuperPoint, Deep Corner, and Wei et al.), while maintaining competitive detection rate, false-alarm rate, and runtime. Physiological plausibility validation experiments show that the correlation coefficient between the detection deviation of this method and human perceptual deviation reaches 0.923, indicating that the output of the framework aligns with previously reported human perceptual bias patterns and supporting its biological plausibility as a biologically inspired—rather than mechanistic—model of corner perception. The source code, dataset, and experimental results are publicly available (see Data Availability Statement). Full article
Show Figures

Figure 1

27 pages, 6783 KB  
Article
A Robust Intelligent CNN Model Enhanced with Gabor-Based Feature Extraction, SMOTE Balancing, and Adam Optimization for Multi-Grade Diabetic Retinopathy Classification
by Asri Mulyani, Muljono, Purwanto and Moch Arief Soeleman
J. Imaging 2026, 12(5), 188; https://doi.org/10.3390/jimaging12050188 - 27 Apr 2026
Viewed by 206
Abstract
Diabetic retinopathy (DR) is a leading cause of vision impairment and permanent blindness worldwide, requiring accurate and automated systems for multi-grade severity classification. However, standard Convolutional Neural Networks (CNNs) often struggle to capture fine, high-frequency microvascular patterns critical for diagnosis. This study proposes [...] Read more.
Diabetic retinopathy (DR) is a leading cause of vision impairment and permanent blindness worldwide, requiring accurate and automated systems for multi-grade severity classification. However, standard Convolutional Neural Networks (CNNs) often struggle to capture fine, high-frequency microvascular patterns critical for diagnosis. This study proposes a Robust Intelligent CNN Model (RICNN) that integrates Gabor-based feature extraction with deep learning to improve DR classification. Specifically, Gabor filters are applied during preprocessing to extract orientation- and frequency-sensitive texture features, which are transformed into feature maps and concatenated with CNN feature representations at the fully connected layer (feature-level fusion). The model also incorporates the Synthetic Minority Oversampling Technique (SMOTE) for data balancing and the Adam optimizer for efficient convergence. This integration enhances sensitivity to microvascular structures such as microaneurysms and hemorrhages. The proposed RICNN was evaluated on the Messidor dataset (1200 images) across four severity levels: Mild, Moderate, Severe, and Proliferative DR. The model achieved an accuracy of 89%, a precision of 88.75%, a recall of 89%, and an F1-score of 89%, with AUCs of 97% for Severe DR and 99% for Proliferative DR. Comparative analysis confirms that the proposed texture-aware Gabor enhancement significantly outperforms LBP and Color Histogram approaches, indicating its potential for reliable clinical decision support. Full article
(This article belongs to the Section Medical Imaging)
Show Figures

Figure 1

13 pages, 750 KB  
Article
Evaluating Handcrafted Image Descriptors for Defect Detection in the X-Ray Inspection of Turbine Blade Castings: A Feature Separability Study
by Andrzej Burghardt and Wojciech Łabuński
Appl. Sci. 2026, 16(8), 3905; https://doi.org/10.3390/app16083905 - 17 Apr 2026
Viewed by 163
Abstract
The industrial X-ray inspection of turbine blade castings requires reliable and auditable decision support, yet defect indications are subtle, and data availability is limited. This study quantitatively assesses the diagnostic potential of handcrafted image descriptors by evaluating class separability in feature space, independently [...] Read more.
The industrial X-ray inspection of turbine blade castings requires reliable and auditable decision support, yet defect indications are subtle, and data availability is limited. This study quantitatively assesses the diagnostic potential of handcrafted image descriptors by evaluating class separability in feature space, independently of any trained classifier. The dataset comprises 1600 16-bit DICOM radiograms of 200 blades (eight views per blade), including 156 defective images with 207 localized defects. Standardized 32 × 32 ROI patches were sampled randomly in the vicinity of indications and from defect-free regions to reduce sample correlation and to emulate localization uncertainty. Feature vectors were extracted using five descriptor families—first-order statistics, GLCM/Haralick, FFT and wavelet (DWT) features, Gabor filters, and LBP—and the standardized z-score. Separability was ranked using complementary distribution-based and distance-based metrics grouped into three sets, and the results were min–max-normalized to enable TOP-5 comparisons. Spectral descriptors, particularly DWT wavelets and FFT combined with DWT, consistently achieved the highest scores in distributional metrics, supporting a lightweight screening profile. In contrast, richer combinations dominated multidimensional geometric metrics, indicating benefits from multi-perspective representations for offline analysis. The proposed metric-driven framework provides an interpretable basis for representation selection prior to classifier development under industrial constraints. Full article
Show Figures

Figure 1

25 pages, 15197 KB  
Article
Semi-Automated Computational Identification of Fibrosis for Enhanced Histopathological Decision Support
by Alexandru-George Berciu, Diana Rus-Gonciar, Teodora Mocan, Lucia Agoston-Coldea, Carmen Cionca and Eva-Henrietta Dulf
J. Imaging 2026, 12(4), 152; https://doi.org/10.3390/jimaging12040152 - 31 Mar 2026
Viewed by 315
Abstract
Myocardial fibrosis is a critical prognostic marker involving a progressive cascade of pathological conditions. Accurate assessment of fibrosis in myocardial samples is a routine but difficult procedure for pathologists. This article presents a semi-automated system designed to ease this task while providing pixel-level [...] Read more.
Myocardial fibrosis is a critical prognostic marker involving a progressive cascade of pathological conditions. Accurate assessment of fibrosis in myocardial samples is a routine but difficult procedure for pathologists. This article presents a semi-automated system designed to ease this task while providing pixel-level accuracy that exceeds manual estimation capabilities. The proposed innovative approach combines Gabor filters with CIELAB color space analysis to ensure the efficiency and interpretability of calculations. Testing on histopathological samples, differentiating between fibrous, healthy, and variant tissues, yielded a promising accuracy of 87.5% for images with fibrosis and 80% for all 45 images tested. This system successfully establishes a solid foundation for automated diagnosis, providing pathologists with a reliable and highly accurate tool for quantitative analysis of cardiac tissue. Full article
(This article belongs to the Section AI in Imaging)
Show Figures

Figure 1

24 pages, 8810 KB  
Article
FreqPose: Frequency-Aware Diffusion with Fractional Gabor Filters and Global Pose–Semantic Alignment
by Meng Wang, Bing Wang, Huiling Chen, Jing Ren and Xueping Tang
Sensors 2026, 26(4), 1334; https://doi.org/10.3390/s26041334 - 19 Feb 2026
Viewed by 403
Abstract
The task of pose-guided person image generation has long been confronted with two major challenges: high-frequency texture details tend to blur and be lost during appearance transfer, while the semantic identity of the person is difficult to maintain consistently during pose changes. To [...] Read more.
The task of pose-guided person image generation has long been confronted with two major challenges: high-frequency texture details tend to blur and be lost during appearance transfer, while the semantic identity of the person is difficult to maintain consistently during pose changes. To address these issues, this paper proposes a diffusion-based generative framework that integrates frequency awareness and global semantic alignment. The framework consists of two core modules: a multi-level fractional-order Gabor frequency-aware network, which accurately extracts and reconstructs high-frequency texture features such as hair strands and fabric wrinkles, enhances image detail fidelity through fractional-order filtering and complex domain modeling; and a global semantic-pose alignment module that utilizes a cross-modal attention mechanism to establish a global mapping between pose features and appearance semantics, ensuring pose-driven semantic alignment and appearance consistency. The collaborative function of these two modules ensures that the generated results maintain structural integrity and natural textures even under complex pose variations and large-angle rotations. The experimental results on the DeepFashion and Market1501 datasets demonstrate that the proposed method outperforms existing state-of-the-art approaches in terms of SSIM, FID, and perceptual quality, validating the effectiveness of the model in enhancing texture fidelity and semantic consistency. Full article
(This article belongs to the Section Intelligent Sensors)
Show Figures

Figure 1

14 pages, 421 KB  
Article
Artificial Intelligence-Based Evaluation of Permanent First Molar Extraction Indications in Children Using Panoramic Radiographs
by Serap Gülçin Çetin, Ömer Faruk Ertuğrul, Nursezen Kavasoğlu and Veysel Eratilla
Children 2026, 13(2), 277; https://doi.org/10.3390/children13020277 - 17 Feb 2026
Viewed by 467
Abstract
Background: The aim of this study was to develop an artificial intelligence (AI)-based decision support model for evaluating the extraction indication of permanent first molars in pediatric patients using panoramic radiographs, and to investigate the potential contribution of this model to the clinical [...] Read more.
Background: The aim of this study was to develop an artificial intelligence (AI)-based decision support model for evaluating the extraction indication of permanent first molars in pediatric patients using panoramic radiographs, and to investigate the potential contribution of this model to the clinical decision-making process. Methods: This retrospective observational study analyzed 1000 panoramic radiographs obtained from children aged 8–10 years who attended the Clinics of Batman University Faculty of Dentistry for routine dental examination. Among the radiographs meeting the inclusion criteria, a total of 176 panoramic images were selected based on dental maturation assessment using the Demirjian tooth development staging system. Cases in which the permanent second molar was classified as Demirjian stages E–F were labeled as “extraction indication present”, while the remaining stages were labeled as “extraction indication absent”. A balanced dataset was created, consisting of 88 cases in each group. Image features were extracted using Gabor filters and Histogram of Oriented Gradients (HOG). The selected features were analyzed using a Support Vector Machine (SVM) classifier with a radial basis function (RBF) kernel. Model performance was evaluated using accuracy, sensitivity, specificity, F1-score, and area under the receiver operating characteristic curve (ROC–AUC). Results: The proposed Gabor–HOG–SVM-based AI model achieved an overall classification accuracy of 77.78% with an AUC value of 0.77 in distinguishing between “extraction indication present” and “extraction indication absent” cases. For the extraction-indicated group, the sensitivity was 0.81 and the F1-score was 0.79, whereas for the non-indicated group, the sensitivity and F1-score were 0.74 and 0.77, respectively. No statistically significant differences were observed between the groups in terms of age or sex distribution (p > 0.05). Conclusions: This study demonstrates that artificial intelligence-based analysis of panoramic radiographic images can provide an objective and reproducible decision support approach for evaluating extraction indications of permanent first molars in pediatric patients. The proposed model should be considered as an adjunctive tool to reduce observer-dependent variability rather than a replacement for clinical judgment, and its clinical applicability should be further validated through multicenter and multi-parametric studies. Full article
(This article belongs to the Section Pediatric Dentistry & Oral Medicine)
Show Figures

Figure 1

19 pages, 2617 KB  
Article
Topic-Modeling Guided Semantic Clustering for Enhancing CNN-Based Image Classification Using Scale-Invariant Feature Transform and Block Gabor Filtering
by Natthaphong Suthamno and Jessada Tanthanuch
J. Imaging 2026, 12(2), 70; https://doi.org/10.3390/jimaging12020070 - 9 Feb 2026
Viewed by 447
Abstract
This study proposes a topic-modeling guided framework that enhances image classification by introducing semantic clustering prior to CNN training. Images are processed through two key-point extraction pipelines: Scale-Invariant Feature Transform (SIFT) with Sobel edge detection and Block Gabor Filtering (BGF), to obtain local [...] Read more.
This study proposes a topic-modeling guided framework that enhances image classification by introducing semantic clustering prior to CNN training. Images are processed through two key-point extraction pipelines: Scale-Invariant Feature Transform (SIFT) with Sobel edge detection and Block Gabor Filtering (BGF), to obtain local feature descriptors. These descriptors are clustered using K-means to build a visual vocabulary. Bag of Words histograms then represent each image as a visual document. Latent Dirichlet Allocation is applied to uncover latent semantic topics, generating coherent image clusters. Cluster-specific CNN models, including AlexNet, GoogLeNet, and several ResNet variants, are trained under identical conditions to identify the most suitable architecture for each cluster. Two topic guided integration strategies, the Maximum Proportion Topic (MPT) and the Weight Proportion Topic (WPT), are then used to assign test images to the corresponding specialized model. Experimental results show that both the SIFT-based and BGF-based pipelines outperform non-clustered CNN models and a baseline method using Incremental PCA, K-means, Same-Cluster Prediction, and unweighted Ensemble Voting. The SIFT pipeline achieves the highest accuracy of 95.24% with the MPT strategy, while the BGF pipeline achieves 93.76% with the WPT strategy. These findings confirm that semantic structure introduced through topic modeling substantially improves CNN classification performance. Full article
(This article belongs to the Topic Machine Learning and Deep Learning in Medical Imaging)
Show Figures

Figure 1

24 pages, 5019 KB  
Article
A Dual Stream Deep Learning Framework for Alzheimer’s Disease Detection Using MRI Sonification
by Nadia A. Mohsin and Mohammed H. Abdul Ameer
J. Imaging 2026, 12(1), 46; https://doi.org/10.3390/jimaging12010046 - 15 Jan 2026
Viewed by 573
Abstract
Alzheimer’s Disease (AD) is an advanced brain illness that affects millions of individuals across the world. It causes gradual damage to the brain cells, leading to memory loss and cognitive dysfunction. Although Magnetic Resonance Imaging (MRI) is widely used in AD diagnosis, the [...] Read more.
Alzheimer’s Disease (AD) is an advanced brain illness that affects millions of individuals across the world. It causes gradual damage to the brain cells, leading to memory loss and cognitive dysfunction. Although Magnetic Resonance Imaging (MRI) is widely used in AD diagnosis, the existing studies rely solely on the visual representations, leaving alternative features unexplored. The objective of this study is to explore whether MRI sonification can provide complementary diagnostic information when combined with conventional image-based methods. In this study, we propose a novel dual-stream multimodal framework that integrates 2D MRI slices with their corresponding audio representations. MRI images are transformed into audio signals using a multi-scale, multi-orientation Gabor filtering, followed by a Hilbert space-filling curve to preserve spatial locality. The image and sound modalities are processed using a lightweight CNN and YAMNet, respectively, then fused via logistic regression. The experimental results of the multimodal achieved the highest accuracy in distinguishing AD from Cognitively Normal (CN) subjects at 98.2%, 94% for AD vs. Mild Cognitive Impairment (MCI), and 93.2% for MCI vs. CN. This work provides a new perspective and highlights the potential of audio transformation of imaging data for feature extraction and classification. Full article
(This article belongs to the Section AI in Imaging)
Show Figures

Figure 1

19 pages, 2960 KB  
Article
Gabor Transform-Based Deep Learning System Using CNN for Melanoma Detection
by S. Deivasigamani, C. Senthilpari, Siva Sundhara Raja. D, A. Thankaraj, G. Narmadha and K. Gowrishankar
Computers 2026, 15(1), 54; https://doi.org/10.3390/computers15010054 - 13 Jan 2026
Viewed by 592
Abstract
Melanoma is highly dangerous and can spread rapidly to other parts of the body. It has an increasing fatality rate among different types of cancer. Timely detection of skin malignancies can reduce overall mortality. Therefore, clinical screening methods require more time and accuracy [...] Read more.
Melanoma is highly dangerous and can spread rapidly to other parts of the body. It has an increasing fatality rate among different types of cancer. Timely detection of skin malignancies can reduce overall mortality. Therefore, clinical screening methods require more time and accuracy for diagnosis. An automated, computer-aided system would facilitate earlier melanoma detection, thereby increasing patient survival rates. This paper identifies melanoma images using a Convolutional Neural Network. Skin images are preprocessed using Histogram Equalization and Gabor transforms. A Gabor filter-based Convolutional Neural Network (CNN) classifier trains and classifies the extracted features. We adopt Gabor filters because they are bandpass filters that transform a pixel into a multi-resolution kernel matrix, providing detailed information about the image. This study suggests a method with accuracy, sensitivity, and specificity of 98.58%, 98.66%, and 98.75%, respectively. This research supports SDGs 3 and 4 by facilitating early melanoma detection and enhancing AI-driven medical education. Full article
Show Figures

Figure 1

26 pages, 6100 KB  
Article
A New Change Detection Method for Heterogeneous Remote Sensing Images Via an Automatic Differentiable Adversarial Search
by Hui Li, Jing Liu, Yan Zhang, Jie Chen, Hongcheng Zeng, Wei Yang, Jie Chen, Zhixiang Huang and Long Sun
Remote Sens. 2026, 18(1), 94; https://doi.org/10.3390/rs18010094 - 26 Dec 2025
Cited by 1 | Viewed by 1116
Abstract
Heterogeneous remote sensing image change detection (Hete-CD) holds significant research value in military and civilian fields. The existing methods often rely on expert experience to design fixed deep network architectures for cross-modal feature alignment and fusion purposes. However, when faced with diverse land [...] Read more.
Heterogeneous remote sensing image change detection (Hete-CD) holds significant research value in military and civilian fields. The existing methods often rely on expert experience to design fixed deep network architectures for cross-modal feature alignment and fusion purposes. However, when faced with diverse land cover types, these methods often lead to blurred change boundaries and structural distortions, resulting in significant performance degradations. To address this, we propose an adaptive adversarial learning-based heterogeneous remote sensing image change detection method based on the differentiable filter combination search (DFCS) strategy to provide enhanced generalizability and dynamic learning capabilities for diverse scenarios. First, a fully reconfigurable self-learning discriminator is designed to dynamically synthesize the optimal convolutional architecture from a library of atomic filters containing basic operators. This provides highly adaptive adversarial supervision to the generator, enabling joint dynamic learning between the generator and discriminator. To further mitigate modality differences in the input stage, we integrate a feature fusion module based on the Gabor and local normalized cross-correlation (G-LNCC) to extract modality-invariant texture and structure features. Finally, a geometric structure-based collaborative supervision (GSCS) loss function is constructed to impose fine-grained constraints on the change map from the perspectives of regions, boundaries, and structures, thereby enforcing physical properties. Comparative experimental results obtained on five public Hete-CD datasets show that our method achieves the best F1 values and overall accuracy levels, especially on the Gloucester I and Gloucester II datasets, achieving F1 scores of 93.7% and 95.0%, respectively, demonstrating the strong generalizability of our method in complex scenarios. Full article
Show Figures

Figure 1

23 pages, 4279 KB  
Article
DCT Underwater Image Enhancement Based on Attenuation Analysis
by Leyuan Wang, Miao Yang, Can Pan and Jiaju Tao
Sensors 2025, 25(23), 7192; https://doi.org/10.3390/s25237192 - 25 Nov 2025
Viewed by 910
Abstract
Underwater images often suffer from color distortion, reduced contrast, and blurred details due to the selective absorption and scattering of light by water, which limits the performance of underwater visual tasks. To address these issues, this paper proposes an underwater image enhancement method [...] Read more.
Underwater images often suffer from color distortion, reduced contrast, and blurred details due to the selective absorption and scattering of light by water, which limits the performance of underwater visual tasks. To address these issues, this paper proposes an underwater image enhancement method that integrates multi-channel attenuation analysis and discrete cosine transform (DCT). First, the color statistics of an in situ-captured underwater image are mapped to those of a reference image selected from a well-illuminated natural image dataset with standard color distribution; no pristine underwater image is required. This mapping yields a color transfer image, i.e., an intermediate color-corrected result obtained via statistical matching. Subsequently, this image is fused with an attenuation weight map and the original input to produce the final color-corrected result. Secondly, taking advantage of the median’s resistance to extreme value interference and the Sigmoid function’s flexible control of gray-scale transformation, the gray-scale range is adjusted in different regions through nonlinear mapping to achieve global contrast balance. Finally, considering the visual system’s sensitivity to high-frequency details, a saliency map is extracted using Gabor filtering, and the frequency characteristics are analyzed through block DCT transformation. Adaptive gain is applied to high-frequency details to enhance them. Experiments were conducted on the UIEB, EUVP, and LSUI datasets and compared with existing methods. Through qualitative and quantitative analysis, it was verified that the proposed algorithm not only effectively enhances underwater images but also significantly improves image clarity. Full article
Show Figures

Figure 1

14 pages, 5908 KB  
Article
A Novel Multi-Source Image Registration of Porcine Body for Multi-Feature Detection
by Zhen Zhong and Shengfei Zhi
Sensors 2025, 25(22), 6918; https://doi.org/10.3390/s25226918 - 12 Nov 2025
Cited by 1 | Viewed by 789
Abstract
The safety of animal-related agricultural products has been a hot issue. To obtain a multi-feature representation of porcine bodies for detecting their health, visible and infrared imaging is valuable for exploiting multiple images of a porcine body from different modalities. However, the direct [...] Read more.
The safety of animal-related agricultural products has been a hot issue. To obtain a multi-feature representation of porcine bodies for detecting their health, visible and infrared imaging is valuable for exploiting multiple images of a porcine body from different modalities. However, the direct registration of visible and infrared porcine body images can easily cause the dislocation of structural information and spatial position, due to different resolutions and spectrums of multi-source images. To overcome the problem, a novel multi-source image feature representation method based on contour angle orientation is proposed and named Gabor-Ordinal-based Contour Angle Orientation (GOCAO). Moreover, a visible and infrared porcine body image registration method is described and named GOCAO-Rough to Fine (GOCAO-R2F). First, contour and texture features of the porcine body are acquired using a Gabor filter with variable scales and an ordinal operation. Second, feature points in contours are obtained by curvature scale space (CSS), and the main orientation of each feature point is determined by GOCAO. Third, modified scale-invariant feature transform (MSIFT) features are received on the main orientation and registered with bilateral matching. Finally, accurate registrations are extracted by R2F. Experimental results show that the proposed registration algorithm accurately matches multi-source images for porcine body multi-feature detection and is capable of achieving lower average root-mean-square error than current registration algorithms. Full article
Show Figures

Figure 1

15 pages, 8859 KB  
Article
A Hybrid Estimation Model for Graphite Nodularity of Ductile Cast Iron Based on Multi-Source Feature Extraction
by Yongjian Yang, Yanhui Liu, Yuqian He, Zengren Pan and Zhiwei Li
Modelling 2025, 6(4), 126; https://doi.org/10.3390/modelling6040126 - 13 Oct 2025
Viewed by 835
Abstract
Graphite nodularity is a key indicator for evaluating the microstructure quality of ductile iron and plays a crucial role in ensuring product quality and enhancing manufacturing efficiency. Existing research often only focuses on a single type of feature and fails to utilize multi-source [...] Read more.
Graphite nodularity is a key indicator for evaluating the microstructure quality of ductile iron and plays a crucial role in ensuring product quality and enhancing manufacturing efficiency. Existing research often only focuses on a single type of feature and fails to utilize multi-source information in a coordinated manner. Single-feature methods are difficult to comprehensively capture microstructures, which limits the accuracy and robustness of the model. This study proposes a hybrid estimation model for the graphite nodularity of ductile cast iron based on multi-source feature extraction. A comprehensive feature engineering pipeline was established, incorporating geometric, color, and texture features extracted via Hue-Saturation-Value color space (HSV) histograms, gray level co-occurrence matrix (GLCM), Local Binary Pattern (LBP), and multi-scale Gabor filters. Dimensionality reduction was performed using Principal Component Analysis (PCA) to mitigate redundancy. An improved watershed algorithm combined with intelligent filtering was used for accurate particle segmentation. Several machine learning algorithms, including Support Vector Regression (SVR), Multi-Layer Perceptron (MLP), Random Forest (RF), Gradient Boosting Regressor (GBR), eXtreme Gradient Boosting (XGBoost) and Categorical Boosting (CatBoost), are applied to estimate graphite nodularity based on geometric features (GFs) and feature extraction. Experimental results demonstrate that the CatBoost model trained on fused features achieves high estimation accuracy and stability for geometric parameters, with R-squared (R2) exceeding 0.98. Furthermore, introducing geometric features into the fusion set enhances model generalization and suppresses overfitting. This framework offers an efficient and robust approach for intelligent analysis of metallographic images and provides valuable support for automated quality assessment in casting production. Full article
Show Figures

Figure 1

14 pages, 1202 KB  
Article
Optimization of Gabor Convolutional Networks Using the Taguchi Method and Their Application in Wood Defect Detection
by Ming-Feng Yeh, Ching-Chuan Luo and Yu-Cheng Liu
Appl. Sci. 2025, 15(17), 9557; https://doi.org/10.3390/app15179557 - 30 Aug 2025
Cited by 1 | Viewed by 1084
Abstract
Automated optical inspection (AOI) of wood surfaces is critical for ensuring product quality in the furniture and manufacturing industries; however, existing defect detection systems often struggle to generalize across complex grain patterns and diverse defect types. This study proposes a wood defect recognition [...] Read more.
Automated optical inspection (AOI) of wood surfaces is critical for ensuring product quality in the furniture and manufacturing industries; however, existing defect detection systems often struggle to generalize across complex grain patterns and diverse defect types. This study proposes a wood defect recognition model employing a Gabor Convolutional Network (GCN) that integrates convolutional neural networks (CNNs) with Gabor filters. To systematically optimize the network’s architecture and improve both detection accuracy and computational efficiency, the Taguchi method is employed to tune key hyperparameters, including convolutional kernel size, filter number, and Gabor parameters (frequency, orientation, and phase offset). Additionally, image tiling and augmentation techniques are employed to effectively increase the training dataset, thereby enhancing the model’s stability and accuracy. Experiments conducted on the MVTec Anomaly Detection dataset (wood category) demonstrate that the Taguchi-optimized GCN achieves an accuracy of 98.92%, outperforming a baseline Taguchi-optimized CNN by 2.73%. Results confirm that Taguchi-optimized GCNs enhance defect detection performance and computational efficiency, making them valuable for smart manufacturing. Full article
(This article belongs to the Topic Electronic Communications, IOT and Big Data, 2nd Volume)
Show Figures

Figure 1

17 pages, 588 KB  
Article
An Accurate and Efficient Diabetic Retinopathy Diagnosis Method via Depthwise Separable Convolution and Multi-View Attention Mechanism
by Qing Yang, Ying Wei, Fei Liu and Zhuang Wu
Appl. Sci. 2025, 15(17), 9298; https://doi.org/10.3390/app15179298 - 24 Aug 2025
Cited by 1 | Viewed by 1349
Abstract
Diabetic retinopathy (DR), a critical ocular disease that can lead to blindness, demands early and accurate diagnosis to prevent vision loss. Current automated DR diagnosis methods face two core challenges: first, subtle early lesions such as microaneurysms are often missed due to insufficient [...] Read more.
Diabetic retinopathy (DR), a critical ocular disease that can lead to blindness, demands early and accurate diagnosis to prevent vision loss. Current automated DR diagnosis methods face two core challenges: first, subtle early lesions such as microaneurysms are often missed due to insufficient feature extraction; second, there is a persistent trade-off between model accuracy and efficiency—lightweight architectures often sacrifice precision for real-time performance, while high-accuracy models are computationally expensive and difficult to deploy on resource-constrained edge devices. To address these issues, this study presents a novel deep learning framework integrating depthwise separable convolution and a multi-view attention mechanism (MVAM) for efficient DR diagnosis using retinal images. The framework employs multi-scale feature fusion via parallel 3 × 3 and 5 × 5 convolutions to capture lesions of varying sizes and incorporates Gabor filters to enhance vascular texture and directional lesion modeling, improving sensitivity to early structural abnormalities while reducing computational costs. Experimental results on both the diabetic retinopathy (DR) dataset and ocular disease (OD) dataset demonstrate the superiority of the proposed method: it achieves a high accuracy of 0.9697 on the DR dataset and 0.9669 on the OD dataset, outperforming traditional methods such as CNN_eye, VGG, and UNet by more than 1 percentage point. Moreover, its training time is only half that of U-Net (on DR dataset) and VGG (on OD dataset), highlighting its potential for clinical DR screening. Full article
Show Figures

Figure 1

Back to TopTop