Sign in to use this feature.

Years

Between: -

Subjects

remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline

Journals

Article Types

Countries / Regions

Search Results (45)

Search Parameters:
Keywords = VeinCNN

Order results
Result details
Results per page
Select all
Export citation of selected articles as:
22 pages, 4399 KiB  
Article
Deep Learning-Based Fingerprint–Vein Biometric Fusion: A Systematic Review with Empirical Evaluation
by Sarah Almuwayziri, Abeer Al-Nafjan, Hessah Aljumah and Mashael Aldayel
Appl. Sci. 2025, 15(15), 8502; https://doi.org/10.3390/app15158502 - 31 Jul 2025
Viewed by 412
Abstract
User authentication is crucial for safeguarding access to digital systems and services. Biometric authentication serves as a strong and user-friendly alternative to conventional security methods such as passwords and PINs, which are often susceptible to breaches. This study proposes a deep learning-based multimodal [...] Read more.
User authentication is crucial for safeguarding access to digital systems and services. Biometric authentication serves as a strong and user-friendly alternative to conventional security methods such as passwords and PINs, which are often susceptible to breaches. This study proposes a deep learning-based multimodal biometric system that combines fingerprint (FP) and finger vein (FV) modalities to improve accuracy and security. The system explores three fusion strategies: feature-level fusion (combining feature vectors from each modality), score-level fusion (integrating prediction scores from each modality), and a hybrid approach that leverages both feature and score information. The implementation involved five pretrained convolutional neural network (CNN) models: two unimodal (FP-only and FV-only) and three multimodal models corresponding to each fusion strategy. The models were assessed using the NUPT-FPV dataset, which consists of 33,600 images collected from 140 subjects with a dual-mode acquisition device in varied environmental conditions. The results indicate that the hybrid-level fusion with a dominant score weight (0.7 score, 0.3 feature) achieved the highest accuracy (99.79%) and the lowest equal error rate (EER = 0.0018), demonstrating superior robustness. Overall, the results demonstrate that integrating deep learning with multimodal fusion is highly effective for advancing scalable and accurate biometric authentication solutions suitable for real-world deployments. Full article
Show Figures

Figure 1

26 pages, 9987 KiB  
Article
Detection of Citrus Huanglongbing in Natural Field Conditions Using an Enhanced YOLO11 Framework
by Liang Cao, Wei Xiao, Zeng Hu, Xiangli Li and Zhongzhen Wu
Mathematics 2025, 13(14), 2223; https://doi.org/10.3390/math13142223 - 8 Jul 2025
Viewed by 602
Abstract
Citrus Huanglongbing (HLB) is one of the most devastating diseases in the global citrus industry, but its early detection under complex field conditions remains a major challenge. Existing methods often suffer from insufficient dataset diversity and poor generalization, and struggle to accurately detect [...] Read more.
Citrus Huanglongbing (HLB) is one of the most devastating diseases in the global citrus industry, but its early detection under complex field conditions remains a major challenge. Existing methods often suffer from insufficient dataset diversity and poor generalization, and struggle to accurately detect subtle early-stage lesions and multiple HLB symptoms in natural backgrounds. To address these issues, we propose an enhanced YOLO11-based framework, DCH-YOLO11. We constructed a multi-symptom HLB leaf dataset (MS-HLBD) containing 9219 annotated images across five classes: Healthy (1862), HLB blotchy mottling (2040), HLB Zinc deficiency (1988), HLB yellowing (1768), and Canker (1561), collected under diverse field conditions. To improve detection performance, the DCH-YOLO11 framework incorporates three novel modules: the C3k2 Dynamic Feature Fusion (C3k2_DFF) module, which enhances early and subtle lesion detection through dynamic feature fusion; the C2PSA Context Anchor Attention (C2PSA_CAA) module, which leverages context anchor attention to strengthen feature extraction in complex vein regions; and the High-efficiency Dynamic Feature Pyramid Network (HDFPN) module, which optimizes multi-scale feature interaction to boost detection accuracy across different object sizes. On the MS-HLBD dataset, DCH-YOLO11 achieved a precision of 91.6%, recall of 87.1%, F1-score of 89.3, and mAP50 of 93.1%, surpassing Faster R-CNN, SSD, RT-DETR, YOLOv7-tiny, YOLOv8n, YOLOv9-tiny, YOLOv10n, YOLO11n, and YOLOv12n by 13.6%, 8.8%, 5.3%, 3.2%, 2.0%, 1.6%, 2.6%, 1.8%, and 1.6% in mAP50, respectively. On a publicly available citrus HLB dataset, DCH-YOLO11 achieved a precision of 82.7%, recall of 81.8%, F1-score of 82.2, and mAP50 of 89.4%, with mAP50 improvements of 8.9%, 4.0%, 3.8%, 3.2%, 4.7%, 3.2%, and 3.4% over RT-DETR, YOLOv7-tiny, YOLOv8n, YOLOv9-tiny, YOLOv10n, YOLO11n, and YOLOv12n, respectively. These results demonstrate that DCH-YOLO11 achieves both state-of-the-art accuracy and excellent generalization, highlighting its strong potential for robust and practical citrus HLB detection in real-world applications. Full article
(This article belongs to the Special Issue Deep Learning and Adaptive Control, 3rd Edition)
Show Figures

Figure 1

28 pages, 1737 KiB  
Article
Finger Vein Recognition Based on Unsupervised Spiking Convolutional Neural Network with Adaptive Firing Threshold
by Li Yang, Qiong Yao and Xiang Xu
Sensors 2025, 25(7), 2279; https://doi.org/10.3390/s25072279 - 3 Apr 2025
Viewed by 513
Abstract
Currently, finger vein recognition (FVR) stands as a pioneering biometric technology, with convolutional neural networks (CNNs) and Transformers, among other advanced deep neural networks (DNNs), consistently pushing the boundaries of recognition accuracy. Nevertheless, these DNNs are inherently characterized by static, continuous-valued neuron activations, [...] Read more.
Currently, finger vein recognition (FVR) stands as a pioneering biometric technology, with convolutional neural networks (CNNs) and Transformers, among other advanced deep neural networks (DNNs), consistently pushing the boundaries of recognition accuracy. Nevertheless, these DNNs are inherently characterized by static, continuous-valued neuron activations, necessitating intricate network architectures and extensive parameter training to enhance performance. To address these challenges, we introduce an adaptive firing threshold-based spiking neural network (ATSNN) for FVR. ATSNN leverages discrete spike encodings to transforms static finger vein images into spike trains with spatio-temporal dynamic features. Initially, Gabor and difference of Gaussian (DoG) filters are employed to convert image pixel intensities into spike latency encodings. Subsequently, these spike encodings are fed into the ATSNN, where spiking features are extracted using biologically plausible local learning rules. Our proposed ATSNN dynamically adjusts the firing thresholds of neurons based on average potential tensors, thereby enabling adaptive modulation of the neuronal input-output response and enhancing network robustness. Ultimately, the spiking features with the earliest emission times are retained and utilized for classifier training via a support vector machine (SVM). Extensive experiments conducted across three benchmark finger vein datasets reveal that our ATSNN model not only achieves remarkable recognition accuracy but also excels in terms of reduced parameter count and model complexity, surpassing several existing FVR methods. Furthermore, the sparse and event-driven nature of our ATSNN renders it more biologically plausible compared to traditional DNNs. Full article
(This article belongs to the Section Biosensors)
Show Figures

Figure 1

30 pages, 7517 KiB  
Article
MixCFormer: A CNN–Transformer Hybrid with Mixup Augmentation for Enhanced Finger Vein Attack Detection
by Zhaodi Wang, Shuqiang Yang, Huafeng Qin, Yike Liu and Junqiang Wang
Electronics 2025, 14(2), 362; https://doi.org/10.3390/electronics14020362 - 17 Jan 2025
Cited by 2 | Viewed by 1323
Abstract
Finger vein recognition has gained significant attention for its importance in enhancing security, safeguarding privacy, and ensuring reliable liveness detection. As a foundation of vein recognition systems, vein detection faces challenges, including low feature extraction efficiency, limited robustness, and a heavy reliance on [...] Read more.
Finger vein recognition has gained significant attention for its importance in enhancing security, safeguarding privacy, and ensuring reliable liveness detection. As a foundation of vein recognition systems, vein detection faces challenges, including low feature extraction efficiency, limited robustness, and a heavy reliance on real-world data. Additionally, environmental variability and advancements in spoofing technologies further exacerbate data privacy and security concerns. To address these challenges, this paper proposes MixCFormer, a hybrid CNN–transformer architecture that incorporates Mixup data augmentation to improve the accuracy of finger vein liveness detection and reduce dependency on large-scale real datasets. First, the MixCFormer model applies baseline drift elimination, morphological filtering, and Butterworth filtering techniques to minimize the impact of background noise and illumination variations, thereby enhancing the clarity and recognizability of vein features. Next, finger vein video data are transformed into feature sequences, optimizing feature extraction and matching efficiency, effectively capturing dynamic time-series information and improving discrimination between live and forged samples. Furthermore, Mixup data augmentation is used to expand sample diversity and decrease dependency on extensive real datasets, thereby enhancing the model’s ability to recognize forged samples across diverse attack scenarios. Finally, the CNN and transformer architecture leverages both local and global feature extraction capabilities to capture vein feature correlations and dependencies. Residual connections improve feature propagation, enhancing the stability of feature representations in liveness detection. Rigorous experimental evaluations demonstrate that MixCFormer achieves a detection accuracy of 99.51% on finger vein datasets, significantly outperforming existing methods. Full article
Show Figures

Figure 1

13 pages, 2244 KiB  
Article
Dual-Stream Enhanced Deep Network for Transmission Near-Infrared Dorsal Hand Vein Age Estimation with Attention Mechanisms
by Zhenghua Shu, Zhihua Xie and Xiaowei Zou
Photonics 2024, 11(12), 1113; https://doi.org/10.3390/photonics11121113 - 25 Nov 2024
Viewed by 943
Abstract
Dorsal hand vein recognition, with unique stable and reliable advantages, has attracted considerable attention from numerous researchers. In this case, the dorsal hand vein images captured by the means of transmission infrared imaging are clearer than those collected by other infrared methods, enabling [...] Read more.
Dorsal hand vein recognition, with unique stable and reliable advantages, has attracted considerable attention from numerous researchers. In this case, the dorsal hand vein images captured by the means of transmission infrared imaging are clearer than those collected by other infrared methods, enabling it to be more suitable for the biometric applications. However, less attention is paid to individual age estimation based on dorsal hand veins. To this end, this paper proposes an efficient dorsal hand vein age estimation model using a deep neural network with attention mechanisms. Specifically, a convolutional neural network (CNN) is developed to extract the expressive features for age estimation. Simultaneously, another deep residual network is leveraged to strengthen the representation ability on subtle dorsal vein textures. Moreover, variable activation functions and pooling layers are integrated into the respective streams to enhance the nonlinearity modeling of the dual-stream model. Finally, a dynamic attention mechanism module is embedded into the dual-stream network to achieve multi-modal collaborative enhancement, guiding the model to concentrate on salient age-specific features. To evaluate the performance of dorsal hand vein age estimation, this work collects dorsal hand vein images using the transmission near-infrared spectrum from 300 individuals across various age groups. The experimental results show that the dual-stream enhanced network with the attention mechanism significantly improves the accuracy of dorsal hand vein age estimation in comparison with other deep learning approaches, indicating the potential of using near-infrared dorsal hand vein imaging and deep learning technology for efficient human age estimation. Full article
(This article belongs to the Special Issue Optical Sensing Technologies, Devices and Their Data Applications)
Show Figures

Figure 1

18 pages, 7349 KiB  
Article
YOLOv8-GABNet: An Enhanced Lightweight Network for the High-Precision Recognition of Citrus Diseases and Nutrient Deficiencies
by Qiufang Dai, Yungao Xiao, Shilei Lv, Shuran Song, Xiuyun Xue, Shiyao Liang, Ying Huang and Zhen Li
Agriculture 2024, 14(11), 1964; https://doi.org/10.3390/agriculture14111964 - 1 Nov 2024
Cited by 6 | Viewed by 1459
Abstract
Existing deep learning models for detecting citrus diseases and nutritional deficiencies grapple with issues related to recognition accuracy, complex backgrounds, occlusions, and the need for lightweight architecture. In response, we developed an improved YOLOv8-GABNet model designed specifically for citrus disease and nutritional deficiency [...] Read more.
Existing deep learning models for detecting citrus diseases and nutritional deficiencies grapple with issues related to recognition accuracy, complex backgrounds, occlusions, and the need for lightweight architecture. In response, we developed an improved YOLOv8-GABNet model designed specifically for citrus disease and nutritional deficiency detection, which effectively addresses these challenges. This model incorporates several key enhancements: A lightweight ADown subsampled convolutional block is utilized to reduce both the model’s parameter count and its computational demands, replacing the traditional convolutional module. Additionally, a weighted Bidirectional Feature Pyramid Network (BiFPN) supersedes the original feature fusion network, enhancing the model’s ability to manage complex backgrounds and achieve multiscale feature extraction and integration. Furthermore, we introduced important features through the Global to Local Spatial Aggregation module (GLSA), focusing on crucial image details to enhance both the accuracy and robustness of the model. This study processed the collected images, resulting in a dataset of 1102 images. Using LabelImg, bounding boxes were applied to annotate leaves affected by diseases. The dataset was constructed to include three types of citrus diseases—anthracnose, canker, and yellow vein disease—as well as two types of nutritional deficiencies, namely magnesium deficiency and manganese deficiency. This dataset was expanded to 9918 images through data augmentation and was used for experimental validation. The results show that, compared to the original YOLOv8, our YOLOv8-GABNet model reduces the parameter count by 43.6% and increases the mean Average Precision (mAP50) by 4.3%. Moreover, the model size was reduced from 50.1 MB to 30.2 MB, facilitating deployment on mobile devices. When compared with mainstream models like YOLOv5s, Faster R-CNN, SSD, YOLOv9t, and YOLOv10n, the YOLOv8-GABNet model demonstrates superior performance in terms of size and accuracy, offering an optimal balance between performance, size, and speed. This study confirms that the model effectively identifies the common diseases and nutritional deficiencies of citrus from Conghua’s “Citrus Planet”. Future deployment to mobile devices will provide farmers with instant and precise support. Full article
(This article belongs to the Special Issue Machine Vision Solutions and AI-Driven Systems in Agriculture)
Show Figures

Figure 1

22 pages, 3918 KiB  
Article
A Prior-Guided Dual Branch Multi-Feature Fusion Network for Building Segmentation in Remote Sensing Images
by Yingbin Wu, Peng Zhao, Fubo Wang, Mingquan Zhou, Shengling Geng and Dan Zhang
Buildings 2024, 14(7), 2006; https://doi.org/10.3390/buildings14072006 - 2 Jul 2024
Cited by 1 | Viewed by 1432
Abstract
The domain of remote sensing image processing has witnessed remarkable advancements in recent years, with deep convolutional neural networks (CNNs) establishing themselves as a prominent approach for building segmentation. Despite the progress, traditional CNNs, which rely on convolution and pooling for feature extraction [...] Read more.
The domain of remote sensing image processing has witnessed remarkable advancements in recent years, with deep convolutional neural networks (CNNs) establishing themselves as a prominent approach for building segmentation. Despite the progress, traditional CNNs, which rely on convolution and pooling for feature extraction during the encoding phase, often fail to precisely delineate global pixel interactions, potentially leading to the loss of vital semantic details. Moreover, conventional CNN-based segmentation models frequently neglect the nuanced semantic differences between shallow and deep features during the decoding phase, which can result in subpar feature integration through rudimentary addition or concatenation techniques. Additionally, the unique boundary characteristics of buildings in remote sensing images, which offer a rich vein of prior information, have not been fully harnessed by traditional CNNs. This paper introduces an innovative approach to building segmentation in remote sensing images through a prior-guided dual branch multi-feature fusion network (PDBMFN). The network is composed of a prior-guided branch network (PBN) in the encoding process, a parallel dilated convolution module (PDCM) designed to incorporate prior information, and a multi-feature aggregation module (MAM) in the decoding process. The PBN leverages prior region and edge information derived from superpixels and edge maps to enhance edge detection accuracy during the encoding phase. The PDCM integrates features from both branches and applies dilated convolution across various scales to expand the receptive field and capture a more comprehensive semantic context. During the decoding phase, the MAM utilizes deep semantic information to direct the fusion of features, thereby optimizing segmentation efficacy. Through a sequence of aggregations, the MAM gradually merges deep and shallow semantic information, culminating in a more enriched and holistic feature representation. Extensive experiments are conducted across diverse datasets, such as WHU, Inria Aerial, and Massachusetts, revealing that PDBMFN outperforms other sophisticated methods in terms of segmentation accuracy. In the key segmentation metrics, including mIoU, precision, recall, and F1 score, PDBMFN shows a marked superiority over contemporary techniques. The ablation studies further substantiate the performance improvements conferred by the PBN’s prior information guidance and the efficacy of the PDCM and MAM modules. Full article
(This article belongs to the Section Construction Management, and Computers & Digitization)
Show Figures

Figure 1

14 pages, 2920 KiB  
Article
Zero-FVeinNet: Optimizing Finger Vein Recognition with Shallow CNNs and Zero-Shuffle Attention for Low-Computational Devices
by Nghi C. Tran, Bach-Tung Pham, Vivian Ching-Mei Chu, Kuo-Chen Li, Phuong Thi Le, Shih-Lun Chen, Aufaclav Zatu Kusuma Frisky, Yung-Hui Li and Jia-Ching Wang
Electronics 2024, 13(9), 1751; https://doi.org/10.3390/electronics13091751 - 1 May 2024
Viewed by 1973
Abstract
In the context of increasing reliance on mobile devices, robust personal security solutions are critical. This paper presents Zero-FVeinNet, an innovative, lightweight convolutional neural network (CNN) tailored for finger vein recognition on mobile and embedded devices, which are typically resource-constrained. The model integrates [...] Read more.
In the context of increasing reliance on mobile devices, robust personal security solutions are critical. This paper presents Zero-FVeinNet, an innovative, lightweight convolutional neural network (CNN) tailored for finger vein recognition on mobile and embedded devices, which are typically resource-constrained. The model integrates cutting-edge features such as Zero-Shuffle Coordinate Attention and a blur pool layer, enhancing architectural efficiency and recognition accuracy under various imaging conditions. A notable reduction in computational demands is achieved through an optimized design involving only 0.3 M parameters, thereby enabling faster processing and reduced energy consumption, which is essential for mobile applications. An empirical evaluation on several leading public finger vein datasets demonstrates that Zero-FVeinNet not only outperforms traditional biometric systems in speed and efficiency but also establishes new standards in biometric identity verification. The Zero-FVeinNet achieves a Correct Identification Rate (CIR) of 99.9% on the FV-USM dataset, with a similarly high accuracy on other datasets. This paper underscores the potential of Zero-FVeinNet to significantly enhance security features on mobile devices by merging high accuracy with operational efficiency, paving the way for advanced biometric verification technologies. Full article
(This article belongs to the Special Issue Emerging Artificial Intelligence Technologies and Applications)
Show Figures

Figure 1

18 pages, 3683 KiB  
Article
Finger Vein Identification Based on Large Kernel Convolution and Attention Mechanism
by Meihui Li, Yufei Gong and Zhaohui Zheng
Sensors 2024, 24(4), 1132; https://doi.org/10.3390/s24041132 - 9 Feb 2024
Cited by 7 | Viewed by 1710
Abstract
FV (finger vein) identification is a biometric identification technology that extracts the features of FV images for identity authentication. To address the limitations of CNN-based FV identification, particularly the challenge of small receptive fields and difficulty in capturing long-range dependencies, an FV identification [...] Read more.
FV (finger vein) identification is a biometric identification technology that extracts the features of FV images for identity authentication. To address the limitations of CNN-based FV identification, particularly the challenge of small receptive fields and difficulty in capturing long-range dependencies, an FV identification method named Let-Net (large kernel and attention mechanism network) was introduced, which combines local and global information. Firstly, Let-Net employs large kernels to capture a broader spectrum of spatial contextual information, utilizing deep convolution in conjunction with residual connections to curtail the volume of model parameters. Subsequently, an integrated attention mechanism is applied to augment information flow within the channel and spatial dimensions, effectively modeling global information for the extraction of crucial FV features. The experimental results on nine public datasets show that Let-Net has excellent identification performance, and the EER and accuracy rate on the FV_USM dataset can reach 0.04% and 99.77%. The parameter number and FLOPs of Let-Net are only 0.89M and 0.25G, which means that the time cost of training and reasoning of the model is low, and it is easier to deploy and integrate into various applications. Full article
(This article belongs to the Section Biomedical Sensors)
Show Figures

Figure 1

18 pages, 5012 KiB  
Article
Hybrid Feature Extractor Using Discrete Wavelet Transform and Histogram of Oriented Gradient on Convolutional-Neural-Network-Based Palm Vein Recognition
by Meirista Wulandari, Rifai Chai, Basari Basari and Dadang Gunawan
Sensors 2024, 24(2), 341; https://doi.org/10.3390/s24020341 - 6 Jan 2024
Cited by 4 | Viewed by 2507
Abstract
Biometric recognition techniques have become more developed recently, especially in security and attendance systems. Biometrics are features attached to the human body that are considered safer and more reliable since they are difficult to imitate or lose. One of the popular biometrics considered [...] Read more.
Biometric recognition techniques have become more developed recently, especially in security and attendance systems. Biometrics are features attached to the human body that are considered safer and more reliable since they are difficult to imitate or lose. One of the popular biometrics considered in research is palm veins. They are an intrinsic biometric located under the human skin, so they have several advantages when developing verification systems. However, palm vein images obtained based on infrared spectra have several disadvantages, such as nonuniform illumination and low contrast. This study, based on a convolutional neural network (CNN), was conducted on five public datasets from CASIA, Vera, Tongji, PolyU, and PUT, with three parameters: accuracy, AUC, and EER. Our proposed VeinCNN recognition method, called verification scheme with VeinCNN, uses hybrid feature extraction from a discrete wavelet transform (DWT) and histogram of oriented gradient (HOG). It shows promising results in terms of accuracy, AUC, and EER values, especially in the total parameter values. The best result was obtained for the CASIA dataset with 99.85% accuracy, 99.80% AUC, and 0.0083 EER. Full article
(This article belongs to the Special Issue Computational Intelligence Based-Brain-Body Machine Interface)
Show Figures

Figure 1

17 pages, 5836 KiB  
Article
Interpretable Detection of Diabetic Retinopathy, Retinal Vein Occlusion, Age-Related Macular Degeneration, and Other Fundus Conditions
by Wenlong Li, Linbo Bian, Baikai Ma, Tong Sun, Yiyun Liu, Zhengze Sun, Lin Zhao, Kang Feng, Fan Yang, Xiaona Wang, Szyyann Chan, Hongliang Dou and Hong Qi
Diagnostics 2024, 14(2), 121; https://doi.org/10.3390/diagnostics14020121 - 5 Jan 2024
Cited by 6 | Viewed by 3142
Abstract
Diabetic retinopathy (DR), retinal vein occlusion (RVO), and age-related macular degeneration (AMD) pose significant global health challenges, often resulting in vision impairment and blindness. Automatic detection of these conditions is crucial, particularly in underserved rural areas with limited access to ophthalmic services. Despite [...] Read more.
Diabetic retinopathy (DR), retinal vein occlusion (RVO), and age-related macular degeneration (AMD) pose significant global health challenges, often resulting in vision impairment and blindness. Automatic detection of these conditions is crucial, particularly in underserved rural areas with limited access to ophthalmic services. Despite remarkable advancements in artificial intelligence, especially convolutional neural networks (CNNs), their complexity can make interpretation difficult. In this study, we curated a dataset consisting of 15,089 color fundus photographs (CFPs) obtained from 8110 patients who underwent fundus fluorescein angiography (FFA) examination. The primary objective was to construct integrated models that merge CNNs with an attention mechanism. These models were designed for a hierarchical multilabel classification task, focusing on the detection of DR, RVO, AMD, and other fundus conditions. Furthermore, our approach extended to the detailed classification of DR, RVO, and AMD according to their respective subclasses. We employed a methodology that entails the translation of diagnostic information obtained from FFA results into CFPs. Our investigation focused on evaluating the models’ ability to achieve precise diagnoses solely based on CFPs. Remarkably, our models showcased improvements across diverse fundus conditions, with the ConvNeXt-base + attention model standing out for its exceptional performance. The ConvNeXt-base + attention model achieved remarkable metrics, including an area under the receiver operating characteristic curve (AUC) of 0.943, a referable F1 score of 0.870, and a Cohen’s kappa of 0.778 for DR detection. For RVO, it attained an AUC of 0.960, a referable F1 score of 0.854, and a Cohen’s kappa of 0.819. Furthermore, in AMD detection, the model achieved an AUC of 0.959, an F1 score of 0.727, and a Cohen’s kappa of 0.686. Impressively, the model demonstrated proficiency in subclassifying RVO and AMD, showcasing commendable sensitivity and specificity. Moreover, our models enhanced interpretability by visualizing attention weights on fundus images, aiding in the identification of disease findings. These outcomes underscore the substantial impact of our models in advancing the detection of DR, RVO, and AMD, offering the potential for improved patient outcomes and positively influencing the healthcare landscape. Full article
(This article belongs to the Special Issue Artificial Intelligence in Ophthalmology)
Show Figures

Figure 1

25 pages, 5346 KiB  
Article
An Improved Multimodal Biometric Identification System Employing Score-Level Fuzzification of Finger Texture and Finger Vein Biometrics
by Syed Aqeel Haider, Shahzad Ashraf, Raja Masood Larik, Nusrat Husain, Hafiz Abdul Muqeet, Usman Humayun, Ashraf Yahya, Zeeshan Ahmad Arfeen and Muhammad Farhan Khan
Sensors 2023, 23(24), 9706; https://doi.org/10.3390/s23249706 - 8 Dec 2023
Cited by 8 | Viewed by 6617
Abstract
This research work focuses on a Near-Infra-Red (NIR) finger-images-based multimodal biometric system based on Finger Texture and Finger Vein biometrics. The individual results of the biometric characteristics are fused using a fuzzy system, and the final identification result is achieved. Experiments are performed [...] Read more.
This research work focuses on a Near-Infra-Red (NIR) finger-images-based multimodal biometric system based on Finger Texture and Finger Vein biometrics. The individual results of the biometric characteristics are fused using a fuzzy system, and the final identification result is achieved. Experiments are performed for three different databases, i.e., the Near-Infra-Red Hand Images (NIRHI), Hong Kong Polytechnic University (HKPU) and University of Twente Finger Vein Pattern (UTFVP) databases. First, the Finger Texture biometric employs an efficient texture feature extracting algorithm, i.e., Linear Binary Pattern. Then, the classification is performed using Support Vector Machine, a proven machine learning classification algorithm. Second, the transfer learning of pre-trained convolutional neural networks (CNNs) is performed for the Finger Vein biometric, employing two approaches. The three selected CNNs are AlexNet, VGG16 and VGG19. In Approach 1, before feeding the images for the training of the CNN, the necessary preprocessing of NIR images is performed. In Approach 2, before the pre-processing step, image intensity optimization is also employed to regularize the image intensity. NIRHI outperforms HKPU and UTFVP for both of the modalities of focus, in a unimodal setup as well as in a multimodal one. The proposed multimodal biometric system demonstrates a better overall identification accuracy of 99.62% in comparison with 99.51% and 99.50% reported in the recent state-of-the-art systems. Full article
(This article belongs to the Section Biosensors)
Show Figures

Figure 1

7 pages, 5745 KiB  
Proceeding Paper
Facial Beauty Prediction Using an Ensemble of Deep Convolutional Neural Networks
by Djamel Eddine Boukhari, Ali Chemsa, Abdelmalik Taleb-Ahmed, Riadh Ajgou and Mohamed taher Bouzaher
Eng. Proc. 2023, 56(1), 125; https://doi.org/10.3390/ASEC2023-15400 - 27 Oct 2023
Cited by 3 | Viewed by 2199
Abstract
The topic of facial beauty analysis has emerged as a crucial and fascinating subject of human culture. With various applications and significant attention from researchers, recent studies have investigated the relationship between facial features and age, emotions, and other factors using multidisciplinary approaches. [...] Read more.
The topic of facial beauty analysis has emerged as a crucial and fascinating subject of human culture. With various applications and significant attention from researchers, recent studies have investigated the relationship between facial features and age, emotions, and other factors using multidisciplinary approaches. Facial beauty prediction is a significant visual recognition problem in the assessment of facial attractiveness, which is consistent with human perception. Overcoming the challenges associated with facial beauty prediction requires considerable effort due to the field’s novelty and lack of resources. In this vein, a deep learning method has recently demonstrated remarkable abilities in feature representation and analysis. Accordingly, this paper proposes an ensemble based on pre-trained convolutional neural network models to identify scores for facial beauty prediction. These ensembles are three separate deep convolutional neural networks, each with a unique structural representation built by previously trained models from Inceptionv3, Mobilenetv2, and a new simple network based on Convolutional Neural Networks (CNNs) for facial beauty prediction problems. According to the SCUT-FBP5500 benchmark dataset, the obtained 0.9350 Pearson coefficient experimental result demonstrated that using this ensemble of deep networks leads to a better prediction of facial beauty closer to human evaluation than conventional technology that spreads facial beauty. Finally, potential research directions are suggested for future research on facial beauty prediction. Full article
(This article belongs to the Proceedings of The 4th International Electronic Conference on Applied Sciences)
Show Figures

Figure 1

15 pages, 6489 KiB  
Article
TTH-Net: Two-Stage Transformer–CNN Hybrid Network for Leaf Vein Segmentation
by Peng Song, Yonghong Yu and Yang Zhang
Appl. Sci. 2023, 13(19), 11019; https://doi.org/10.3390/app131911019 - 6 Oct 2023
Cited by 2 | Viewed by 1745
Abstract
Leaf vein segmentation is crucial in species classification and smart agriculture. The existing methods combine manual features and machine learning techniques to segment coarse leaf veins. However, the extraction of the intricate patterns is time consuming. To address the issues, we propose a [...] Read more.
Leaf vein segmentation is crucial in species classification and smart agriculture. The existing methods combine manual features and machine learning techniques to segment coarse leaf veins. However, the extraction of the intricate patterns is time consuming. To address the issues, we propose a coarse-to-fine two-stage hybrid network termed TTH-Net, which combines a transformer and CNN to accurately extract veins. Specifically, the proposed TTH-Net consists of two stages and a cross-stage semantic enhancement module. The first stage utilizes the Vision Transformer (base version) to extract globally high-level feature representations. Based on these features, the second stage identifies fine-grained vein features via CNN. To enhance the interaction between the two stages, a cross-stage semantic enhancement module is designed to integrate the strengths of the transformer and CNN, which also improves the segmentation accuracy of the decoder. Extensive experiments on the public dataset LVN are conducted, and the results prove that TTH-Net has significant advantages over other methods in leaf vein segmentation. Full article
Show Figures

Figure 1

23 pages, 5380 KiB  
Article
SoftVein-WELM: A Weighted Extreme Learning Machine Model for Soft Biometrics on Palm Vein Images
by David Zabala-Blanco, Ruber Hernández-García and Ricardo J. Barrientos
Electronics 2023, 12(17), 3608; https://doi.org/10.3390/electronics12173608 - 26 Aug 2023
Cited by 7 | Viewed by 1824
Abstract
Contactless biometric technologies such as palm vein recognition have gained more relevance in the present and immediate future due to the COVID-19 pandemic. Since certain soft biometrics like gender and age can generate variations in the visualization of palm vein patterns, these soft [...] Read more.
Contactless biometric technologies such as palm vein recognition have gained more relevance in the present and immediate future due to the COVID-19 pandemic. Since certain soft biometrics like gender and age can generate variations in the visualization of palm vein patterns, these soft traits can reduce the penetration rate on large-scale databases for mass individual recognition. Due to the limited availability of public databases, few works report on the existing approaches to gender and age classification through vein pattern images. Moreover, soft biometric classification commonly faces the problem of imbalanced data class distributions, representing a limitation of the reported approaches. This paper introduces weighted extreme learning machine (W-ELM) models for gender and age classification based on palm vein images to address imbalanced data problems, improving the classification performance. The highlights of our proposal are that it avoids using a feature extraction process and can incorporate a weight matrix in optimizing the ELM model by exploiting the imbalanced nature of the data, which guarantees its application in realistic scenarios. In addition, we evaluate a new class distribution for soft biometrics on the VERA dataset and a new multi-label scheme identifying gender and age simultaneously. The experimental results demonstrate that both evaluated W-ELM models outperform previous existing approaches and a novel CNN-based method in terms of the accuracy and G-mean metrics, achieving accuracies of 98.91% and 99.53% for gender classification on VERA and PolyU, respectively. In more challenging scenarios for age and gender–age classifications on the VERA dataset, the proposed method reaches accuracies of 97.05% and 96.91%, respectively. The multi-label classification results suggest that further studies can be conducted on multi-task ELM for palm vein recognition. Full article
Show Figures

Figure 1

Back to TopTop