Sign in to use this feature.

Years

Between: -

Subjects

remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline

Journals

Article Types

Countries / Regions

Search Results (16)

Search Parameters:
Keywords = blind modulation classification

Order results
Result details
Results per page
Select all
Export citation of selected articles as:
24 pages, 7886 KiB  
Article
AdaptiveSwin-CNN: Adaptive Swin-CNN Framework with Self-Attention Fusion for Robust Multi-Class Retinal Disease Diagnosis
by Imran Qureshi
AI 2025, 6(2), 28; https://doi.org/10.3390/ai6020028 - 6 Feb 2025
Cited by 1 | Viewed by 1593
Abstract
Retinal diseases account for a large fraction of global blinding disorders, requiring sophisticated diagnostic tools for early management. In this study, the author proposes a hybrid deep learning framework in the form of AdaptiveSwin-CNN that combines Swin Transformers and Convolutional Neural Networks (CNNs) [...] Read more.
Retinal diseases account for a large fraction of global blinding disorders, requiring sophisticated diagnostic tools for early management. In this study, the author proposes a hybrid deep learning framework in the form of AdaptiveSwin-CNN that combines Swin Transformers and Convolutional Neural Networks (CNNs) for the classification of multi-class retinal diseases. In contrast to traditional architectures, AdaptiveSwin-CNN utilizes a brand-new Self-Attention Fusion Module (SAFM) to effectively combine multi-scale spatial and contextual options to alleviate class imbalance and give attention to refined retina lesions. Utilizing the adaptive baseline augmentation and dataset-driven preprocessing of input images, the AdaptiveSwin-CNN model resolves the problem of the variability of fundus images in the dataset. AdaptiveSwin-CNN achieved a mean accuracy of 98.89%, sensitivity of 95.2%, specificity of 96.7%, and F1-score of 97.2% on RFMiD and ODIR benchmarks, outperforming other solutions. An additional lightweight ensemble XGBoost classifier to reduce overfitting and increase interpretability also increased diagnostic accuracy. The results highlight AdaptiveSwin-CNN as a robust and computationally efficient decision-support system. Full article
(This article belongs to the Special Issue Multimodal Artificial Intelligence in Healthcare)
Show Figures

Figure 1

17 pages, 1237 KiB  
Article
Hybrid Deep Learning Model for Cataract Diagnosis Assistance
by Zonghong Feng, Kai Xu, Liangchang Li and Yong Wang
Appl. Sci. 2024, 14(23), 11314; https://doi.org/10.3390/app142311314 - 4 Dec 2024
Viewed by 1506
Abstract
With the population aging globally, cataracts have become one of the main causes of vision impairment. Early diagnosis and treatment of cataracts are crucial for preventing blindness. However, the use of deep learning models for assisting in the diagnosis of cataracts is limited [...] Read more.
With the population aging globally, cataracts have become one of the main causes of vision impairment. Early diagnosis and treatment of cataracts are crucial for preventing blindness. However, the use of deep learning models for assisting in the diagnosis of cataracts is limited due to reasons such as scarce data labeling, small sample size, uneven distribution, and poor generalization ability in the field. Therefore, this paper proposes a hybrid deep learning network for assisting in the diagnosis of cataract fundus images, attempting to solve the above problems and limitations. The network is based on the idea of transfer learning for feature extraction of fundus images, and introduces the Squeeze-and-Excitation (SE) module and prototype network for feature enhancement and classification, improving the model’s generalization ability for new disease samples. Finally, this paper verifies the role of each part of the model through ablation experiments, especially the significant contribution of the SE_block module and the prototype network classifier in enhancing the model’s performance. The experimental results show that the proposed model achieves excellent performance in the task of cataract fundus image recognition, with an accuracy of 0.9875, AUC value of 0.9984, and F1 score of 0.9855. The establishment of this hybrid model not only provides an effective tool for the auxiliary diagnosis of cataracts but also provides a new perspective and method for the application of deep learning in the field of ophthalmic disease recognition. Full article
Show Figures

Figure 1

11 pages, 2245 KiB  
Article
Metasurface-Based Image Classification Using Diffractive Deep Neural Network
by Kaiyang Cheng, Cong Deng, Fengyu Ye, Hongqiang Li, Fei Shen, Yuancheng Fan and Yubin Gong
Nanomaterials 2024, 14(22), 1812; https://doi.org/10.3390/nano14221812 - 12 Nov 2024
Viewed by 1919
Abstract
The computer-assisted inverse design of photonic computing, especially by leveraging artificial intelligence algorithms, offers great convenience to accelerate the speed of development and improve calculation accuracy. However, traditional thickness-based modulation methods are hindered by large volume and difficult fabrication process, making it hard [...] Read more.
The computer-assisted inverse design of photonic computing, especially by leveraging artificial intelligence algorithms, offers great convenience to accelerate the speed of development and improve calculation accuracy. However, traditional thickness-based modulation methods are hindered by large volume and difficult fabrication process, making it hard to meet the data-driven requirements of flexible light modulation. Here, we propose a diffractive deep neural network (D2NN) framework based on a three-layer all-dielectric phased transmitarray as hidden layers, which can perform the classification of handwritten digits. By tailoring the radius of a silicon nanodisk of a meta-atom, the metasurface can realize the phase profile calculated by D2NN and maintain a relative high transmittance of 0.9 at a wavelength of 600 nm. The designed image classifier consists of three layers of phase-only metasurfaces, each of which contains 1024 units, mimicking a fully connected neural network through the diffraction of light fields. The classification task of handwriting digits from the ‘0’ to ‘5’ dataset is verified, with an accuracy of over 90% on the blind test dataset, as well as demonstrated by the full-wave simulation. Furthermore, the performance of the more complex animal image classification task is also validated by increasing the number of neurons to enhance the connectivity of the neural network. This study may provide a possible solution for practical applications such as biomedical detection, image processing, and machine vision based on all-optical computing. Full article
(This article belongs to the Special Issue Linear and Nonlinear Optical Properties of Nanomaterials)
Show Figures

Figure 1

16 pages, 4854 KiB  
Article
Point-Sim: A Lightweight Network for 3D Point Cloud Classification
by Jiachen Guo and Wenjie Luo
Algorithms 2024, 17(4), 158; https://doi.org/10.3390/a17040158 - 15 Apr 2024
Viewed by 2326
Abstract
Analyzing point clouds with neural networks is a current research hotspot. In order to analyze the 3D geometric features of point clouds, most neural networks improve the network performance by adding local geometric operators and trainable parameters. However, deep learning usually requires a [...] Read more.
Analyzing point clouds with neural networks is a current research hotspot. In order to analyze the 3D geometric features of point clouds, most neural networks improve the network performance by adding local geometric operators and trainable parameters. However, deep learning usually requires a large amount of computational resources for training and inference, which poses challenges to hardware devices and energy consumption. Therefore, some researches have started to try to use a nonparametric approach to extract features. Point-NN combines nonparametric modules to build a nonparametric network for 3D point cloud analysis, and the nonparametric components include operations such as trigonometric embedding, farthest point sampling (FPS), k-nearest neighbor (k-NN), and pooling. However, Point-NN has some blindness in feature embedding using the trigonometric function during feature extraction. To eliminate this blindness as much as possible, we utilize a nonparametric energy function-based attention mechanism (ResSimAM). The embedded features are enhanced by calculating the energy of the features by the energy function, and then the ResSimAM is used to enhance the weights of the embedded features by the energy to enhance the features without adding any parameters to the original network; Point-NN needs to compute the similarity between each feature at the naive feature similarity matching stage; however, the magnitude difference of the features in vector space during the feature extraction stage may affect the final matching result. We use the Squash operation to squeeze the features. This nonlinear operation can make the features squeeze to a certain range without changing the original direction in the vector space, thus eliminating the effect of feature magnitude, and we can ultimately better complete the naive feature matching in the vector space. We inserted these modules into the network and build a nonparametric network, Point-Sim, which performs well in 3D classification tasks. Based on this, we extend the lightweight neural network Point-SimP by adding some trainable parameters for the point cloud classification task, which requires only 0.8 M parameters for high performance analysis. Experimental results demonstrate the effectiveness of our proposed algorithm in the point cloud shape classification task. The corresponding results on ModelNet40 and ScanObjectNN are 83.9% and 66.3% for 0 M parameters—without any training—and 93.3% and 86.6% for 0.8 M parameters. The Point-SimP reaches a test speed of 962 samples per second on the ModelNet40 dataset. The experimental results show that our proposed method effectively improves the performance on point cloud classification networks. Full article
(This article belongs to the Special Issue Machine Learning for Pattern Recognition)
Show Figures

Figure 1

2 pages, 139 KiB  
Abstract
Comparative Evaluation of a Dietary Fiber Mixture in an Intestinal Screening Platform and a Crossover Intervention Study
by Femke P. M. Hoevenaars, Tim J. van den Broek, Boukje Eveleens Maarse, Matthijs Moerland, Ines Warnke, Hannah Eggink and Frank H. J. Schuren
Proceedings 2023, 91(1), 418; https://doi.org/10.3390/proceedings2023091418 - 27 Mar 2024
Viewed by 1011
Abstract
In personalized nutrition, specific recommendations are often based on extensive phenotyping. In the world of microbiome research, classification is often based on the bacteriological composition of gut microbiota and enterotypes. We investigated if there is a possibility of translating outcomes from an intestinal [...] Read more.
In personalized nutrition, specific recommendations are often based on extensive phenotyping. In the world of microbiome research, classification is often based on the bacteriological composition of gut microbiota and enterotypes. We investigated if there is a possibility of translating outcomes from an intestinal screening platform to an intervention study that makes use of phenotyping. A 12-week double-blind, randomized, placebo-controlled, crossover intervention study (8-week wash-out period) with a dietary fiber mixture of acacia gum and carrot powder (ratio 3.33:1) was performed in healthy volunteers (N = 54, 45–70 years, BMI 27.3 ± 1.4) to modulate their microbiome. Fecal samples were collected every 4 weeks during the 32-week study period. Before and after the intervention a standardized mixed meal challenge was performed and plasma samples were taken (0, 30, 60, 120, and 240 min). Postprandial responses were used for sub-group cluster analysis to identify the metabolic phenotype. The individual participants’ samples were cultured anaerobically for 24 h with the mixture and the individual fibers. Compositional 16s rRNA data of exposed in vitro (24 h) and in vivo samples (4, 8, and 12 weeks) was compared and linked to the metabolic cluster analysis. The comparison between the clinical intervention’s effect on microbiota composition after 12 weeks and a single 24 h exposure in vitro showed a statistically significant association in microbiome effects between in vivo and in vitro exposures (p < 0.05) for the fiber intervention. Analysis of the metabolic postprandial responses revealed a division between improvement and deterioration in response to the fiber intervention indicating two distinct clusters (metabolic phenotypes). Cluster 1 contained the lowest triglycerides-, total cholesterol-, and non-esterified fatty acids responses, while cluster 2 contained the highest triglycerides- and total cholesterol responses. Interestingly, the beta diversity of the microbiota was linked to these two clusters, resembling two different responses to the fiber intervention. Our study in healthy individuals demonstrates that a short-term in vitro exposure of individual microbiome samples to the fiber mixture is predictive of a long-term in vivo effect and relates to a distinct phenotypic cluster. This paves the way for using the in vitro platform as a pre-screen for intervention studies. Full article
(This article belongs to the Proceedings of The 14th European Nutrition Conference FENS 2023)
18 pages, 2992 KiB  
Article
Automatic Detection and Classification of Hypertensive Retinopathy with Improved Convolution Neural Network and Improved SVM
by Usharani Bhimavarapu, Nalini Chintalapudi and Gopi Battineni
Bioengineering 2024, 11(1), 56; https://doi.org/10.3390/bioengineering11010056 - 5 Jan 2024
Cited by 8 | Viewed by 3236
Abstract
Hypertensive retinopathy (HR) results from the microvascular retinal changes triggered by hypertension, which is the most common leading cause of preventable blindness worldwide. Therefore, it is necessary to develop an automated system for HR detection and evaluation using retinal images. We aimed to [...] Read more.
Hypertensive retinopathy (HR) results from the microvascular retinal changes triggered by hypertension, which is the most common leading cause of preventable blindness worldwide. Therefore, it is necessary to develop an automated system for HR detection and evaluation using retinal images. We aimed to propose an automated approach to identify and categorize the various degrees of HR severity. A new network called the spatial convolution module (SCM) combines cross-channel and spatial information, and the convolution operations extract helpful features. The present model is evaluated using publicly accessible datasets ODIR, INSPIREVR, and VICAVR. We applied the augmentation to artificially increase the dataset of 1200 fundus images. The different HR severity levels of normal, mild, moderate, severe, and malignant are finally classified with the reduced time when compared to the existing models because in the proposed model, convolutional layers run only once on the input fundus images, which leads to a speedup and reduces the processing time in detecting the abnormalities in the vascular structure. According to the findings, the improved SVM had the highest detection and classification accuracy rate in the vessel classification with an accuracy of 98.99% and completed the task in 160.4 s. The ten-fold classification achieved the highest accuracy of 98.99%, i.e., 0.27 higher than the five-fold classification accuracy and the improved KNN classifier achieved an accuracy of 98.72%. When computation efficiency is a priority, the proposed model’s ability to quickly recognize different HR severity levels is significant. Full article
Show Figures

Figure 1

17 pages, 4564 KiB  
Article
A Multi-Task Learning and Multi-Branch Network for DR and DME Joint Grading
by Xiaoxue Xing, Shenbo Mao, Minghan Yan, He Yu, Dongfang Yuan, Cancan Zhu, Cong Zhang, Jian Zhou and Tingfa Xu
Appl. Sci. 2024, 14(1), 138; https://doi.org/10.3390/app14010138 - 22 Dec 2023
Cited by 1 | Viewed by 1357
Abstract
Diabetic Retinopathy (DR) is one of the most common microvascular complications of diabetes. Diabetic Macular Edema (DME) is a concomitant symptom of DR. As the grade of lesion of DR and DME increase, the possibility of blindness can also increase significantly. In order [...] Read more.
Diabetic Retinopathy (DR) is one of the most common microvascular complications of diabetes. Diabetic Macular Edema (DME) is a concomitant symptom of DR. As the grade of lesion of DR and DME increase, the possibility of blindness can also increase significantly. In order to take the early interventions as soon as possible to reduce the likelihood of blindness, it is necessary to perform both DR and DME grading. We design a joint grading model based on multi-task learning and multi-branch networks (MaMNet) for DR and DME grading. The model mainly includes a multi-branch network (MbN), a feature fusion module, and a disease classification module. The MbN is formed by four branch structures, which can extract the low-level feature information of DME and DR in a targeted way; the feature fusion module is composed of a self-feature extraction module (SFEN), cross-feature extraction module (CFEN) and atrous spatial pyramid pooling module (ASPP). By combining various features collected from the aforementioned modules, the feature fusion module can provide more thorough discriminative features, which benefits the joint grading accuracy. The ISBI-2018-IDRiD challenge dataset is used to evaluate the performance of the proposed model. The experimental results show that based on the multi-task strategy the two grading tasks of DR and DME can provide each other with additional useful information. The joint accuracy of the model, the accuracy of DR and the accuracy of DME are 61.2%, 64.1% and 79.4% respectively. Full article
(This article belongs to the Special Issue Artificial Intelligence for Health and Well-Being)
Show Figures

Figure 1

24 pages, 13431 KiB  
Article
Toward Lightweight Diabetic Retinopathy Classification: A Knowledge Distillation Approach for Resource-Constrained Settings
by Niful Islam, Md. Mehedi Hasan Jony, Emam Hasan, Sunny Sutradhar, Atikur Rahman and Md. Motaharul Islam
Appl. Sci. 2023, 13(22), 12397; https://doi.org/10.3390/app132212397 - 16 Nov 2023
Cited by 6 | Viewed by 3093
Abstract
Diabetic retinopathy (DR), a consequence of diabetes, is one of the prominent contributors to blindness. Effective intervention necessitates accurate classification of DR; this is a need that computer vision-based technologies address. However, using large-scale deep learning models for DR classification presents difficulties, especially [...] Read more.
Diabetic retinopathy (DR), a consequence of diabetes, is one of the prominent contributors to blindness. Effective intervention necessitates accurate classification of DR; this is a need that computer vision-based technologies address. However, using large-scale deep learning models for DR classification presents difficulties, especially when integrating them into devices with limited resources, particularly in places with poor technological infrastructure. In order to address this, our research presents a knowledge distillation-based approach, where we train a fusion model, composed of ResNet152V2 and Swin Transformer, as the teacher model. The knowledge learned from the heavy teacher model is transferred to the lightweight student model of 102 megabytes, which consists of Xception with a customized convolutional block attention module (CBAM). The system also integrates a four-stage image enhancement technique to improve the image quality. We compared the model against eight state-of-the-art classifiers on five evaluation metrics; the experiments show superior performance of the model over other methods on two datasets (APTOS and IDRiD). The model performed exceptionally well on the APTOS dataset, achieving 100% accuracy in binary classification and 99.04% accuracy in multi-class classification. On the IDRiD dataset, the results were 98.05% for binary classification accuracy and 94.17% for multi-class accuracy. The proposed approach shows promise for practical applications, enabling accessible DR assessment even in technologically underdeveloped environments. Full article
(This article belongs to the Special Issue AI Technologies in Biomedical Image Processing and Analysis)
Show Figures

Figure 1

32 pages, 1148 KiB  
Review
A Survey of Blind Modulation Classification Techniques for OFDM Signals
by Anand Kumar, Sudhan Majhi, Guan Gui, Hsiao-Chun Wu and Chau Yuen
Sensors 2022, 22(3), 1020; https://doi.org/10.3390/s22031020 - 28 Jan 2022
Cited by 22 | Viewed by 7901
Abstract
Blind modulation classification (MC) is an integral part of designing an adaptive or intelligent transceiver for future wireless communications. Blind MC has several applications in the adaptive and automated systems of sixth generation (6G) communications to improve spectral efficiency and power efficiency, and [...] Read more.
Blind modulation classification (MC) is an integral part of designing an adaptive or intelligent transceiver for future wireless communications. Blind MC has several applications in the adaptive and automated systems of sixth generation (6G) communications to improve spectral efficiency and power efficiency, and reduce latency. It will become a integral part of intelligent software-defined radios (SDR) for future communication. In this paper, we provide various MC techniques for orthogonal frequency division multiplexing (OFDM) signals in a systematic way. We focus on the most widely used statistical and machine learning (ML) models and emphasize their advantages and limitations. The statistical-based blind MC includes likelihood-based (LB), maximum a posteriori (MAP) and feature-based methods (FB). The ML-based automated MC includes k-nearest neighbors (KNN), support vector machine (SVM), decision trees (DTs), convolutional neural networks (CNNs), recurrent neural networks (RNNs), and long short-term memory (LSTM) based MC methods. This survey will help the reader to understand the main characteristics of each technique, their advantages and disadvantages. We have also simulated some primary methods, i.e., statistical- and ML-based algorithms, under various constraints, which allows a fair comparison among different methodologies. The overall system performance in terms bit error rate (BER) in the presence of MC is also provided. We also provide a survey of some practical experiment works carried out through National Instrument hardware over an indoor propagation environment. In the end, open problems and possible directions for blind MC research are briefly discussed. Full article
(This article belongs to the Special Issue Feature Papers in Communications Section 2022)
Show Figures

Figure 1

14 pages, 4702 KiB  
Article
Improved Salp Swarm Optimization Algorithm: Application in Feature Weighting for Blind Modulation Identification
by Sarra Ben Chaabane, Akram Belazi, Sofiane Kharbech, Ammar Bouallegue and Laurent Clavier
Electronics 2021, 10(16), 2002; https://doi.org/10.3390/electronics10162002 - 19 Aug 2021
Cited by 13 | Viewed by 2387
Abstract
In modulation identification issues, like in any other classification problem, the performance of the classification task is significantly impacted by the feature characteristics. Feature weighting boosts the performance of machine learning algorithms, particularly the class of instance-based learning algorithms such as the Minimum [...] Read more.
In modulation identification issues, like in any other classification problem, the performance of the classification task is significantly impacted by the feature characteristics. Feature weighting boosts the performance of machine learning algorithms, particularly the class of instance-based learning algorithms such as the Minimum Distance (MD) classifier, in which the distance measure is highly sensitive to the magnitude of features. In this paper, we propose an improved version of the Salp Swarm optimization Algorithm (SSA), called ISSA, that will be applied to optimize feature weights for an MD classifier. The aim is to improve the performance of a blind digital modulation detection approach in the context of multiple-antenna systems. The improvements introduced to SSA mainly rely on the opposition-based learning technique. Computer simulations show that the ISSA outperforms the SSA as well as the algorithms that derive from it. The ISSA also exhibits the best performance once it is applied for feature weighting in the above context. Full article
(This article belongs to the Special Issue Cognitive Radio Applications in Wireless Communication System)
Show Figures

Figure 1

16 pages, 4868 KiB  
Article
Multi-Scale Feature Fusion with Adaptive Weighting for Diabetic Retinopathy Severity Classification
by Runze Fan, Yuhong Liu and Rongfen Zhang
Electronics 2021, 10(12), 1369; https://doi.org/10.3390/electronics10121369 - 8 Jun 2021
Cited by 35 | Viewed by 3735
Abstract
Diabetic retinopathy (DR) is the prime cause of blindness in people who suffer from diabetes. Automation of DR diagnosis could help a lot of patients avoid the risk of blindness by identifying the disease and making judgments at an early stage. The main [...] Read more.
Diabetic retinopathy (DR) is the prime cause of blindness in people who suffer from diabetes. Automation of DR diagnosis could help a lot of patients avoid the risk of blindness by identifying the disease and making judgments at an early stage. The main focus of the present work is to propose a feasible scheme of DR severity level detection under the MobileNetV3 backbone network based on a multi-scale feature of the retinal fundus image and improve the classification performance of the model. Firstly, a special residual attention module RCAM for multi-scale feature extraction from different convolution layers was designed. Then, the feature fusion by an innovative operation of adaptive weighting was carried out in each layer. The corresponding weight of the convolution block is updated in the model training automatically, with further global average pooling (GAP) and division process to avoid over-fitting of the model and removing non-critical features. In addition, Focal Loss is used as a loss function due to the data imbalance of DR images. The experimental results based on Kaggle APTOS 2019 contest dataset show that our proposed method for DR severity classification achieves an accuracy of 85.32%, a kappa statistic of 77.26%, and an AUC of 0.97. The comparison results also indicate that the model obtained is superior to the existing models and presents superior classification performance on the dataset. Full article
(This article belongs to the Special Issue Recent Advances in Multimedia Signal Processing and Communications)
Show Figures

Figure 1

32 pages, 6873 KiB  
Article
AMC2N: Automatic Modulation Classification Using Feature Clustering-Based Two-Lane Capsule Networks
by Dhamyaa H. Al-Nuaimi, Muhammad F. Akbar, Laith B. Salman, Intan S. Zainal Abidin and Nor Ashidi Mat Isa
Electronics 2021, 10(1), 76; https://doi.org/10.3390/electronics10010076 - 4 Jan 2021
Cited by 13 | Viewed by 3914
Abstract
The automatic modulation classification (AMC) of a detected signal has gained considerable prominence in recent years owing to its numerous facilities. Numerous studies have focused on feature-based AMC. However, improving accuracy under low signal-to-noise ratio (SNR) rates is a serious issue in AMC. [...] Read more.
The automatic modulation classification (AMC) of a detected signal has gained considerable prominence in recent years owing to its numerous facilities. Numerous studies have focused on feature-based AMC. However, improving accuracy under low signal-to-noise ratio (SNR) rates is a serious issue in AMC. Moreover, research on the enhancement of AMC performance under low and high SNR rates is limited. Motivated by these issues, this study proposes AMC using a feature clustering-based two-lane capsule network (AMC2N). In the AMC2N, accuracy of the MC process is improved by designing a new two-layer capsule network (TL-CapsNet), and classification time is reduced by introducing a new feature clustering approach in the TL-CapsNet. Firstly, the AMC2N executes blind equalization, sampling, and quantization in trilevel preprocessing. Blind equalization is executed using a binary constant modulus algorithm to avoid intersymbol interference. To extract features from the preprocessed signal and classify signals accurately, the AMC2N employs the TL-CapsNet, in which individual lanes are incorporated to process the real and imaginary parts of the signal. In addition, it is robust to SNR variations, that is, low and high SNR rates. The TL-CapsNet extracts features from the real and imaginary parts of the given signal, which are then clustered based on feature similarity. For feature extraction and clustering, the dynamic routing procedure of the TL-CapsNet is adopted. Finally, classification is performed in the SoftMax layer of the TL-CapsNet. This study proves that the AMC2N outperforms existing methods, particularly, convolutional neural network(CNN), Robust-CNN (R-CNN), curriculum learning(CL), and Local Binary Pattern (LBP), in terms of accuracy, precision, recall, F-score, and computation time. All metrics are validated in two scenarios, and the proposed method shows promising results in both. Full article
(This article belongs to the Section Computer Science & Engineering)
Show Figures

Figure 1

19 pages, 1102 KiB  
Article
Time–Frequency-Analysis-Based Blind Modulation Classification for Multiple-Antenna Systems
by Weiheng Jiang, Xiaogang Wu, Yimou Wang, Bolin Chen, Wenjiang Feng and Yi Jin
Sensors 2021, 21(1), 231; https://doi.org/10.3390/s21010231 - 1 Jan 2021
Cited by 17 | Viewed by 3546
Abstract
Blind modulation classification is an important step in implementing cognitive radio networks. The multiple-input multiple-output (MIMO) technique is widely used in military and civil communication systems. Due to the lack of prior information about channel parameters and the overlapping of signals in MIMO [...] Read more.
Blind modulation classification is an important step in implementing cognitive radio networks. The multiple-input multiple-output (MIMO) technique is widely used in military and civil communication systems. Due to the lack of prior information about channel parameters and the overlapping of signals in MIMO systems, the traditional likelihood-based and feature-based approaches cannot be applied in these scenarios directly. Hence, in this paper, to resolve the problem of blind modulation classification in MIMO systems, the time–frequency analysis method based on the windowed short-time Fourier transform was used to analyze the time–frequency characteristics of time-domain modulated signals. Then, the extracted time–frequency characteristics are converted into red–green–blue (RGB) spectrogram images, and the convolutional neural network based on transfer learning was applied to classify the modulation types according to the RGB spectrogram images. Finally, a decision fusion module was used to fuse the classification results of all the receiving antennas. Through simulations, we analyzed the classification performance at different signal-to-noise ratios (SNRs); the results indicate that, for the single-input single-output (SISO) network, our proposed scheme can achieve 92.37% and 99.12% average classification accuracy at SNRs of −4 and 10 dB, respectively. For the MIMO network, our scheme achieves 80.42% and 87.92% average classification accuracy at −4 and 10 dB, respectively. The proposed method greatly improves the accuracy of modulation classification in MIMO networks. Full article
(This article belongs to the Section Communications)
Show Figures

Figure 1

16 pages, 2606 KiB  
Article
A Multi-Scale Residual Attention Network for Retinal Vessel Segmentation
by Yun Jiang, Huixia Yao, Chao Wu and Wenhuan Liu
Symmetry 2021, 13(1), 24; https://doi.org/10.3390/sym13010024 - 24 Dec 2020
Cited by 24 | Viewed by 4507
Abstract
Accurate segmentation of retinal blood vessels is a key step in the diagnosis of fundus diseases, among which cataracts, glaucoma, and diabetic retinopathy (DR) are the main diseases that cause blindness. Most segmentation methods based on deep convolutional neural networks can effectively extract [...] Read more.
Accurate segmentation of retinal blood vessels is a key step in the diagnosis of fundus diseases, among which cataracts, glaucoma, and diabetic retinopathy (DR) are the main diseases that cause blindness. Most segmentation methods based on deep convolutional neural networks can effectively extract features. However, convolution and pooling operations also filter out some useful information, and the final segmented retinal vessels have problems such as low classification accuracy. In this paper, we propose a multi-scale residual attention network called MRA-UNet. Multi-scale inputs enable the network to learn features at different scales, which increases the robustness of the network. In the encoding phase, we reduce the negative influence of the background and eliminate noise by using the residual attention module. We use the bottom reconstruction module to aggregate the feature information under different receptive fields, so that the model can extract the information of different thicknesses of blood vessels. Finally, the spatial activation module is used to process the up-sampled image to further increase the difference between blood vessels and background, which promotes the recovery of small blood vessels at the edges. Our method was verified on the DRIVE, CHASE, and STARE datasets. Respectively, the segmentation accuracy rates reached 96.98%, 97.58%, and 97.63%; the specificity reached 98.28%, 98.54%, and 98.73%; and the F-measure scores reached 82.93%, 81.27%, and 84.22%. We compared the experimental results with some state-of-art methods, such as U-Net, R2U-Net, and AG-UNet in terms of accuracy, sensitivity, specificity, F-measure, and AUCROC. Particularly, MRA-UNet outperformed U-Net by 1.51%, 3.44%, and 0.49% on DRIVE, CHASE, and STARE datasets, respectively. Full article
(This article belongs to the Section Computer)
Show Figures

Figure 1

19 pages, 4171 KiB  
Article
Deep Learning for Optic Disc Segmentation and Glaucoma Diagnosis on Retinal Images
by Syna Sreng, Noppadol Maneerat, Kazuhiko Hamamoto and Khin Yadanar Win
Appl. Sci. 2020, 10(14), 4916; https://doi.org/10.3390/app10144916 - 17 Jul 2020
Cited by 149 | Viewed by 12462
Abstract
Glaucoma is a major global cause of blindness. As the symptoms of glaucoma appear, when the disease reaches an advanced stage, proper screening of glaucoma in the early stages is challenging. Therefore, regular glaucoma screening is essential and recommended. However, eye screening is [...] Read more.
Glaucoma is a major global cause of blindness. As the symptoms of glaucoma appear, when the disease reaches an advanced stage, proper screening of glaucoma in the early stages is challenging. Therefore, regular glaucoma screening is essential and recommended. However, eye screening is currently subjective, time-consuming and labor-intensive and there are insufficient eye specialists available. We present an automatic two-stage glaucoma screening system to reduce the workload of ophthalmologists. The system first segmented the optic disc region using a DeepLabv3+ architecture but substituted the encoder module with multiple deep convolutional neural networks. For the classification stage, we used pretrained deep convolutional neural networks for three proposals (1) transfer learning and (2) learning the feature descriptors using support vector machine and (3) building ensemble of methods in (1) and (2). We evaluated our methods on five available datasets containing 2787 retinal images and found that the best option for optic disc segmentation is a combination of DeepLabv3+ and MobileNet. For glaucoma classification, an ensemble of methods performed better than the conventional methods for RIM-ONE, ORIGA, DRISHTI-GS1 and ACRIMA datasets with the accuracy of 97.37%, 90.00%, 86.84% and 99.53% and Area Under Curve (AUC) of 100%, 92.06%, 91.67% and 99.98%, respectively, and performed comparably with CUHKMED, the top team in REFUGE challenge, using REFUGE dataset with an accuracy of 95.59% and AUC of 95.10%. Full article
(This article belongs to the Special Issue Advanced Intelligent Imaging Technology Ⅱ)
Show Figures

Figure 1

Back to TopTop