Sign in to use this feature.

Years

Between: -

Subjects

remove_circle_outline
remove_circle_outline

Journals

Article Types

Countries / Regions

Search Results (5)

Search Parameters:
Keywords = multi-layer attention module (MAM)

Order results
Result details
Results per page
Select all
Export citation of selected articles as:
19 pages, 3011 KiB  
Article
Hyperspectral Image Classification Based on Multi-Scale Convolutional Features and Multi-Attention Mechanisms
by Qian Sun, Guangrui Zhao, Xinyuan Xia, Yu Xie, Chenrong Fang, Le Sun, Zebin Wu and Chengsheng Pan
Remote Sens. 2024, 16(12), 2185; https://doi.org/10.3390/rs16122185 - 16 Jun 2024
Cited by 5 | Viewed by 2778
Abstract
Convolutional neural network (CNN)-based and Transformer-based methods for hyperspectral image (HSI) classification have rapidly advanced due to their unique characterization capabilities. However, the fixed kernel sizes in convolutional layers limit the comprehensive utilization of multi-scale features in HSI land cover analysis, while the [...] Read more.
Convolutional neural network (CNN)-based and Transformer-based methods for hyperspectral image (HSI) classification have rapidly advanced due to their unique characterization capabilities. However, the fixed kernel sizes in convolutional layers limit the comprehensive utilization of multi-scale features in HSI land cover analysis, while the Transformer’s multi-head self-attention (MHSA) mechanism faces challenges in effectively encoding feature information across various dimensions. To tackle this issue, this article introduces an HSI classification method, based on multi-scale convolutional features and multi-attention mechanisms (i.e., MSCF-MAM). Firstly, the model employs a multi-scale convolutional module to capture features across different scales in HSIs. Secondly, to enhance the integration of local and global channel features and establish long-range dependencies, a feature enhancement module based on pyramid squeeze attention (PSA) is employed. Lastly, the model leverages a classical Transformer Encoder (TE) and linear layers to encode and classify the transformed spatial–spectral features. The proposed method is evaluated on three publicly available datasets—Salina Valley (SV), WHU-Hi-HanChuan (HC), and WHU-Hi-HongHu (HH). Extensive experimental results have demonstrated that the MSCF-MAM method outperforms several representative methods in terms of classification performance. Full article
(This article belongs to the Special Issue Advances in Hyperspectral Remote Sensing Image Processing)
Show Figures

Figure 1

24 pages, 14625 KiB  
Article
Semantic Segmentation of Urban Airborne LiDAR Point Clouds Based on Fusion Attention Mechanism and Multi-Scale Features
by Jingxue Wang, Huan Li, Zhenghui Xu and Xiao Xie
Remote Sens. 2023, 15(21), 5248; https://doi.org/10.3390/rs15215248 - 5 Nov 2023
Cited by 5 | Viewed by 3073
Abstract
Semantic segmentation of point clouds provided by airborne LiDAR survey in urban scenes is a great challenge. This is due to the fact that point clouds at boundaries of different types of objects are easy to be mixed and have geometric spatial similarity. [...] Read more.
Semantic segmentation of point clouds provided by airborne LiDAR survey in urban scenes is a great challenge. This is due to the fact that point clouds at boundaries of different types of objects are easy to be mixed and have geometric spatial similarity. In addition, the 3D descriptions of the same type of objects have different scales. To address above problems, a fusion attention convolutional network (SMAnet) was proposed in this study. The fusion attention module includes a self-attention module (SAM) and multi-head attention module (MAM). The SAM can capture feature information according to correlation of adjacent point cloud and it can distinguish the mixed point clouds with similar geometric features effectively. The MAM strengthens connections among point clouds according to different subspace features, which is beneficial for distinguishing point clouds at different scales. In feature extraction, lightweight multi-scale feature extraction layers are used to effectively utilize local information of different neighbor fields. Additionally, in order to solve the feature externalization problem and expand the network receptive field, the SoftMax-stochastic pooling (SSP) algorithm is proposed to extract global features. The ISPRS 3D Semantic Labeling Contest dataset was chosen in this study for point cloud segmentation experimentation. Results showed that the overall accuracy and average F1-score of SMAnet reach 85.7% and 75.1%, respectively. It is therefore superior to common algorithms at present. The proposed model also achieved good results on the GML(B) dataset, which proves that the model has good generalization ability. Full article
(This article belongs to the Topic Geocomputation and Artificial Intelligence for Mapping)
Show Figures

Graphical abstract

26 pages, 14471 KiB  
Article
Gas Sensor Array Fault Diagnosis Based on Multi-Dimensional Fusion, an Attention Mechanism, and Multi-Task Learning
by Pengyu Huang, Qingfeng Wang, Haotian Chen and Geyu Lu
Sensors 2023, 23(18), 7836; https://doi.org/10.3390/s23187836 - 12 Sep 2023
Cited by 3 | Viewed by 2682
Abstract
With the development of gas sensor arrays and computational technology, machine olfactory systems have been widely used in environmental monitoring, medical diagnosis, and other fields. The reliable and stable operation of gas sensing systems depends heavily on the accuracy of the sensors outputs. [...] Read more.
With the development of gas sensor arrays and computational technology, machine olfactory systems have been widely used in environmental monitoring, medical diagnosis, and other fields. The reliable and stable operation of gas sensing systems depends heavily on the accuracy of the sensors outputs. Therefore, the realization of accurate gas sensor array fault diagnosis is essential to monitor the working status of sensor arrays and ensure the normal operation of the whole system. The existing methods extract features from a single dimension and require the separate training of models for multiple diagnosis tasks, which limits diagnostic accuracy and efficiency. To address these limitations, for this study, a novel fault diagnosis network based on multi-dimensional feature fusion, an attention mechanism, and multi-task learning, MAM-Net, was developed and applied to gas sensor arrays. First, feature fusion models were applied to extract deep and comprehensive features from the original data in multiple dimensions. A residual network equipped with convolutional block attention modules and a Bi-LSTM network were designed for two-dimensional and one-dimensional signals to capture spatial and temporal features simultaneously. Subsequently, a concatenation layer was constructed using feature stitching to integrate the fault details of different dimensions and avoid ignoring useful information. Finally, a multi-task learning module was designed for the parallel learning of the sensor fault diagnosis to effectively improve the diagnosis capability. The experimental results derived from using the proposed framework on gas sensor datasets across different amounts of data, balanced and unbalanced datasets, and different experimental settings show that the proposed framework outperforms the other available methods and demonstrates good recognition accuracy and robustness. Full article
Show Figures

Figure 1

15 pages, 3060 KiB  
Article
An Effective Multi-Layer Attention Network for SAR Ship Detection
by Zhiling Suo, Yongbo Zhao and Yili Hu
J. Mar. Sci. Eng. 2023, 11(5), 906; https://doi.org/10.3390/jmse11050906 - 23 Apr 2023
Cited by 5 | Viewed by 2072
Abstract
The use of deep learning-based techniques has improved the performance of synthetic aperture radar (SAR) image-based applications, such as ship detection. However, all existing methods have limited object detection performance under the conditions of varying ship sizes and complex background noise, to the [...] Read more.
The use of deep learning-based techniques has improved the performance of synthetic aperture radar (SAR) image-based applications, such as ship detection. However, all existing methods have limited object detection performance under the conditions of varying ship sizes and complex background noise, to the best of our knowledge. In this paper, to solve both the multi-scale problem and the noisy background issues, we propose a multi-layer attention approach based on the thorough analysis of both location and semantic information. The solution works by exploring the richness of spatial information of the low-level feature maps generated by a backbone and the richness of semantic information of the high-level feature maps created by the same method. Additionally, we integrate an attention mechanism into the network to exclusively extract useful features from the input maps. Tests involving multiple SAR datasets show that our proposed solution enables significant improvements to the accuracy of ship detection regardless of vessel size and background complexity. Particularly for the widely-adopted High-Resolution SAR Images Dataset (HRSID), the new method provides a 1.3% improvement in the average precision for detection. The proposed new method can be potentially used in other feature-extraction-based classification, detection, and segmentation. Full article
Show Figures

Figure 1

19 pages, 2169 KiB  
Article
Age and Gender Recognition Using a Convolutional Neural Network with a Specially Designed Multi-Attention Module through Speech Spectrograms
by Anvarjon Tursunov, Mustaqeem, Joon Yeon Choeh and Soonil Kwon
Sensors 2021, 21(17), 5892; https://doi.org/10.3390/s21175892 - 1 Sep 2021
Cited by 75 | Viewed by 9740
Abstract
Speech signals are being used as a primary input source in human–computer interaction (HCI) to develop several applications, such as automatic speech recognition (ASR), speech emotion recognition (SER), gender, and age recognition. Classifying speakers according to their age and gender is a challenging [...] Read more.
Speech signals are being used as a primary input source in human–computer interaction (HCI) to develop several applications, such as automatic speech recognition (ASR), speech emotion recognition (SER), gender, and age recognition. Classifying speakers according to their age and gender is a challenging task in speech processing owing to the disability of the current methods of extracting salient high-level speech features and classification models. To address these problems, we introduce a novel end-to-end age and gender recognition convolutional neural network (CNN) with a specially designed multi-attention module (MAM) from speech signals. Our proposed model uses MAM to extract spatial and temporal salient features from the input data effectively. The MAM mechanism uses a rectangular shape filter as a kernel in convolution layers and comprises two separate time and frequency attention mechanisms. The time attention branch learns to detect temporal cues, whereas the frequency attention module extracts the most relevant features to the target by focusing on the spatial frequency features. The combination of the two extracted spatial and temporal features complements one another and provide high performance in terms of age and gender classification. The proposed age and gender classification system was tested using the Common Voice and locally developed Korean speech recognition datasets. Our suggested model achieved 96%, 73%, and 76% accuracy scores for gender, age, and age-gender classification, respectively, using the Common Voice dataset. The Korean speech recognition dataset results were 97%, 97%, and 90% for gender, age, and age-gender recognition, respectively. The prediction performance of our proposed model, which was obtained in the experiments, demonstrated the superiority and robustness of the tasks regarding age, gender, and age-gender recognition from speech signals. Full article
Show Figures

Figure 1

Back to TopTop