Next Article in Journal
A Method for the Front-End Design of Electric SUVs Integrating Kansei Engineering and the Seagull Optimization Algorithm
Previous Article in Journal
Building Consensus with Enhanced K-means++ Clustering: A Group Consensus Method Based on Minority Opinion Handling and Decision Indicator Set-Guided Opinion Divergence Degrees
Previous Article in Special Issue
Secure and Lightweight Firmware Over-the-Air Update Mechanism for Internet of Things
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Enhancing Subband Speech Processing: Integrating Multi-View Attention Module into Inter-SubNet for Superior Speech Enhancement

Department of Electrical Engineering, National Chi Nan University, Nantou 545, Taiwan
*
Author to whom correspondence should be addressed.
Electronics 2025, 14(8), 1640; https://doi.org/10.3390/electronics14081640
Submission received: 23 February 2025 / Revised: 10 April 2025 / Accepted: 16 April 2025 / Published: 18 April 2025
(This article belongs to the Special Issue IoT Security in the Age of AI: Innovative Approaches and Technologies)

Abstract

:
The Inter-SubNet speechenhancement network improves subband interaction by enabling the exchange of complementary information across frequency bands, ensuring robust feature refinement while significantly reducing computational load through lightweight, subband-specific modules. Despite its compact design, it outperforms state-of-the-art models such as FullSubNet, FullSubNet+, Conv-TasNet, and DCCRN+, offering a highly efficient yet powerful solution. To further enhance its performance, we propose integrating a Multi-view Attention (MA) module as a front-end or intermediate component. The MA module utilizes attention mechanisms across channel, global, and local views to emphasize critical features, ensuring comprehensive speech signal processing. Evaluations on the VoiceBank-DEMAND dataset show that incorporating the MA module significantly improves metrics like SI-SNR, PESQ, and STOI, demonstrating its effectiveness in enhancing subband feature extraction and overall speech enhancement performance.

1. Introduction

Speech enhancement (SE) has seen remarkable progress in recent years, evolving from traditional statistics-based methods to advanced deep learning techniques. This transition has significantly improved the ability to enhance speech intelligibility and quality in complex noisy environments. Deep Neural Networks (DNNs) have been at the forefront of this revolution, enabling models to learn intricate patterns from large datasets and outperform conventional approaches in separating speech from noise [1,2,3,4,5,6].
The field has witnessed several breakthrough innovations. Hybrid architectures, such as the TRAMBA model [7], combine transformers and Mamba deep learning models to enhance both acoustic and bone conduction speech, achieving high-quality enhancement with a small memory footprint suitable for mobile devices. End-to-end systems have gained prominence, with frameworks like [8] mapping noisy audio directly to enhanced audio without intermediate representations. Attention mechanisms have further refined SE performance by dynamically focusing on relevant input features [9].
Innovative models like GaGNet [10] and TaylorSENet [11] have pushed the boundaries of the field by employing guided learning mechanisms and efficient gating structures for real-time processing. Mixed-domain approaches, combining time-domain and time–frequency-domain frameworks, have enhanced both magnitude and phase representations of speech signals, yielding superior results across various metrics.
To meet real-time demands, researchers have optimized DNN architectures for efficiency. For instance, [12] introduced a lightweight network that maintains high performance while significantly reducing computational complexity, making it ideal for practical applications.
State-of-the-art deep learning SE techniques can be broadly categorized into several approaches:
  • Masking-based methods estimate time–frequency masks to suppress noise and enhance speech. Recent advancements include complex-valued masks for modifying both magnitude and phase [13], attention mechanisms focusing on relevant time-frequency regions [14], and multi-task learning to jointly optimize speech quality and intelligibility [15].
  • Mapping-based techniques directly estimate clean speech features from noisy inputs. Modern approaches encompass end-to-end models operating on raw waveforms [16], generative adversarial networks (GANs) for realistic speech reconstruction [17], and self-supervised pre-training leveraging large unlabeled datasets [18].
  • Advanced architectures pushing the state-of-the-art include Conformer-based models combining convolution and self-attention [19], diffusion probabilistic models for high-quality speech generation [20], and adaptive convolution networks that dynamically adjust to input characteristics [21].
In particular, when using a complex-valued STFT spectrogram as the main feature for speech enhancement, a series of advanced frameworks have been presented recently. For example, CFTNet (Complex Frequency Transformation Network) [22] leverages a complex-valued U-Net and frequency transformation blocks to capture contextual information, achieving significant improvements over state-of-the-art methods by effectively recovering speech from noisy spectrograms without creating musical artifacts. AERO [23] (Audio Super-Resolution in the Spectral Domain) is an audio super-resolution model operating in the spectral domain. It uses an encoder–decoder architecture with U-Net-like skip connections to process complex-valued spectrograms. AERO optimizes performance with a combination of time- and frequency-domain loss functions, achieving superior results in speech and music enhancement across various sample rates. DCCTN (Deep Complex Convolution Transformer Network) [24] is a speech enhancement framework designed to address challenges faced by cochlear implant (CI) users in noisy environments by jointly enhancing both magnitude and phase components of speech signals. DCCTN combines a complex-valued U-Net with a transformer-based bottleneck layer and frequency transformation blocks to capture harmonic correlations and contextual time–frequency details, enabling accurate reconstruction of clean speech from noisy inputs. Unlike conventional methods that focus solely on magnitude, DCCTN employs multi-task learning to optimize spectral accuracy and perceptual quality simultaneously.
Among recent advancements in deep neural network-based speech enhancement systems, Inter-SubNet [25] stands out for its innovative approach. This architecture refines and streamlines its predecessors, FullSubNet [26] and FullSubNet+ [27], achieving superior performance with a significantly smaller model size. Although it primarily uses magnitude spectrograms as input, Inter-SubNet is also a complex-valued speech enhancement network that effectively addresses phase reconstruction. This is made possible by its use of a complex ideal ratio mask (cIRM) [13], which modifies both the magnitude and phase components of the original complex spectrogram. The cIRM is generated through a hierarchical process involving the SubInter modules and LSTM layers, which process the input magnitude spectrogram. By applying the cIRM to the complex spectrogram, Inter-SubNet enhances both the amplitude and temporal characteristics of the speech signal, showcasing its ability to handle phase reconstruction efficiently. This approach enables Inter-SubNet to benefit from complex-valued processing while maintaining a lightweight architecture suitable for real-time applications.
Inter-SubNet’s key innovation lies in balancing high-quality enhancement with reduced model complexity, addressing critical challenges in real-time processing and deployment on resource-constrained devices. Rigorous evaluations using the DNS Challenge dataset have consistently demonstrated Inter-SubNet’s superiority over other leading methods, including DCCRN-E [28], Conv-TasNet [16], PoCoNet [29], DCCRN+ [30], TRU-Net [31], and CTS-Net [32].
This study proposes enhancing Inter-SubNet by incorporating a Multi-view Attention (MA) network, a crucial element of the MANNER architecture [33]. The MA module extracts and combines key information from input spectrograms across three dimensions: global time context, local spectral patterns, and channel-wise features. This comprehensive approach aims to enhance the model’s capacity to identify and maintain speech components while effectively reducing noise.
In particular, we introduce two integration strategies:
  • IS-IMA (Inter-SubNet with Internal MA): Incorporates the MA module at the initial stage of Inter-SubNet.
  • IS-EMA (Inter-SubNet with External MA): Positions the MA module at an intermediate stage of Inter-SubNet.
These enhanced models are expected to achieve more refined and precise speech enhancement by employing attention mechanisms designed to concentrate on the most significant spectrogram elements. This integration combines Inter-SubNet’s efficient structure with MA’s advanced feature extraction capabilities, potentially leading to improved performance in speech enhancement tasks.
To briefly sum up, the presented new framework in this study has the following key contributions:
  • Novel Integration Strategies: We propose two distinct approaches (IS-IMA and IS-EMA) for integrating the MA module into the Inter-SubNet framework, enhancing its speech enhancement capabilities.
  • Improved Performance: IS-IMA and IS-EMA demonstrate improvements over the baseline Inter-SubNet across various speech enhancement metrics, particularly in PESQ, STOI, and SI-SDR.
  • Efficient Architecture: The proposed models achieve performance gains with a relatively small increase in model parameters, maintaining computational efficiency.
  • Comprehensive Ablation Study: We conduct a detailed analysis of different attention configurations (channel-only, global-only, local-only) for both IS-IMA and IS-EMA, providing insights into the effectiveness of each attention mechanism.
  • Adaptability: The IS-EMA model, particularly, shows strong performance when isolating specific attention views, demonstrating its adaptability to different speech enhancement requirements.
The remainder of this paper is organized as follows: Section 2 and Section 3 provide a brief overview of the two backbone modules, Inter-SubNet and the MA block. Section 4 introduces the two newly proposed architectures that integrate Inter-SubNet and MA. The experimental setup, results, and discussions regarding the comparison methods are presented in Section 5 and Section 6. Lastly, Section 7 concludes the paper with a summary.

2. Inter-SubNet

This section outlines the core architecture of the Inter-SubNet framework, which forms the basis for subsequent enhancements. As depicted in the left part of Figure 1, the Inter-SubNet pipeline operates through the following key stages:

2.1. Signal Transformation

The input audio signal x [ n ] undergoes conversion to a complex spectrogram representation via the short-time Fourier transform (STFT), yielding X = X r + j X i R F × T , where F and T represent frequency bins and temporal frames, respectively. The magnitude spectrogram X m is subsequently extracted through element-wise absolute value computation.

2.2. Spectral Decomposition

Through a frequency-unfolding operation, X m is partitioned into overlapping subband units { b f } f = 1 F . Each subband b f aggregates spectral information from N adjacent bins through concatenation:
b f = X f N m ; ; X f m ; ; X f + N m R ( 2 N + 1 ) × T ,
with circular boundary handling for edge frequencies.

2.3. Hierarchical Feature Processing

The subband units undergo sequential processing through two SubInter-LSTM (SIL) blocks followed by a fully connected output layer to generate the complex ideal ratio mask (cIRM). Each SIL block integrates three core components: SubInter Module, LSTM Network, and Group Normalization (G-Norm). The SubInter module, essential for extracting both local and global spectral information, is shown in the right part of Figure 1 and functions as described below:
  • Feature Transformation and Initial Processing:
    Input subband units undergo dimension transformation and linear processing:
    { h f } f = 1 F = F transform ( { b f } f = 1 F ) ,
    where { b f } f = 1 F is the initial hidden representation and F is a linear layer.
  • Global Spectral Information Aggregation:
    Cross-band dependencies are captured by averaging and processing the hidden representation:
    h ˜ = R 1 F f = 1 F h f ,
    where R is another linear layer and h ˜ encapsulates global spectral information.
  • Feature Fusion and Enhancement:
    Local and global information are combined, processed, and added to the original input:
    { b ^ f } i = 1 F = { b f } i = 1 F + transform P ( { [ h f , h ˜ ] } f = 1 F ) ,
    where P is a linear layer and { b ^ f } i = 1 F are the enhanced features.

2.4. Final Processing and Output

The enhanced features are processed through an LSTM network and normalized. A complex ideal ratio mask (cIRM) is generated and applied to the input spectrogram:
X ^ = ( M r + j M i ) ( X r + j X i ) .
The enhanced spectrogram X ^ is converted back to the time domain, producing the enhanced signal x ^ [ n ] .

3. Multi-View Attention Module

The Multi-view Attention (MA) block, a key component of the MANNER speech enhancement framework [33], operates within a convolutional encoder–decoder framework. Its primary function is to extract and process three distinct perspectives of the noisy speech signal: channel view, global view, and local view.
By leveraging attention mechanisms across these multiple viewpoints, the MA block highlights critical features, resulting in a more comprehensive and precise representation of the speech signal. Specifically, channel attention enhances representations derived from compressed channels, while a dual-path mechanism incorporating both global and local attention effectively captures extended sequential features. This multi-faceted approach allows the MA block to synthesize a rich, context-aware representation of the speech signal, potentially leading to more effective noise suppression and speech enhancement. The structure and operation of the MA block are visually depicted in Figure 2.
The MA block in the MANNER framework [33] transforms latent representations through three parallel processing pathways. The input signal first undergoes channel reduction through convolutional layers that compress the channel dimension from C to C / 3 in each pathway. For global and local processing streams, the reduced features are partitioned into overlapping chunks through a 50 % overlap operation, reformatting the input as x R C / 3 × P × S where P denotes chunk size and S the number of chunks. This structural separation enables efficient handling of both extended temporal patterns and localized spectral features.
The channel attention pathway enhances compressed channel representations through dual pooling operations. For input x c R C / 3 × L , average and max pooling generate aggregated descriptors x c a v g and x c m a x . These descriptors are processed through a dense network with C / 6 hidden nodes and sigmoid activation to compute attention weights:
α c = σ W 1 W 0 x c a v g + W 1 W 0 x c m a x ,
where W 0 and W 1 represent network parameters. The resulting weights are broadcast temporally and applied to the original features to produce refined channel representations x c .
The global attention pathway employs Transformer-based multi-head self-attention on chunked inputs x g R C / 3 × P × S . The mechanism computes the following:
x g = W softmax Q K d k V ,
where the Q, K, and V matrices derive from linear projections of the input, enabling cross-chunk dependency modeling through scaled dot-product attention.
Local feature processing utilizes depthwise convolution with S / 2 1 kernels on chunked inputs x l R P × C / 3 × S . Concatenated average and max pooling outputs feed into a convolutional layer generating attention weights:
α l = σ ( F ( [ x l a v g ; x l m a x ] ) ) .
These weights modulate the original features through element-wise multiplication:
x l = x l × α l .
Final integration occurs through concatenation of all pathway outputs followed by convolutional blending. A residual gating mechanism combining sigmoid, tanh, and ReLU activations refines the combined features before residual summation with the original input completes the transformation. This multi-view architecture enables simultaneous exploitation of channel-wise relationships, global temporal dependencies, and local spectral patterns for comprehensive speech representation learning.

4. Presented Method: MA-Equipped Inter-SubNet

In this work, we propose integrating the MA module into Inter-SubNet to achieve improved speech enhancement performance. Two integration strategies are designed, as illustrated in Figure 3 and Figure 4, and are described as follows:
  • Strategy I: IS-IMA (Inter-SubNet with Internal MA)
    This method involves embedding the MA block within the first SubInter-LSTM (SIL) block of the Inter-SubNet, shown in Figure 3. Specifically, we insert it immediately after the linear layer F in the SubInter module. This positioning enables the MA block to process the transformed features prior to their input into the subsequent interaction layers.
    This approach embeds the MA block within the first SubInter-LSTM (SIL) block of the Inter-SubNet, as shown in Figure 3. Specifically, the MA block is inserted immediately after the linear layer F in the SubInter module. This positioning allows the MA block to process transformed subband features before they are passed to subsequent interaction layers. By placing the attention mechanism early in the pipeline, IS-IMA emphasizes critical spectral features at an early stage, ensuring that downstream layers receive refined inputs. This strategy is particularly effective in filtering irrelevant noise and enhancing feature representations for subsequent processing.
  • Strategy II: IS-EMA (Inter-SubNet with External MA)
    In contrast, IS-EMA positions the MA block directly after the first complete SIL block of the Inter-SubNet, as depicted in Figure 4. This arrangement allows the MA block to process intermediate features that already incorporate local and global spectral information extracted by the first SIL block. By refining these structured representations, IS-EMA aligns enhanced features with the needs of the speech enhancement task before passing them to the second SIL block. This strategy leverages both low-level and high-level feature interactions, enabling a more comprehensive enhancement process.
The strategies for integrating MA mechanisms into the InterSubnet framework show key differences that affect how they process information. These differences mainly relate to where the attention mechanism is placed in the signal pathway and how they handle features:
  • The first major difference is where the MA module is located in relation to feature transformations. In the IS-IMA approach, the MA block works right after the linear transformation of sub-channel features, meaning it processes pure sub-band features. In contrast, IS-EMA waits until after the first SIL block to apply MA, which means it deals with features that already include both sub-band details and global spectral patterns.
  • The second difference is how they interact with residual pathways. IS-IMA uses a parallel processing method where the original sub-band features skip the MA-enhanced main pathway through residual connections. On the other hand, IS-EMA places its MA module before the second SIL block, allowing both main and residual pathways to be influenced by attention-modified features.
These structural differences lead to different learning behaviors during model training. IS-IMA’s approach keeps raw sub-band features in residual paths, which might improve gradient flow and maintain high-frequency details. Meanwhile, IS-EMA’s use of MA could enable more complex interactions between features but may also increase complexity. Testing these approaches would help understand their trade-offs across various applications and datasets.
In the following, we further investigate the potential advantage and limitation of the presented two strategies:
  • Strategy I places the MA block before the first SIL block in Inter-SubNet, enabling it to preprocess input features. This early placement helps the MA block emphasize important features and filter out irrelevant ones, providing SIL blocks with cleaner inputs, which can improve performance in noisy conditions.
    However, this approach has drawbacks. Since the MA block processes raw input, it may miss out on the refined sub-band features produced by the initial SIL block. As a result, it might focus on less useful or noisy elements, reducing its effectiveness and causing weaker interaction with later SIL blocks, limiting performance improvements.
  • Strategy II places the MA module between the two SIL blocks in Inter-SubNet, refining intermediate features by processing the structured subband representations from the first SIL block. This helps the MA module focus on both local and global spectral information, aligning enhanced features with the speech enhancement task before passing them to the second SIL block, potentially improving performance.
    However, since it skips raw input features and processes only intermediate outputs, Strategy II may have reduced effectiveness in refining low-level features and could introduce inefficiencies in the processing pipeline.

5. Experimental Setup

To assess the effectiveness of the proposed method, we utilize the VoiceBank-DEMAND dataset [34,35], which combines speech from VoiceBank with noise from the DEMAND database. The training set consists of 11,572 utterances from 28 speakers, while the test set includes 824 utterances from 2 speakers. The training data are corrupted with ten types of DEMAND noise at four signal-to-noise ratios (SNRs): 0, 5, 10, and 15 dB. The test set uses five different DEMAND noises at SNRs of 2.5, 7.5, 12.5, and 17.5 dB, with approximately 200 utterances reserved for validation.
For the Inter-SubNet framework, we set the following parameters: a Hanning window of 32 ms frame size with a 16 ms frame shift is used to generate spectrograms; the sub-band unit consists of 2 N + 1 = 31 frequency bins; the input–target sequence length during training is T = 192 frames; the first SIL block has 102 hidden units in the SubInter module and 384 hidden units in the LSTM; the second SIL block has 307 hidden units in the SubInter module and 384 hidden units in the LSTM; and the batch size is reduced to two due to GPU limitations.
The MA block parameters include a channel count (CM) of 512, four attention heads, attention dimensions of 256, and an attention dropout rate of 0.1.
To assess the SE performance of the comparison models, several objective metrics are used:
  • Perceptual Evaluation of Speech Quality (PESQ) [36]: This metric measures the quality difference between enhanced and clean speech signals, ranging from 0.5 to 4.5, with higher values indicating better quality.
  • Short-time objective intelligibility (STOI) [37]: STOI evaluates the intelligibility of short-time time–frequency regions in an utterance using discrete-time Fourier transform (DFT). It ranges from 0 to 1, where higher scores indicate better intelligibility.
  • Scale-invariant signal-to-distortion ratio (SI-SNR) [38]: This metric quantifies the distortion between the processed utterance s ^ and its clean counterpart s . It is calculated as follows:
    S I - S D R = 10 log 10 s t a r g e t 2 e n o i s e 2 ,
    where
    s t a r g e t = s ^ , s s s 2 ,
    e n o i s e = s ^ s t a r g e t ,
    with s ^ , s being the inner product of s ^ and s .
  • Composite Measure for Overall Quality (COVL) [39]: COVL predicts the Mean Opinion Score (MOS) for overall speech quality, ranging from 0 to 5. Higher scores reflect better perceived quality by listeners.
  • Composite Measure for Signal Distortion (CSIG): [37]: CSIG predicts MOS for signal distortion, also ranging from 0 to 5. Higher scores indicate less distortion and better preservation of the original speech signal during enhancement.

6. Results and Discussions

6.1. Overall Performance Evaluation

Table 1 presents the performance comparison of three models: Inter-SubNet (baseline), IS-IMA, and IS-EMA, across various metrics commonly used in speech enhancement (SE) tasks. Below is an analysis and discussion of the results:
  • In terms of PESQ, which measures perceived speech quality, both IS-IMA and IS-EMA demonstrate slight improvements over the baseline, achieving a score of 3.590 compared to 3.584. Although the improvement is marginal, it indicates progress in enhancing the perceived quality of speech.
  • For STOI, a metric that evaluates speech intelligibility, both proposed models outperform the baseline, with IS-IMA scoring 0.942 and IS-EMA achieving the highest score of 0.946. This suggests that IS-EMA significantly enhances intelligibility compared to both the baseline and IS-IMA. However, when it comes to SI-SDR, which measures signal reconstruction accuracy, the results are mixed. While IS-IMA slightly underperforms relative to the baseline (18.504 versus 18.646), IS-EMA surpasses both with a score of 18.782, indicating its superior capability in reconstructing signals.
  • The COVL metric, which reflects overall quality, presents a different trend. IS-IMA improves upon the baseline score of 3.429 by reaching 3.469, but IS-EMA experiences a decline to 3.286, suggesting a trade-off in overall quality for other performance gains. Similarly, for CSIG, which evaluates signal distortion, IS-IMA achieves the best score at 4.058, surpassing both the baseline (4.012) and IS-EMA (3.959). This highlights IS-IMA’s strength in handling signal distortion effectively.
  • The placement of the MA module significantly affects performance. IS-EMA’s late-stage MA integration enhances speech patterns critical for intelligibility (STOI) by refining intermediate features. In contrast, IS-IMA’s early MA placement focuses on raw feature filtering, effectively reducing noise but potentially limiting later refinements, which improves distortion reduction (CSIG).
  • The MA placement also influences how each model processes spectral information. IS-EMA’s ability to refine structured features aligns with the need for clear speech patterns in low-SNR conditions, improving intelligibility. IS-IMA’s early attention maintains overall quality by reducing noise early on, though it may not capture nuanced speech details as effectively as IS-EMA.
  • In terms of model complexity, as measured by the number of parameters (#Para.), Inter-SubNet is the most lightweight with only 2.29M parameters. However, this simplicity comes at the cost of lower performance in most metrics compared to the proposed models. IS-IMA increases complexity to 3.06M parameters, while IS-EMA has the highest parameter count at 3.81M. This increase in complexity correlates with improved performance in some areas but also raises computational demands.
In summary, while both IS-IMA and IS-EMA outperform Inter-SubNet in various aspects, their strengths differ depending on the metric considered. IS-EMA excels in intelligibility (STOI) and signal reconstruction (SI-SDR), whereas IS-IMA demonstrates stronger performance in overall quality (COVL) and reduced signal distortion (CSIG). These results highlight trade-offs between different performance metrics and model complexity across the three approaches.

6.2. Performance Evaluation Across Individual SNR Conditions and Low-SNR Scenarios

We present a comprehensive evaluation of the proposed IS-IMA and IS-EMA models under individual SNR conditions ranging from 17.5 dB to 2.5 dB using the standard test set. To ensure clarity and readability, we report results for the key metrics—PESQ, STOI, and SI-SDR—summarized in Table 2, Table 3 and Table 4, respectively.
To further evaluate the robustness of the proposed methods, additional experiments were conducted using manually prepared noisy test sets at 0 dB and −5 dB SNR levels. These results are also included in Table 2, Table 3 and Table 4.
From these tables, we have the following findings:
  • Performance under standard SNR conditions:
    • PESQ:
      (a)
      High SNR (17.5 dB and 12.5 dB): IS-IMA achieves PESQ scores comparable to Inter-SubNet, indicating minimal degradation in clean conditions. For instance, at 17.5 dB, IS-IMA scores 3.904 versus the baseline’s 3.910.
      (b)
      Mid SNR (7.5 dB): Both IS-IMA and IS-EMA exhibit improvements over the baseline, with IS-IMA achieving the highest PESQ score of 3.530, reflecting its ability to enhance perceptual quality in moderately noisy environments.
      (c)
      Low SNR (2.5 dB): IS-EMA slightly outperforms IS-IMA and the baseline, achieving a PESQ score of 3.210 compared to Inter-SubNet’s 3.198. This suggests that IS-EMA is better suited for handling challenging noise conditions while preserving speech quality.
    • STOI:
      IS-EMA consistently outperforms both IS-IMA and Inter-SubNet across all standard SNR levels in terms of STOI, achieving the highest intelligibility scores in nearly every condition:
      (a)
      At 17.5 dB, IS-EMA achieves an STOI score of 0.970 compared to Inter-SubNet’s 0.967, showcasing its ability to maintain intelligibility even in clean conditions.
      (b)
      At 7.5 dB, IS-EMA reaches an STOI score of 0.947, surpassing both IS-IMA (0.942) and the baseline (0.943). This demonstrates its robustness in moderately noisy scenarios where clarity is critical.
    • SI-SDR:
      IS-EMA achieves higher SI-SDR scores than both Inter-SubNet and IS-IMA across most standard SNR levels, with notable differences at mid-to-low SNRs:
      (a)
      At 12.5 dB, IS-EMA achieves an SI-SDR of 19.866 compared to Inter-SubNet’s 19.759, reflecting its ability to reconstruct high-fidelity signals under moderate noise conditions.
      (b)
      At 7.5 dB, IS-EMA achieves an SI-SDR of 18.246 versus the baseline’s 18.056, indicating its robustness in preserving speech signal fidelity under noisy environments.
  • Performance under extra-low-SNR conditions:
    Both IS-IMA and IS-EMA demonstrate measurable improvements over the baseline at critical low-SNR conditions:
    • PESQ:
      (a)
      At 0 dB SNR, IS-IMA achieves a PESQ of 3.065 (+0.042 over baseline 3.023), demonstrating superior perceptual quality in moderate noise, while IS-EMA matches the baseline (3.023). This highlights IS-IMA’s early MA integration effectively filters raw subband features, enhancing speech quality prior to downstream processing.
      (b)
      At −5 dB SNR, IS-EMA slightly outperforms the baseline (2.788 vs. 2.773), but IS-IMA achieves a higher PESQ of 2.819 (+1.7% relative improvement). This underscores IS-IMA’s robustness in extreme noise, leveraging early MA integration to preserve perceptual quality under severe acoustic challenges.
    • STOI:
      IS-EMA consistently achieves superior STOI scores over IS-IMA and Inter-SubNet in low-SNR conditions, demonstrating robust speech intelligibility preservation even under heavy noise. This stems from its late-stage MA module integration, which refines intermediate features enriched by the initial SIL block’s local–global spectral analysis.
    • SI-SDR:
      Similarly, IS-EMA demonstrates superior SI-SDR performance compared to both Inter-SubNet and IS-IMA at 0 dB and −5 dB SNR levels. This highlights its ability to effectively handle signal distortion in challenging noise environments. By leveraging its late-stage MA integration, IS-EMA refines intermediate features, ensuring robust reconstruction of speech signals even under severe acoustic interference, making it highly effective in low-SNR scenarios.

6.3. Ablation Analysis of Attention Mechanisms

The ablation study evaluates the performance of IS-IMA and IS-EMA under different view attention configurations in their MA blocks.
  • IS-IMA Results (Table 5):
    The baseline model achieves balanced performance across the following metrics: PESQ (3.590), STOI (0.942), SI-SDR (18.504), and COVL (3.469), with 3.06k parameters.
    Channel-only attention slightly improves PESQ (3.586 → 3.592) and SI-SDR (18.504 → 18.569), with a small increase in CSIG (4.058 → 4.075).
    Global-only attention achieves the highest PESQ (3.592) and SI-SDR (18.616) but reduces STOI and CSIG slightly.
    Local-only attention delivers the highest SI-SDR (18.625) but performs worse in PESQ (3.535), STOI (0.939), and COVL (3.296).
  • IS-EMA Results (Table 6):
    The baseline model achieves strong performance across the following metrics: PESQ (3.590), STOI (0.946), SI-SDR (18.787), COVL (3.386), and CSIG (3.959), with 3.82k parameters.
    Channel-only attention improves PESQ to 3.624 and SI-SDR to 18.801 while maintaining STOI at 0.946; COVL and CSIG also increase to 3.549 and 4.138, respectively.
    Global-only attention achieves a similar PESQ score (3.623) but lowers STOI to 0.941 and SI-SDR to 18.177, though it slightly improves COVL to 3.449.
    Local-only attention produces the highest PESQ (3.637) and improves CSIG to 4.143 but slightly reduces SI-SDR to 18.552.
Overall, isolating specific views in IS-EMA delivers better results than in IS-IMA, with stronger improvements across key metrics like PESQ, SI-SDR, and CSIG:
  • Channel-only attention in IS-EMA achieves a PESQ of 3.624 and SI-SDR of 18.801 while maintaining STOI at 0.946, outperforming IS-IMA in similar setups.
  • Local-only attention in IS-EMA excels with the highest PESQ (3.637) and CSIG (4.143), demonstrating its ability to enhance speech quality and reduce distortion.
In contrast, isolating views in IS-IMA shows limited improvements with greater trade-offs:
  • Global-only attention achieves the highest PESQ (3.592) and SI-SDR (18.616) but reduces STOI and CSIG compared to the baseline.
  • Local-only attention performs poorly in PESQ (3.535) and COVL (3.296), highlighting its limitations.
Thus, while integrating all three views ensures balanced performance for both models, IS-EMA’s isolated configurations demonstrate greater adaptability and effectiveness, making it a stronger choice for tasks requiring specific attention mechanisms.
Table 5. Ablation study: the various SE performance scores of IS-IMA with different assignments of view attention in the MA block.
Table 5. Ablation study: the various SE performance scores of IS-IMA with different assignments of view attention in the MA block.
IS-IMAPESQSTOISI-SDRCOVLCSIG# Para. (M)
all three views (baseline)3.5900.94218.5043.4694.0583.06
channel only3.5860.94218.5693.4884.0752.82
global only3.5920.94118.6163.4494.0032.84
local only3.5350.93918.6253.2963.8422.81
Table 6. Ablation study: the various SE performance scores of IS-EMA with different assignments of view attention in the MA block.
Table 6. Ablation study: the various SE performance scores of IS-EMA with different assignments of view attention in the MA block.
IS-EMAPESQSTOISI-SDRCOVLCSIG# Para. (M)
all three views (baseline)3.5900.94618.7873.3863.9593.82
channel only3.6240.94618.8013.5494.1383.81
global only3.6230.94118.1773.4494.0083.81
local only3.6370.94618.5523.5494.1433.78

Results of Ablation Study for Each SNR Level

Following the methodology outlined in Section 6.2, we conducted an ablation study evaluating IS-IMA and IS-EMA under isolated attention views across individual SNR conditions. These results, presented in Table 7, Table 8 and Table 9, reveal the following key findings:
  • IS-EMA’s Local-Only MA Configuration:
    Achieves the highest PESQ scores across almost all SNR levels, demonstrating superior speech quality enhancement in both standard and low-SNR scenarios
    Maintains robust STOI performance (0.923 at 0 dB, 0.895 at −5 dB), outperforming IS-IMA in preserving speech intelligibility under extreme noise
  • IS-IMA’s Global-Only MA Configuration:
    Excels at mid-SNR conditions (7.5 dB: PESQ 3.53) but struggles in low-SNR scenarios (−5 dB: PESQ 2.793), highlighting limitations in severe noise environments.
  • SI-SDR Trade-Offs:
    IS-EMA’s local-only setup dominates extreme SNR scenarios (−5 dB: 8.241), while IS-IMA’s channel-only configuration performs best at 0 dB (10.589).
These findings underscore the importance of MA placement and attention view selection, with IS-EMA demonstrating superior adaptability to challenging noise conditions compared to IS-IMA. The late-stage MA integration in IS-EMA enables effective refinement of structured spectral features, particularly under low-SNR scenarios, while IS-IMA’s early integration prioritizes raw feature filtering with inherent noise robustness limitations.

6.4. Comparison with Some State-of-the-Art SE Methods

Finally, we compare the presented IS-IMA and IS-EMA variants with several state-of-the-art speech enhancement frameworks, including FullSubNet [26] and WaveUNet [40,41], which uses frequency- and time-domain features as input, respectively, SEGAN [17], CMGAN [42], and MP-SENet [43], which primarily utilize generative adversarial networks (GANs), along with MANNER [33] and SE-Conformer [44], which mainly employ attention mechanisms.
The results, as presented in Table 10, are based on the PESQ and STOI metrics obtained from the VoiceBank-DEMAND dataset, as reported in the literature. The findings indicate that IS-IMA and IS-EMA achieve the highest PESQ scores (3.59 and 3.64, respectively) among all evaluated methods, showcasing their superior speech quality enhancement capabilities. While STOI scores are generally comparable across most methods, FullSubNet, CMGAN, and MP-SENet achieve the highest STOI score of 0.96. Nevertheless, our proposed models demonstrate a strong balance between speech quality (PESQ) and intelligibility (STOI), highlighting their effectiveness in advancing speech enhancement performance.
These findings confirm that our methods perform better than or are competitive with state-of-the-art techniques, making them highly suitable for real-world applications. Notably, resource-intensive models like DCCTN [24] (10.1M parameters) are excluded from this comparison due to hardware constraints—our lab infrastructure cannot support the computational demands of such large architectures. In contrast, IS-IMA (3.06M parameters) and IS-EMA (3.81M parameters) emphasize parameter efficiency while retaining competitive enhancement quality, making them uniquely suited for real-time deployment in resource-limited settings.

7. Conclusions and Future Work

This study focuses on improving Inter-SubNet, a leading speech enhancement framework, by integrating a Multi-view Attention (MA) module. Experiments on the VoiceBank-DEMAND dataset show that the proposed IS-IMA and IS-EMA models significantly enhance Inter-SubNet’s performance across various metrics. Notably, isolating individual view attentions in IS-EMA achieves near-optimal results in multiple SE metrics, highlighting the effectiveness of the proposed model.
Furthermore, building on the insights from this study, we have several promising directions for future research:
  • Dual-MA Architecture: As suggested, integrating Multi-view Attention (MA) modules at both early and late stages of the Inter-SubNet framework could combine the strengths of IS-IMA (noise suppression) and IS-EMA (intelligibility enhancement). While this dual-MA configuration may introduce computational overhead, a targeted ablation study could evaluate whether the performance gains in distortion reduction and speech quality justify the increased complexity. This approach might better balance feature filtering and refinement across the processing pipeline.
  • Layer-Specific Attention Mechanisms: Exploring the strategic placement of different attention types (channel, global, local) at specific layers could optimize feature extraction. For instance, local attention in early layers might capture fine-grained phonetic details (e.g., formants, harmonics), while global attention in later layers could ensure temporal coherence of enhanced speech. Channel attention could be dynamically weighted across layers to emphasize critical frequency bands. Such a structured approach could yield more targeted enhancements while maintaining computational efficiency.
  • Dynamic Attention Allocation: Beyond static configurations, future work could investigate adaptive mechanisms where attention types are dynamically selected based on input SNR or spectral characteristics. For example, local attention might dominate in high-noise conditions, while global attention could activate in cleaner speech scenarios. This adaptability could further improve robustness across diverse acoustic environments.
These directions aim to advance the flexibility and efficacy of attention-based speech enhancement while addressing practical constraints such as computational cost and real-time applicability.

Author Contributions

Conceptualization, J.-W.H.; methodology, J.-W.H., T.-J.L. and B.-Y.S.; software, T.-J.L. and B.-Y.S.; validation, J.-W.H., T.-J.L. and B.-Y.S.; formal analysis, J.-W.H. and T.-J.L.; investigation, J.-W.H.; resources, J.-W.H.; data curation, J.-W.H., T.-J.L. and B.-Y.S.; writing—original draft preparation, J.-W.H.; writing—review and editing, J.-W.H.; visualization, J.-W.H., T.-J.L. and B.-Y.S.; supervision, J.-W.H.; project administration, J.-W.H.; funding acquisition, J.-W.H. All authors have read and agreed to the published version of the manuscript.

Funding

This research received no external funding.

Data Availability Statement

The original contributions presented in this study are included in the article. Further inquiries can be directed to the corresponding author.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Ochieng, P. Deep neural network techniques for monaural speech enhancement: State of the art analysis. arXiv 2023, arXiv:2212.00369. [Google Scholar] [CrossRef]
  2. Xu, L.; Zhang, T. Fractional feature-based speech enhancement with deep neural network. Speech Commun. 2023, 153, 102971. [Google Scholar] [CrossRef]
  3. Hao, X.; Xu, C.; Xie, L. Neural speech enhancement with unsupervised pre-training and mixture training. Neural Netw. 2023, 158, 216–227. [Google Scholar] [CrossRef] [PubMed]
  4. Wang, Y.; Han, J.; Zhang, T.; Qing, D. Speech Enhancement from Fused Features Based on Deep Neural Network and Gated Recurrent Unit Network. EURASIP J. Adv. Signal Process. 2021, 2021, 104. [Google Scholar] [CrossRef]
  5. Skariah, D.; Thomas, J. Review of Speech Enhancement Methods using Generative Adversarial Networks. In Proceedings of the 2023 International Conference on Control, Communication and Computing (ICCC), Brighton, UK, 12–17 May 2023; pp. 1–4. [Google Scholar] [CrossRef]
  6. Karjol, P.; Ajay Kumar, M.; Ghosh, P.K. Speech Enhancement Using Multiple Deep Neural Networks. In Proceedings of the 2018 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP), Calgary, AB, Canada, 15–20 April 2018; pp. 5049–5052. [Google Scholar] [CrossRef]
  7. Sui, Y.; Zhao, M.; Xia, J.; Jiang, X.; Xia, S. TRAMBA: A Hybrid Transformer and Mamba Architecture for Practical Audio and Bone Conduction Speech Super Resolution and Enhancement on Mobile and Wearable Platforms. arXiv 2024, arXiv:2405.01242. [Google Scholar] [CrossRef]
  8. Pang, J.; Li, H.; Jiang, T.; Wang, H.; Liao, X.; Luo, L.; Liu, H. A Dual-Channel End-to-End Speech Enhancement Method Using Complex Operations in the Time Domain. Appl. Sci. 2023, 13, 7698. [Google Scholar] [CrossRef]
  9. Fan, J.; Yang, J.; Zhang, X.; Yao, Y. Real-time single-channel speech enhancement based on causal attention mechanism. Appl. Acoust. 2022, 201, 109084. [Google Scholar] [CrossRef]
  10. Li, A.; Zheng, C.; Zhang, L.; Li, X. Glance and Gaze: A Collaborative Learning Framework for Single-channel Speech Enhancement. arXiv 2021, arXiv:2106.11789. [Google Scholar] [CrossRef]
  11. Li, A.; Liu, C.; Tobing, P.L.; Ren, Y.; Xu, Y.; Chen, Z.; Qian, Y.; Li, X. Taylor, Can You Hear Me Now? A Taylor-Unfolding Framework for Monaural Speech Enhancement. arXiv 2022, arXiv:2211.16764. [Google Scholar]
  12. Yang, L.; Liu, W.; Meng, R.; Lee, G.; Baek, S.; Moon, H.G. Fspen: An Ultra-Lightweight Network for Real Time Speech Enahncment. In Proceedings of the ICASSP 2024—2024 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP), Seoul, Republic of Korea, 14–19 April 2024; pp. 10671–10675. [Google Scholar] [CrossRef]
  13. Williamson, D.S.; Wang, Y.; Wang, D. Complex Ratio Masking for Monaural Speech Separation. IEEE/ACM Trans. Audio Speech Lang. Process. 2016, 24, 483–492. [Google Scholar] [CrossRef]
  14. Kim, J.; El-Khamy, M.; Lee, J. T-GSA: Transformer with Gaussian-weighted self-attention for speech enhancement. In Proceedings of the ICASSP 2020—2020 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP), Barcelona, Spain, 4–8 May 2020; IEEE: New York, NY, USA, 2020; pp. 6649–6653. [Google Scholar]
  15. Xu, Y.; Weng, C.; Hui, L.; Liu, J.; Yu, M.; Su, D. Joint training of complex ratio mask based beamformer and acoustic model for noise robust ASR. In Proceedings of the ICASSP 2019—2019 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP), Brighton, UK, 12–17 May 2019; IEEE: New York, NY, USA, 2019; pp. 6934–6938. [Google Scholar]
  16. Luo, Y.; Mesgarani, N. Conv-TasNet: Surpassing Ideal Time–Frequency Magnitude Masking for Speech Separation. IEEE/ACM Trans. Audio Speech Lang. Process. 2019, 27, 1256–1266. [Google Scholar] [CrossRef] [PubMed]
  17. Pascual, S.; Bonafonte, A.; Serra, J. SEGAN: Speech enhancement generative adversarial network. arXiv 2017, arXiv:1703.09452. [Google Scholar]
  18. Hsu, W.N.; Bolte, B.; Tsai, Y.H.H.; Lakhotia, K.; Salakhutdinov, R.; Mohamed, A. HuBERT: Self-supervised speech representation learning by masked prediction of hidden units. IEEE/ACM Trans. Audio Speech Lang. Process. 2021, 29, 3451–3460. [Google Scholar] [CrossRef]
  19. Gulati, A.; Qin, J.; Chiu, C.C.; Parmar, N.; Zhang, Y.; Yu, J.; Han, W.; Wang, S.; Zhang, Z.; Wu, Y.; et al. Conformer: Convolution-augmented transformer for speech recognition. arXiv 2020, arXiv:2005.08100. [Google Scholar]
  20. Kong, Z.; Ping, W.; Huang, J.; Zhao, K.; Catanzaro, B. DiffWave: A versatile diffusion model for audio synthesis. arXiv 2021, arXiv:2009.09761. [Google Scholar]
  21. Wang, D. Adaptive Convolution for CNN-based Speech Enhancement Models. arXiv 2025, arXiv:2502.14224. [Google Scholar]
  22. Mamun, N.; Hansen, J.H.L. CFTNet: Complex-valued Frequency Transformation Network for Speech Enhancement. In Proceedings of the INTERSPEECH 2023, Convention Centre Dublin, Dublin, Ireland, 20–24 August 2023; pp. 809–813. [Google Scholar]
  23. Mandel, M.; Tal, O.; Adi, Y. AERO: Audio Super Resolution in the Spectral Domain. arXiv 2023, arXiv:2211.12232. [Google Scholar]
  24. Mamun, N.; Hansen, J.H.L. Speech Enhancement for Cochlear Implant Recipients Using Deep Complex Convolution Transformer With Frequency Transformation. IEEE/ACM Trans. Audio Speech Lang. Process. 2024, 32, 2616–2629. [Google Scholar] [CrossRef]
  25. Chen, J.; Rao, W.; Wang, Z.; Lin, J.; Wu, Z.; Wang, Y.; Shang, S.; Meng, H. Inter-Subnet: Speech Enhancement with Subband Interaction. In Proceedings of the ICASSP 2023—2023 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP), Rhodes Island, Greece, 4–10 June 2023; pp. 1–5. [Google Scholar] [CrossRef]
  26. Hao, X.; Su, X.; Horaud, R.; Li, X. Fullsubnet: A Full-Band and Sub-Band Fusion Model for Real-Time Single-Channel Speech Enhancement. In Proceedings of the ICASSP 2021—2021 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP), Toronto, ON, Canada, 6–11 June 2021; pp. 6633–6637. [Google Scholar] [CrossRef]
  27. Chen, J.; Wang, Z.; Tuo, D.; Wu, Z.; Kang, S.; Meng, H. FullSubNet+: Channel Attention Fullsubnet with Complex Spectrograms for Speech Enhancement. In Proceedings of the ICASSP 2022—2022 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP), Singapore, 23–27 May 2022; pp. 7857–7861. [Google Scholar] [CrossRef]
  28. Hu, Y.; Liu, Y.; Lv, S.; Zhang, S.; Wu, J.; Zhang, B.; Xie, L. DCCRN: Deep Complex Convolution Recurrent Network for Phase-Aware Speech Enhancement. In Proceedings of the Interspeech 2020, Shanghai, China, 25–29 October 2020; pp. 2473–2477. [Google Scholar] [CrossRef]
  29. Isik, U.; Giri, R.; Phansalkar, N.; Valin, J.M.; Helwani, K.; Krishnaswamy, A. PoCoNet: Better Speech Enhancement with Frequency-Positional Embeddings, Semi-Supervised Conversational Data, and Biased Loss. In Proceedings of the Interspeech 2020, Shanghai, China, 25–29 October 2020; pp. 2487–2491. [Google Scholar]
  30. Lv, S.; Hu, Y.; Wu, J.; Xie, L. DCCRN+: Channel-wise Subband DCCRN with SNR Estimation for Speech Enhancement. In Proceedings of the Interspeech 2021, Brno, Czechia, 30 August—3 September 2021; pp. 2816–2820. [Google Scholar]
  31. Choi, H.S.; Park, S.; Lee, J.H.; Heo, H.; Jeon, D.; Lee, K. Real-Time Denoising and Dereverberation wtih Tiny Recurrent U-Net. In Proceedings of the ICASSP 2021—2021 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP), Toronto, ON, Canada, 6–11 June 2021; pp. 5789–5793. [Google Scholar] [CrossRef]
  32. Li, A.; Liu, W.; Zheng, C.; Fan, C.; Li, X. Two Heads are Better Than One: A Two-Stage Complex Spectral Mapping Approach for Monaural Speech Enhancement. IEEE/ACM Trans. Audio Speech Lang. Process. 2021, 29, 1829–1843. [Google Scholar] [CrossRef]
  33. Park, H.J.; Kang, B.H.; Shin, W.; Kim, J.S.; Han, S.W. MANNER: Multi-View Attention Network For Noise Erasure. In Proceedings of the ICASSP 2022—2022 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP), Singapore, 23–27 May 2022; IEEE: New York, NY, USA, 2022; pp. 7842–7846. [Google Scholar]
  34. Valentini-Botinhao, C.; Wang, X.; Takaki, S.; Yamagishi, J. Investigating RNN-based speech enhancement methods for noise-robust Text-to-Speech. In Proceedings of the 9th ISCA Workshop on Speech Synthesis Workshop (SSW 9), Sunnyvale, CA, USA, 13–15 September 2016; pp. 146–152. [Google Scholar] [CrossRef]
  35. Thiemann, J.; Ito, N.; Vincent, E. The Diverse Environments Multi-channel Acoustic Noise Database (DEMAND): A database of multichannel environmental noise recordings. In Proceedings of the 21st International Congress on Acoustics, Montreal, QC, Canada, 2–7 June 2013; pp. 1–6. [Google Scholar]
  36. ITU-T. Perceptual Evaluation of Speech Quality (PESQ), an Objective Method for End-to-End Speech Quality Assessment of Narrowband Telephone Networks and Speech Codecs; Technical Report P.862; International Telecommunication Union: Geneva, Switzerland, 2001. [Google Scholar]
  37. Taal, C.H.; Hendriks, R.C.; Heusdens, R.; Jensen, J. An Algorithm for Intelligibility Prediction of Time–Frequency Weighted Noisy Speech. IEEE Trans. Audio Speech Lang. Process. 2011, 19, 2125–2136. [Google Scholar] [CrossRef]
  38. Roux, J.L.; Wisdom, S.; Erdogan, H.; Hershey, J.R. SDR – Half-baked or Well Done? In Proceedings of the ICASSP 2019—2019 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP), Brighton, UK, 12–17 May 2019; pp. 626–630. [Google Scholar] [CrossRef]
  39. Hu, Y.; Loizou, P.C. Evaluation of Objective Quality Measures for Speech Enhancement. IEEE Trans. Audio Speech Lang. Process. 2008, 16, 229–238. [Google Scholar] [CrossRef]
  40. Pal, M.; Ramanathan, A.; Wada, T.; Pandey, A. Speech enhancement deep-learning architecture for efficient edge processing. arXiv 2024, arXiv:2405.16834. [Google Scholar]
  41. Daniel Stoller, S.E.; Dixon, S. Wave-U-Net: A Multi-Scale Neural Network for End-to-End Audio Source Separation. In Proceedings of the International Society for Music Information Retrieval Conference (ISMIR), Paris, France, 3–27 September 2018; pp. 334–340. [Google Scholar]
  42. Cao, R.; Abdulatif, S.; Yang, B. CMGAN: Conformer-based Metric GAN for Speech Enhancement. In Proceedings of the Interspeech 2022, Incheon, Republic of Korea, 18–22 September 2022; pp. 936–940. [Google Scholar] [CrossRef]
  43. Lu, Y.X.; Ai, Y.; Ling, Z.H. MP-SENet: A Speech Enhancement Model with Parallel Denoising of Magnitude and Phase Spectra. In Proceedings of the INTERSPEECH 2023, Dublin, Ireland, 20–24 August 2023; pp. 3834–3838. [Google Scholar] [CrossRef]
  44. Kim, E.; Seo, H. SE-Conformer: Time-Domain Speech Enhancement Using Conformer. In Proceedings of the Interspeech 2021, Brno, Czechia, 30 August–3 September 2021; pp. 2736–2740. [Google Scholar] [CrossRef]
Figure 1. The flowcharts of the Inter-SubNet and the SubInter module (redrawn according to [25]) Inter-SubNet processes the magnitude spectrum of the input signal through two SubInter-LSTM (SIL) blocks to predict a complex ideal ratio mask (cIRM), which is then applied to the original complex spectrogram to enhance speech.
Figure 1. The flowcharts of the Inter-SubNet and the SubInter module (redrawn according to [25]) Inter-SubNet processes the magnitude spectrum of the input signal through two SubInter-LSTM (SIL) blocks to predict a complex ideal ratio mask (cIRM), which is then applied to the original complex spectrogram to enhance speech.
Electronics 14 01640 g001
Figure 2. The flowchart of the MA block (redrawn according to [33]). This flowchart illustrates MA’s operation across three parallel pathways: channel view, time-axis attention (i.e., global attention), and frequency-axis attention (i.e., local attention). This multi-faceted approach captures critical features by emphasizing channel-wise relationships, extended temporal dependencies, and localized spectral patterns, leading to a comprehensive representation of the input signal.
Figure 2. The flowchart of the MA block (redrawn according to [33]). This flowchart illustrates MA’s operation across three parallel pathways: channel view, time-axis attention (i.e., global attention), and frequency-axis attention (i.e., local attention). This multi-faceted approach captures critical features by emphasizing channel-wise relationships, extended temporal dependencies, and localized spectral patterns, leading to a comprehensive representation of the input signal.
Electronics 14 01640 g002
Figure 3. The flowchart of the presented IS-IMA (Inter-SubNet with Internal MA).
Figure 3. The flowchart of the presented IS-IMA (Inter-SubNet with Internal MA).
Electronics 14 01640 g003
Figure 4. The flowchart of the presented IS-EMA (Inter-SubNet with External MA).
Figure 4. The flowchart of the presented IS-EMA (Inter-SubNet with External MA).
Electronics 14 01640 g004
Table 1. The various SE performance scores of Inter-SubNet baseline and the presented IS-IMA and IS-EMA. Bold numbers represent the best values among the compared methods.
Table 1. The various SE performance scores of Inter-SubNet baseline and the presented IS-IMA and IS-EMA. Bold numbers represent the best values among the compared methods.
PESQSTOISI-SDRCOVLCSIG#Para. (M)
Inter-SubNet3.5840.94118.6463.4294.0122.29
IS-IMA3.5900.94218.5043.4694.0583.06
IS-EMA3.5900.94618.7823.2863.9593.81
Table 2. The PESQ scores of Inter-SubNet baseline and the presented IS-IMA and IS-EMA at different SNRs. Bold numbers represent the best values among the compared methods.
Table 2. The PESQ scores of Inter-SubNet baseline and the presented IS-IMA and IS-EMA at different SNRs. Bold numbers represent the best values among the compared methods.
PESQ17.5 dB12.5 dB7.5 dB2.5 dB0 dB−5 dB
Inter-SubNet3.9103.7153.5183.1983.0232.773
IS-IMA3.9043.7233.5303.2083.0652.819
IS-EMA3.9023.7273.5273.2103.0232.788
Table 3. The STOI scores of Inter-SubNet baseline and the presented IS-IMA and IS-EMA at different SNRs. Bold numbers represent the best values among the compared methods.
Table 3. The STOI scores of Inter-SubNet baseline and the presented IS-IMA and IS-EMA at different SNRs. Bold numbers represent the best values among the compared methods.
STOI17.5 dB12.5 dB7.5 dB2.5 dB0 dB−5 dB
Inter-SubNet0.9670.9550.9430.9020.9140.883
IS-IMA0.9680.9560.9420.9020.9190.890
IS-EMA0.9700.9610.9470.9070.9240.896
Table 4. The SI-SDR scores of Inter-SubNet baseline and the presented IS-IMA and IS-EMA at different SNRs. Bold numbers represent the best values among the compared methods.
Table 4. The SI-SDR scores of Inter-SubNet baseline and the presented IS-IMA and IS-EMA at different SNRs. Bold numbers represent the best values among the compared methods.
SI-SDR17.5 dB12.5 dB7.5 dB2.5 dB0 dB−5 dB
Inter-SubNet21.83419.75918.05614.97810.4137.617
IS-IMA21.89219.69517.84414.63010.4127.643
IS-EMA21.97219.86618.24615.10910.3897.913
Table 7. Ablation study: the PESQ scores of Inter-SubNet, IS-IMA, and IS-EMA with different assignments of view attention in the MA block at different SNRs. Bold numbers represent the best values among the compared methods.
Table 7. Ablation study: the PESQ scores of Inter-SubNet, IS-IMA, and IS-EMA with different assignments of view attention in the MA block at different SNRs. Bold numbers represent the best values among the compared methods.
PESQ17.5 dB12.5 dB7.5 dB2.5 dB0 dB−5 dB
Inter-SubNet3.9103.7153.5183.1983.0232.773
IS-IMAall three views3.9043.7233.5303.2083.0652.819
channel only3.9213.7213.5173.1873.0402.793
global only3.9203.7253.5313.1963.0492.797
local only3.8703.6653.4643.1443.0392.792
IS-EMAall three views3.9023.7273.5273.2103.0232.788
channel only3.9483.7673.5583.2273.0362.789
global only3.9313.7563.5633.2473.0672.824
local only3.9513.7673.5693.2683.0712.815
Table 8. Ablation study: the STOI scores of Inter-SubNet, IS-IMA and IS-EMA with different assignments of view attention in the MA block at different SNRs. Bold numbers represent the best values among the compared methods.
Table 8. Ablation study: the STOI scores of Inter-SubNet, IS-IMA and IS-EMA with different assignments of view attention in the MA block at different SNRs. Bold numbers represent the best values among the compared methods.
STOI17.5 dB12.5 dB7.5 dB2.5 dB0 dB−5 dB
Inter-SubNet0.9670.9550.9430.9020.9140.883
IS-IMAall three views0.9680.9560.9420.9020.9190.890
channel only0.9680.9560.9430.9030.9220.894
global only0.9660.9540.9420.9010.9170.887
local only0.9670.9540.9390.8970.9190.889
IS-EMAall three views0.9700.9610.9470.9070.9240.896
channel only0.9700.9600.9460.9070.9220.894
global only0.9680.9550.9420.9010.9200.892
local only0.9700.9600.9470.9090.9230.895
Table 9. Ablation study: the SI-SDR scores of Inter-SubNet, IS-IMA and IS-EMA with different assignments of view attention in the MA block at different SNRs. Bold numbers represent the best values among the compared methods.
Table 9. Ablation study: the SI-SDR scores of Inter-SubNet, IS-IMA and IS-EMA with different assignments of view attention in the MA block at different SNRs. Bold numbers represent the best values among the compared methods.
SI-SDR17.5 dB12.5 dB7.5 dB2.5 dB0 dB−5 dB
Inter-SubNet21.83419.75918.05614.97810.4137.617
IS-IMAall three views21.89219.69517.84414.63010.4127.643
channel only21.91619.70517.93814.76310.5867.957
global only21.85719.76218.03114.86010.3197.635
local only21.90319.79418.03014.81810.2687.631
IS-EMAall three views21.97219.86618.24615.10910.3897.913
channel only22.07819.96118.13515.07510.9798.206
global only21.78519.65017.58313.84410.7497.949
local only21.48119.47218.06115.23510.8248.241
Table 10. The PESQ and STOI performance scores (rounded to the nearest two decimal places) of the presented IS-IMA and IS-EMA variants, FullSubNet [26], WaveUNet [40,41], SEGAN [17], CMGAN [42], MP-SENet [43], MANNER [33], and SE-Conformer [44]. Bold numbers represent the best values among the compared methods.
Table 10. The PESQ and STOI performance scores (rounded to the nearest two decimal places) of the presented IS-IMA and IS-EMA variants, FullSubNet [26], WaveUNet [40,41], SEGAN [17], CMGAN [42], MP-SENet [43], MANNER [33], and SE-Conformer [44]. Bold numbers represent the best values among the compared methods.
IS-IMA
(All Three Views)
IS-EMA
(Local Only)
FullSubNet WaveUNet SEGAN
PESQ3.593.643.433.343.41
STOI0.940.950.960.940.92
CMGANMP-SENetMANNERSE-Conformer
PESQ3.413.503.213.13
STOI0.960.960.950.95
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Hung, J.-W.; Li, T.-J.; Su, B.-Y. Enhancing Subband Speech Processing: Integrating Multi-View Attention Module into Inter-SubNet for Superior Speech Enhancement. Electronics 2025, 14, 1640. https://doi.org/10.3390/electronics14081640

AMA Style

Hung J-W, Li T-J, Su B-Y. Enhancing Subband Speech Processing: Integrating Multi-View Attention Module into Inter-SubNet for Superior Speech Enhancement. Electronics. 2025; 14(8):1640. https://doi.org/10.3390/electronics14081640

Chicago/Turabian Style

Hung, Jeih-Weih, Tsung-Jung Li, and Bo-Yu Su. 2025. "Enhancing Subband Speech Processing: Integrating Multi-View Attention Module into Inter-SubNet for Superior Speech Enhancement" Electronics 14, no. 8: 1640. https://doi.org/10.3390/electronics14081640

APA Style

Hung, J.-W., Li, T.-J., & Su, B.-Y. (2025). Enhancing Subband Speech Processing: Integrating Multi-View Attention Module into Inter-SubNet for Superior Speech Enhancement. Electronics, 14(8), 1640. https://doi.org/10.3390/electronics14081640

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop