Next Article in Journal
Application of PN Code in Time Delay Measurement of Telephone Network
Next Article in Special Issue
Simultaneous Speech and Eating Behavior Recognition Using Data Augmentation and Two-Stage Fine-Tuning
Previous Article in Journal
A Framework of State Estimation on Laminar Grinding Based on the CT Image–Force Model
Previous Article in Special Issue
Automated Detection of Gastrointestinal Diseases Using Resnet50*-Based Explainable Deep Feature Engineering Model with Endoscopy Images
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Robust Multi-Subtype Identification of Breast Cancer Pathological Images Based on a Dual-Branch Frequency Domain Fusion Network

School of Mechanical and Electrical Engineering, China Jiliang University, Hangzhou 310018, China
*
Author to whom correspondence should be addressed.
Sensors 2025, 25(1), 240; https://doi.org/10.3390/s25010240
Submission received: 12 December 2024 / Revised: 27 December 2024 / Accepted: 1 January 2025 / Published: 3 January 2025
(This article belongs to the Special Issue AI-Based Automated Recognition and Detection in Healthcare)

Abstract

:
Breast cancer (BC) is one of the most lethal cancers worldwide, and its early diagnosis is critical for improving patient survival rates. However, the extraction of key information from complex medical images and the attainment of high-precision classification present a significant challenge. In the field of signal processing, texture-rich images typically exhibit periodic patterns and structures, which are manifested as significant energy concentrations at specific frequencies in the frequency domain. Given the above considerations, this study is designed to explore the application of frequency domain analysis in BC histopathological classification. This study proposes the dual-branch adaptive frequency domain fusion network (AFFNet), designed to enable each branch to specialize in distinct frequency domain features of pathological images. Additionally, two different frequency domain approaches, namely Multi-Spectral Channel Attention (MSCA) and Fourier Filtering Enhancement Operator (FFEO), are employed to enhance the texture features of pathological images and minimize information loss. Moreover, the contributions of the two branches at different stages are dynamically adjusted by a frequency-domain-adaptive fusion strategy to accommodate the complexity and multi-scale features of pathological images. The experimental results, based on two public BC histopathological image datasets, corroborate the idea that AFFNet outperforms 10 state-of-the-art image classification methods, underscoring its effectiveness and superiority in this domain.

1. Introduction

According to a report by the World Health Organization (WHO), breast cancer (BC) is one of the most common cancers among females and a leading cause of cancer-related deaths in females under 45 years old. In 2022, there were 19.96 million new cancer cases worldwide, and BC ranked second in terms of incidence, becoming the leading cause of cancer-related deaths in 112 countries globally [1]. Early detection and accurate diagnosis of tumors play a critical role in clinical treatment, which can significantly reduce the mortality of cancers, exhibiting profound implications for treatment and prognosis among cancer patients. Therefore, early identification and appropriate treatment contribute to improving the long-term survival of patients with BC [2,3].
There are various methods for the detection and diagnosis of BC. Among them, histopathological examination is considered the “gold standard” for the diagnosis of this cancer [4]. Pathological images can present structural information at the cellular and tissue levels of breast lesions, which is critical for the diagnosis, classification, and grading of BC [5,6,7,8]. However, the intrinsic complexity and diversity of breast histopathological images result in labor-intensive and time-consuming work for pathologists. Furthermore, variations in experience and the subjective application of pathological diagnostic standards often bring inconsistent and non-reproducible diagnostic outcomes. To improve diagnostic accuracy, reduce diagnostic time, and avoid misdiagnosis or missed diagnosis among BC patients, deep learning (DL)-based computer-aided detection and diagnosis systems have received increasing attention from researchers. DL overcomes the limitations of conventional classification methods that require manual feature extraction of lesions. It can autonomously learn the characteristics of BC lesion regions, enabling it to assist clinicians in diagnosing diseases accurately and rapidly, minimizing subjective judgments and reliance on experience, and reducing the risk of misdiagnosis and missed diagnosis.
In recent years, DL algorithms have provided unique solutions for clinical medical image recognition. Spanhol et al. [9] employed various patching strategies to automatically extract features from BC histopathological images using AlexNet [10] while preserving high image resolution. Jiang et al. [11] proposed a Transformer-based [12] fine-grained classification model (Breast TransFG Plus) to perform grading of invasive ductal carcinoma (IDC). This model eliminated the widespread distribution characteristics of tumor cells in hematoxylin–eosin (H&E)-stained pathological images. Su et al. [13] introduced a model-ensemble learning framework that integrated classical Vision Transformer (ViT) modules into a graph neural network (GNN) for cancer tissue image grading. In this framework, graph nodes [14] corresponded to nucleus feature extraction based on convolutional neural networks (CNNs), while the Transformer module captured global graph node features of individual whole-slide images (WSIs). The evaluation results based on two public datasets [15,16] demonstrated that the ensemble model outperformed single models. However, these methods often focus solely on feature extraction in the spatial domain, overlooking the rich information potentially embedded in the frequency domain. The information is crucial for identifying microscopic structures and periodic texture patterns in pathological images, revealing complex changes in tumor cells and interactions between tissues. To address the aforementioned issues, we propose the following content:
  • We propose a dual-branch framework, AFFNet, designed to simultaneously capture fine-grained local details and global structural information in images.
  • The characteristics and attention mechanisms of BC histopathological images are analyzed from the perspective of the frequency domain. MSCA is employed to effectively extract the local high-saliency features of pathological images. Meanwhile, FFEO is utilized to enhance high-frequency information while preserving the low-frequency features of images. These frequency domain features are then transformed into spatial domain feature maps, thus increasing the number of feature maps and enabling the network to focus on contour and texture information simultaneously.
  • A novel frequency-domain-adaptive feature fusion strategy is proposed to adaptively model multi-scale contextual information. This enables the model to efficiently integrate multi-source information at different resolutions, thus enhancing its representational capacity.
  • AFFNet outperforms other state-of-the-art models in the experiments based on two publicly accessible BC histopathological image datasets. Furthermore, the effectiveness and superiority of the proposed method are validated by a T-SNE-based visualization.
In the following, Section 2 presents related work; Section 3 describes the proposed AFFNet model; Section 4 elucidates the extensive experimental results in detail; Section 5 provides the conclusion of this study, highlighting the advantages, limitations, and future potential of this method.

2. Related Work

Frequency domain: The pivotal role of the frequency domain in the field of DL has been increasingly recognized. Through the frequency domain, DL has demonstrated great potential in fields like image processing, signal processing, and time–series analysis [17,18,19]. In recent years, many researchers have employed frequency domain methods to enhance the feature extraction performance of models [20,21,22]. DCCNet [23] strengthens token interactions by applying dynamic circular convolutions in a separable manner along the channel dimension in the frequency domain. Specformer [24] integrates spectral analysis with multi-head self-attention into the original ViT architecture, exhibiting improved performance. Shen et al. [25] adopted branch processing techniques to obtain high, mid, and low frequencies and demonstrated the effectiveness of FFDN in extracting frequency domain information, further fully exploiting multi-band information in the shallow feature maps of blurry images.
Multi-branch networks: In recent years, a multi-branch approach has been adopted in an increasing number of designs to improve network structures and enhance performance. SCNet [26] divides input features into two branches. Specifically, one branch utilizes self-calibrated convolutions to significantly expand the receptive field of each convolutional layer through internal information interaction, thereby enhancing feature representation; while the other branch preserves the integrity of the original features. Wang et al. [27] proposed a dual-path lesion-aware neural network to remove the challenges of high inter-class similarity and significant intra-class variability in colonoscopy images through a dual-network fusion strategy, thereby improving the discriminative ability of the network. Jiang et al. [28] developed an end-to-end dual-branch fusion network to tackle the challenges of recognizing small details and complex nodules in lung nodule segmentation, significantly improving the segmentation performance for hard-to-identify lung nodules. These studies validate that multi-branch structural designs can effectively improve the performance of DL models on specific tasks.

3. Proposed Method

This study proposes a novel medical image analysis framework to assist in the histopathological diagnosis of BC. The architecture of AFFNet is illustrated in Figure 1. This network is established by ingeniously integrating the strengths of ConvNeXt [29] and Swin Transformer [30] through various frequency domain methods. It can not only capture detailed local features of breast tissues but also extract critical global contextual information. The design principles and functional implementation of each component in the network will be elucidated in the following sections.

3.1. Stem

In the AFFNet architecture, the image is first passed through a shared stem layer, where downsampling is performed via convolutions to reduce the input image resolution to one-fourth of its original size. Subsequently, the downsampled image is distributed to two specifically designed feature encoding branches. AFFNet builds hierarchical feature maps across four stages, with each stage’s architecture illustrated in Figure 2. The figure also offers a detailed depiction of two key modules: the Fourier Filtering Enhancement Operator (FFEO) and Adaptive Frequency Domain Fusion (AFDF).

3.2. Frequency Domain Operator

3.2.1. Multi-Spectral Channel Attention

The challenges in classifying histopathological images of BC include the nuclear overlap (which affects the features of individual nuclei) and the potential presence of multiple types of nuclei (which results in the indistinct features of some nuclei), thereby significantly affecting classification performance. Therefore, effectively extracting and utilizing salient features becomes a key to improving classification accuracy. To more effectively capture and leverage key frequency domain components in the signals, we introduced MSCA. This module employs the discrete cosine transform (DCT) to amplify essential information and suppress irrelevant details, followed by channel compression and re-excitation operations to enhance classification accuracy.
For an image f with a size of H × W, its discrete cosine transform can be expressed as follows:
F ( u , v ) = h = 0 H 1 w = 0 W 1 f ( h , w ) φ ( u , v , h , w ) s . t . h { 0 , 1 , , H 1 } , w { 0 , 1 , , W 1 }
Here, h and w are spatial domain indices representing the rows and columns of the image pixels. f(h,w) denotes the pixel value at (h,w) in the spatial domain, while H and W are the image dimensions in the vertical and horizontal directions, respectively. φ ( · ) represents the basis function for the 2D DCT. u and v are frequency domain indices corresponding to the vertical and horizontal directions of the image, and F(u,v) represents the spectral information obtained after the 2D DCT.
First, the input X R C × H × W is divided along the channel dimension into n segments X i R C n × H × W (where C must be divisible by n; in this study, n = 16 ). After division, X is represented as X 0 , X 1 , X 2 , , X n 1 , and the corresponding frequency components are assigned to X i . Consequently, the 2D DCT spectrum of the entire input X is expressed as:
F r e q = C o n c a t ( D C T ( X 0 ) , D C T ( X 1 ) , , D C T ( X n 1 ) )
Next, Freq undergoes the excitation and rescale steps:
Y = M S C A ( X ) = R e s c a l e ( E x c i t a t i o n ( F r e q ) )
The result obtained from Equation (3) represents the output Y of the MSCA module. By incorporating frequency domain information, the original input undergoes a transformation, enhancing its texture features from low to high significance and thereby improving the original features at the texture level.

3.2.2. Fourier Filtering Enhancement Operator

In histopathological image analysis, high-frequency information encapsulates critical microstructural details, which play a pivotal role in improving the classification performance and robustness of the model. In conventional methods, ideal low-pass filters are used to suppress high-frequency noise while preserving the low-frequency features of images. However, the high-frequency details that are crucial for pathological diagnosis may be overlooked, which is detrimental to the accurate diagnosis and classification of pathological images. To overcome this problem, the FFEO is proposed in this study to enhance high-frequency information while retaining low-frequency features of images. By performing precise modulation in the frequency domain, FFEO identifies and retains high-frequency details valuable for pathological analysis while reducing irrelevant noise. The Algorithm 1 section provides a detailed description of the pseudo-algorithm process of FFEO.
Algorithm 1: Fourier Filtering Enhancement Operator
Input:  X in = X R C × H × W
Output:  X out = Y R C × H × W
1:
B , C , H , W shape of X in  
2:
z RFFT 2 ( X in . permute ) . reshape ( B , H , W / / 2 + 1 , N , C / N )  
  Filtering:  
3:
z ˜ u , v L P , z ˜ u , v H P Calculate of low - and high - frequency components by Equations ( 4 ) and ( 5 )  
4:
z ˜ u , v = [ z ˜ u , v L P , z ˜ u , v H P ]  
  Enhancement:  
5:
z ˜ u , v = α × softshrink ( z ˜ u , v ) + ( 1 α ) × ( z ˜ u , v softshrink ( z ˜ u , v ) )  
6:
X i n IRFFT 2 ( z ˜ ) . reshape ( B , C , H , W )  
7:
X out Y = X in + X  
8:
return  X out
Introduction: The Fourier transform simultaneously captures both the low-frequency and high-frequency components of a signal. To retain low-frequency information while enhancing high-frequency components, we apply the Fourier transform to convert the input X from the spatial domain to the frequency domain.
Fourier filter design: First, the input X R C × H × W is transformed into frequency domain features z using the Fourier transform, where each component is represented as z u , v , with u and v as frequency domain indices. Next, a two-layer perceptron (MLP) is applied to the frequency domain features derived from the Fourier Transform, allowing the network to capture complex representations across multiple levels. As shown in Equations (4), (5), (6), the MLP separately models high- and low-frequency features, facilitating comprehensive analysis and processing of the entire frequency spectrum.
z ˜ u , v L P = MLP z u , v L P = W 2 σ W 1 z u , v L P + b 1 + b 2
z ˜ u , v H P = MLP z u , v H P = W 2 σ W 1 z u , v H P + b 1 + b 2
z ˜ u , v = z ˜ u , v L P , z ˜ u , v H P
Inspired by a previous study [31], the weights are structured in a block-diagonal format, with each W comprising N distinct small weight blocks. Each small weight block is shared among all z u , v . The variables z u , v L P and z u , v H P represent the low-pass and high-pass components, where z u , v L P C l o w _ f r e q _ c o u n t , z u , v H P C h i g h _ f r e q _ c o u n t , l o w _ f r e q _ c o u n t = h i g h _ f r e q _ c o u n t = 1 2 ( H × ( W 2 + 1 ) ) .
Design enhancement: In conventional methods, soft-thresholding and shrinkage in the frequency domain are used to remove insignificant high-frequency noise. However, such methods may inadvertently filter out high-frequency details that are critical for pathological diagnosis, such as cell boundaries and subtle structures of nuclei. To eliminate these limitations, the following filtering enhancement method is designed in this study.
z ˜ u , v = α × softshrink ( z ˜ u , v ) + ( 1 α ) × ( z ˜ u , v softshrink ( z ˜ u , v ) )
where α is set to 0.5. Finally, the transformation is switched back from the frequency domain to the time domain and connected with the original input X via a residual connection, resulting in the final output of FFEO:
Y = F F E O ( X ) + X = I F F T ( z ˜ u , v ) + X

3.2.3. Structure

In the histopathological image classification of BC, to achieve accurate recognition of subtle features within images, a CNN–Transformer framework is adopted, incorporating MSCA and FFEO, as described in Equations (9) and (10), which enhances the ability to capture frequency domain features of pathological images.
L i = σ M S C A X i X i , ( i = 1 , 2 , 3 , 4 )
G i = σ F F E O X i X i , ( i = 1 , 2 , 3 , 4 )
In the ConvNeXt branch, an MSCA module is embedded into each block as described in Equation (9). X i represents the output features of the i-th stage ConvNeXt block, while L i denotes the enhanced features of X i after processing through MSCA. The DCT effectively concentrates an image’s energy into low-frequency components, making it ideal for extracting global structural features in pathology images, such as cell distribution and tissue morphology. Integrating the MSCA module into the ConvNeXt branch enables precise suppression of non-essential information while simultaneously enhancing focus on global salient features and extracting fine local details.
In the Swin Transformer branch, an FFEO module is embedded into each block as described in Equation (10). X i represents the output features of the i-th stage Swin Transformer block, while G i denotes the enhanced features of X i after processing through FFEO. This operator employs the Fourier Transform to convert spatial domain information into frequency domain representations and dynamically enhances features within specific frequency bands. Unlike the DCT, which focuses primarily on low-frequency components, Fourier Transform captures both low- and high-frequency information, offering flexibility to emphasize high-frequency features such as cell boundaries and intra-nuclear texture details. This enhancement mechanism, combined with the Swin Transformer, enables the model to maintain global perceptual awareness while enhancing sensitivity to local fine-grained features.

3.3. Adaptive Frequency Domain Fusion

The extensive contextual information required for different cancer types depends on the characteristics of the lesion type and tissue structures. For instance, distinguishing breast lumps may require relatively less contextual information, whereas identifying tubular adenomas in some pathological images may demand a broader range of contextual information. Therefore, inspired by a previous study [32], multi-scale feature maps are constructed to adaptively model long-term contexts, thus assisting the model in identifying and locating objects at the different resolution. This is achieved by explicitly decomposing a large-kernel convolution into a sequence of depthwise convolutions at larger scales to form a series of large-kernel convolutions across different scales. First, the input X i for the current stage is constructed by combining the current stage’s local features L i , global features G i , and the fused output f u s e i 1 from the previous stage, where i denotes the stage number in the network architecture, f u s e 0 = 0 .
X i = L i + G i + f u s e i 1 , ( i = 1 , 2 , 3 , 4 )
To enable X i to capture multi-scale spatial information with varying kernel sizes, depthwise separable convolutions of different kernel sizes are applied. The formulas for U 1 and U 2 at each layer are given as follows:
U 1 = DP 5 × 5 ( X i ) , U 2 = DP 7 × 7 ( U 1 )
where D P ( · ) represents the depthwise separable convolution operation. The adaptive frequency domain fusion module is divided into three branches, with the middle branch and the upper and lower branches derived from Equations (13) and (14):
U = M S C A ( c a t ( U 1 , U 2 ) )
Y u p b , 1 , h , w = i = 1 c U 1 b , c , h , w , Y d o w n b , 1 , h , w = i = 1 c U 2 b , c , h , w
Finally, the middle feature branch is fused with the upper and lower weight branches using Equation (15):
Y i = U × Y u p b , 1 , h , w + U × Y d o w n b , 1 , h , w , ( i = 1 , 2 , 3 , 4 )

4. Experiments and Results

4.1. Dataset

The model is trained and evaluated based on the data from two different BC histopathological datasets.
BreaKHis [33]: The experiment is conducted based on the BreaKHis dataset, which is divided into four magnifications (40×, 100×, 200×, and 400×) and includes 7909 images of eight different BC subtypes. The categories comprise benign lesions, including adenosis (A), fibroadenoma (F), phyllodes tumor (PT), and tubular adenoma (TA), as well as malignant lesions, such as ductal carcinoma (DC), lobular carcinoma (LC), mucinous carcinoma (MC), and papillary carcinoma (PC). The original image size is 700 × 460 pixels, which is resized to 224 × 224 pixels for the experiments. The dataset is divided into the training, validation, and test sets at a ratio of 0.6:0.2:0.2.
BACH [34]: This is a public dataset from the Grand Challenge, comprising 400 H&E-stained histopathological BC images with a size of 2048 × 1536 pixels. It can be categorized into four classes: normal, benign, in situ, and invasive. Due to the large image size and limited dataset, the following efforts are made to prevent overfitting during training. (1) Each image (2048 × 1536 pixels) is divided into smaller patches with a size of 512 × 512 pixels. (2) The patches with nuclei less than 20 are excluded as per a previous study [35]. (3) The images are resized to 224 × 224 pixels. The processed dataset contains a total of 4707 images, with 2825 for training, 941 for validation, and 941 for testing.

4.2. Experimental Setup

To evaluate the performance of the model, Accuracy, Precision, Recall, and F1-score are selected as classification metrics. The definition of these evaluation metrics can be expressed as follows:
A c c u r a c y = T P + T N T P + T N + F P + F N
P r e c i s i o n = T P T P + F P
R e c a l l = T P T P + F N
F 1 = 2 × P r e c i s i o n × R e c a l l P r e c i s i o n + R e c a l l = 2 T P 2 T P + F P + F N
The loss is calculated using the categorical cross-entropy loss function:
C r o s s E n t r o p y L o s s = 1 N i = 1 N c = 1 M y i c log y ^ i c
where N represents the total number of samples; M represents the number of classes. y i c represents the target label, which takes the value 1 if the true class of sample i equals c, and 0 otherwise. y ^ i c represents the probability that the model predicts that sample i belongs to class c.
All experiments are conducted on a single NVIDIA GeForce RTX 3090 GPU with 24 GB of memory. The Adam optimizer is used to train the model in the training process. The initial learning rate is set to 1 × 10 4 , the batch size to 16, and the number of epochs to 100. The learning rate is adjusted using a cosine annealing schedule. A weight decay parameter of 0.01 is employed to prevent overfitting of the model. The performance of the model is optimized by tuning these parameters, thus ensuring greater robustness in the training process.

4.3. Results

To validate the effectiveness of the proposed model, AFFNet is compared with 10 state-of-the-art network models developed in recent years, including ConvNeXt, Swin Transformer, ViT, DFFormer [36], TransNeXt [37], CSwin [38], PoolFormer [39], BiFormer [40], StoHisNet [41], and CAS-ViT [42].

4.3.1. Results Based on the BreaKHis Dataset

Table 1, Table 2, Table 3 and Table 4 present the experimental results based on the BreaKHis dataset under a magnification of 40×, 100×, 200×, and 400× across different networks. The experimental results demonstrate that although the dataset contains various typical and atypical BC subtypes, the proposed algorithm exhibits enhanced BC classification performance, and it outperforms existing methods across all four magnification levels.
Table 1 lists the performance of different comparison methods under a magnification of 40× based on the BreaKHis dataset. It can be observed that the proposed method outperforms other competitors in terms of accuracy, precision, recall, and F1-score. In the experiment based on the test dataset, AFFNet performs well, with the accuracy, precision, recall, and F1-score being 93.00%, 92.75%, 91.69%, and 92.06%, respectively. Compared with ConvNeXt and Swin Transformer, the accuracy of AFFNet increases by 16% and 7.5%, respectively. Among other methods, ViT has the lowest accuracy, which may be attributed to the insufficient number of samples in the dataset. The proposed method also shows advantages compared with stronger baselines such as TransNeXt, BiFormer, and CSwin. The results indicate that the features of images can be captured more effectively by integrating the proposed modules, contributing to improved performance.
Table 1 also lists the total number of parameters and floating-point operations (FLOPs) for this method compared with the aforementioned models. In this study, the depth of the original baseline model is modified, a hierarchical structure of [2, 2, 2, 2] is adopted to fuse the two baseline models, and frequency domain operators are also introduced, thus improving model performance while reducing computational complexity. The FLOPs of AFFNet are only 9.24 G, with 68.45 M parameters, representing a reduction of 6.11 G and 5.92 G in FLOPs and 19.12 M and 18.3 M in parameters compared with ConNeXt and Transformer, respectively. The model structure and computational efficiency of AFFNet are optimized while maintaining high accuracy, making it a more efficient and practical option for real-world applications.
Table 2, Table 3 and Table 4 present the classification results based on the BreaKHis dataset under a magnification of 100×, 200×, and 400×, respectively. It can be found that the proposed model consistently achieves optimal performance across all three magnifications. Under a magnification of 100×, ViT has the lowest accuracy. The reason is that ViT cannot construct hierarchical feature maps in the processing of images, and it directly divides the image into patches and processes them through a Transformer encoder. This approach may alter the original spatial locations of cancer regions, preventing accurate identification of cancer types. The relatively low accuracy of ConvNeXt may be attributed to the significant differences in features and structures between pathological and natural images. Pathological images often contain finer textures and patterns, while ConvNeXt, originally designed for natural images, may fail to capture these subtle features adequately. This limitation results in its suboptimal performance in medical image processing tasks. Additionally, under a magnification of 100×, the accuracy of all models except ConvNeXt and ViT reaches above 80%, with StoHisNet up to 88.70%. This model also employs a dual-branch hybrid architecture, enabling effective extraction of image features for the accurate classification of most cancer types. Similarly, under a magnification of 400×, TransNeXt outperforms other models with the accuracy reaching 89.01%. Its multi-scale reasoning capability effectively avoids deep degradation, thus maintaining efficient classification performance across images of varying magnifications.
As magnification increases, metrics such as accuracy and sensitivity exhibit a declining trend, which is closely associated with the substantial reduction in the field of view. With a reduced field of view, the type and richness of information differ significantly between low-magnification and high-magnification images. At low magnifications, the image reveals a more complete organizational structure and contextual information, with these global features aiding the model in capturing macro-level pathological features, thereby improving classification performance. However, at high magnifications, while the image contains more detailed features, the absence of global contextual support makes it challenging for local features to accurately represent the overall pathological condition, thereby weakening the model’s discriminative ability. Additionally, high-magnification images often exhibit higher noise levels, including uneven staining, artifacts, and increased texture complexity. These noise factors can interfere with the model’s feature extraction process at high magnifications, impairing its ability to recognize key pathological features. Furthermore, the greater variability in cell morphology and distribution in high-magnification images increases data heterogeneity, further complicating the classification task. This phenomenon does not indicate model failure, but rather reflects the natural variation in the characteristics of pathological images at different magnifications. It is important to emphasize that, despite the performance decline at higher magnifications, our model still outperforms all comparison models across all metrics. This suggests that our model has significant advantages in feature extraction and learning capabilities, enabling effective adaptation to and analysis of images across various magnification levels.

4.3.2. Results on Based the BACH Dataset

To further explore the generalization capability of AFFNet, some experiments are also performed based on the BACH dataset (see Table 5).
In the experiment based on the BreaKHis dataset, Swin Transformer demonstrates superior performance compared with ConvNeXt, with the average accuracy increased by 5.33%. However, in the experiment based on the BACH dataset, ConvNeXt significantly outperforms Swin Transformer. This indicates that different models may perform variably based on different datasets. However, AFFNet, by integrating the strengths of both ConvNeXt and Swin Transformer architectures, demonstrates optimal performance on all datasets, highlighting its consistency and robustness compared to other models. The model fusion effectively mitigates the differences in feature distributions across datasets, compensating for the limitations of individual models in certain scenarios. This fusion strategy enhances the model’s adaptability to diverse features, enabling it to exhibit stronger performance across various datasets, thus offering a reliable solution for breast lesion classification.

4.3.3. Result Analysis

The ROC curves and confusion matrices for the baseline models and AFFNet are shown in Figure 3. The errors are analyzed based on the classification results. In the BreaKHis dataset, the classification results for DC and LC are unfavorable, while the classification of other classes is more accurate. The main reasons are elucidated as follows. Firstly, the images in the dataset are scaled, resulting in changes in their resolution. Secondly, some pathological images have unclear background information, which interferes with the judgment of the model. Thirdly, DC and LC are both invasive subtypes of BC with similar cell arrangement patterns, making it difficult for the model to make differentiation during feature extraction. In the BACH dataset, the classification results for benign and in situ are unfavorable. The main reason is that processing high-resolution images by dividing them into smaller patches may result in the loss of critical pathological features and contextual information. However, overall, the proposed model significantly outperforms the two baseline models.
The classification ability of AFFNet is significantly superior to the two baseline networks before hierarchical fusion, which can also be confirmed by the T-SNE visualization in Figure 4. Figure 4a,b illustrate the visualizations of ConvNeXt and Swin Transformer, respectively. As shown in the figures, the inter-class distances are small, the boundaries within the classes are blurry, and the clustering effect is poor. Figure 4c shows the network proposed in this study. Obviously, this network has distinctly separated the different classes. This further confirms that this network achieves better classification performance compared with the two baseline networks in Figure 4a,b.

4.4. Ablation Study

To address the task of multi-subtype classification of breast pathology images, we developed a dual-branch network leveraging frequency domain fusion. This section systematically evaluates the effectiveness of each key module and strategy by adding or modifying specific components of the model. Ablation studies were conducted to address two primary questions: (1) the contribution of each component to the network architecture, and (2) the impact of the fusion strategy.
To evaluate the effectiveness of each component in our model, we utilized 40× magnification data from the BreakHis dataset for ablation experiments. These experiments enabled us to quantify and analyze the specific contribution of each component to the overall model performance. The results are presented in Table 6.
In this study, ConvNeXt and Swin Transformer were chosen as baseline models, denoted as Baseline 1 and Baseline 2, respectively. On the breast pathology image multi-subtype classification task, Baseline 1 achieved an accuracy of 77.00%, and Baseline 2 achieved 85.75%. To investigate the potential of model fusion, we performed the Baseline 1 + Baseline 2 experiment (third row) by employing a simple addition method to fuse the two baseline models. This approach enabled a direct comparison of performance differences between individual models and the fused model. Experimental results demonstrated that the fused network achieved an accuracy of 88.50%, improving performance by 11.50% and 2.75% over Baseline 1 and Baseline 2, respectively. In the Baseline 1 + MSCA experiment (fourth row), the introduction of MSCA enhanced network performance by 12.75%. In the Baseline 2 + FFEO experiment (fifth row), incorporating the FFEO improved the model performance by 5.25%. Finally, an adaptive frequency domain fusion strategy was applied to fuse the enhanced Baseline 1 and Baseline 2 at multiple scales (sixth row). This multi-scale frequency domain fusion method significantly enhanced model performance. Compared to the baseline models, our AFFNet achieved the best results on the breast pathology image multi-subtype classification task, demonstrating substantial performance improvements.
This series of ablation experiments validated the effectiveness of each network component and emphasized the critical role of the fusion strategy in enhancing model performance. Specifically, MSCA and FFEO significantly improved the model’s ability to capture critical features, while the multi-scale frequency domain fusion strategy further enhanced classification accuracy. These findings provide robust experimental evidence supporting the effectiveness of our model design.

5. Discussion

In this study, a dual-branch adaptive frequency domain fusion network (AFFNet) is developed for the histopathological image classification of breast cancer. This study emphasizes the importance of the frequency domain in the classification of histopathological images of breast cancer. In addition, Multi-Spectral Channel Attention is employed in the local feature branch to enhance the representation of local salient features in pathological images. The Fourier Filtering Enhancement Operator is proposed to assist the global feature branch in enhancing high-frequency information while preserving low-frequency features and reducing the interference of irrelevant noise. On that basis, the feature maps of different scales are constructed to adaptively model contextual information for both branches. Moreover, AFFNet is comprehensively evaluated based on the BreaKHis and BACH datasets. The results demonstrate that it outperforms existing state-of-the-art methods. Considering the specificity of pathological images, the performance of this network may be future improved by image enhancement techniques. Furthermore, we plan to explore the potential applications of AFFNet in other medical imaging domains, thus promoting its broader applicability in medical image processing. The findings of this study are expected to provide strong support and innovative insights for various downstream tasks in medical image analysis.

Author Contributions

Conceptualization, K.W. and J.L.; methodology, K.W.; software, K.W.; validation, K.W., J.L. and X.J.; formal analysis, K.W.; investigation, K.W.; resources, J.L.; data curation, X.J.; writing—original draft preparation, K.W.; writing—review and editing, K.W.; visualization, X.J.; supervision, J.L. All authors have read and agreed to the published version of the manuscript.

Funding

This research received no external funding.

Data Availability Statement

The original datasets in the study are available. BreaKHis: https://web.inf.ufpr.br/vri/databases/breast-cancer-histopathological-database-breakhis/ (accessed on 16 April 2023), BACH: https://iciar2018-challenge.grand-challenge.org/Dataset/ (accessed on 18 April 2023).

Conflicts of Interest

The authors declare no conflicts of interest.

References

  1. Bray, F.; Laversanne, M.; Sung, H.; Ferlay, J.; Siegel, R.L.; Soerjomataram, I.; Jemal, A. Global cancer statistics 2022: GLOBOCAN estimates of incidence and mortality worldwide for 36 cancers in 185 countries. CA Cancer J. Clin. 2024, 74, 229–263. [Google Scholar] [CrossRef]
  2. Din, N.M.U.; Dar, R.A.; Rasool, M.; Assad, A. Breast cancer detection using deep learning: Datasets, methods, and challenges ahead. Comput. Biol. Med. 2022, 149, 106073. [Google Scholar] [CrossRef] [PubMed]
  3. Yang, J.; Ju, J.; Guo, L.; Ji, B.; Shi, S.; Yang, Z.; Gao, S.; Yuan, X.; Tian, G.; Liang, Y.; et al. Prediction of HER2-positive breast cancer recurrence and metastasis risk from histopathological images and clinical information via multimodal deep learning. Comput. Struct. Biotechnol. J. 2022, 20, 333–342. [Google Scholar] [CrossRef] [PubMed]
  4. Pan, X.; Lu, Y.; Lan, R.; Liu, Z.; Qin, Z.; Wang, H.; Liu, Z. Mitosis detection techniques in H&E stained breast cancer pathological images: A comprehensive review. Comput. Electr. Eng. 2021, 91, 107038. [Google Scholar]
  5. Abhisheka, B.; Biswas, S.K.; Purkayastha, B. A Comprehensive Review on Breast Cancer Detection, Classification and Segmentation Using Deep Learning. Arch. Comput. Methods Eng. 2023, 30, 5023–5052. [Google Scholar] [CrossRef]
  6. Batool, A.; Byun, Y.-C. Toward Improving Breast Cancer Classification Using an Adaptive Voting Ensemble Learning Algorithm. IEEE Access 2024, 12, 12869–12882. [Google Scholar] [CrossRef]
  7. Gonzalez, R.; Nejat, P.; Saha, A.; Campbell, C.J.V.; Norgan, A.P.; Lokker, C. Performance of externally validated machine learning models based on histopathology images for the diagnosis, classification, prognosis, or treatment outcome prediction in female breast cancer: A systematic review. J. Pathol. Inform. 2024, 15, 100348. [Google Scholar] [CrossRef] [PubMed]
  8. Maurya, R.; Pandey, N.N.; Dutta, M.K.; Karnati, M. FCCS-Net: Breast cancer classification using Multi-Level fully Convolutional-Channel and spatial attention-based transfer learning approach. Biomed. Signal Process. Control 2024, 94, 106258. [Google Scholar] [CrossRef]
  9. Spanhol, F.A.; Oliveira, L.S.; Petitjean, C.; Heutte, L. Breast cancer histopathological image classification using convolutional neural networks. In Proceedings of the 2016 International Joint Conference on Neural Networks (IJCNN), Vancouver, BC, Canada, 24–29 July 2016; pp. 2560–2567. [Google Scholar]
  10. Krizhevsky, A.; Sutskever, I.; Hinton, G.E. ImageNet classification with deep convolutional neural networks. Commun. ACM 2017, 60, 84–90. [Google Scholar] [CrossRef]
  11. Jiang, Z.; Dong, Z.; Fan, J.; Yu, Y.; Xian, Y.; Wang, Z. Breast TransFG Plus: Transformer-based fine-grained classification model for breast cancer grading in Hematoxylin-Eosin stained pathological images. Biomed. Signal Process. Control 2023, 86, 105284. [Google Scholar] [CrossRef]
  12. Dosovitskiy, A.; Beyer, L.; Kolesnikov, A.; Weissenborn, D.; Zhai, X.; Unterthiner, T.; Dehghani, M.; Minderer, M.; Heigold, G.; Gelly, S.; et al. An image is worth 16x16 words: Transformers for image recognition at scale. arXiv 2021, arXiv:2010.11929. [Google Scholar]
  13. Su, Y.; Bai, Y.; Zhang, B.; Zhang, Z.; Wang, W. HAT-Net: A Hierarchical Transformer Graph Neural Network for Grading of Colorectal Cancer Histology Images. In Proceedings of the British Machine Vision Conference (BMVC2021), Virtual, UK, 22–25 November 2021; p. 412. [Google Scholar]
  14. Ju, W.; Fang, Z.; Gu, Y.; Liu, Z.; Long, Q.; Qiao, Z.; Qin, Y.; Shen, J.; Sun, F.; Xiao, Z.; et al. A Comprehensive Survey on Deep Graph Representation Learning. Neural Netw. 2024, 173, 106207. [Google Scholar] [CrossRef]
  15. Awan, R.; Sirinukunwattana, K.; Epstein, D.; Jefferyes, S.; Qidwai, U.; Aftab, Z.; Mujeeb, I.; Snead, D.; Rajpoot, N. Glandular Morphometrics for Objective Grading of Colorectal Adenocarcinoma Histology Images. Sci. Rep. 2017, 7, 16852. [Google Scholar] [CrossRef]
  16. Li, Z.; Zhang, H.; Mu, D.; Guo, L. Random Time Delay Effect on Out-of-Sequence Measurements. IEEE Access 2016, 4, 7509–7518. [Google Scholar] [CrossRef]
  17. Al-Haddad, L.; Jaber, A.; Neranon, P.; Al-Haddad, S. Investigation of Frequency-Domain-Based Vibration Signal Analysis for UAV Unbalance Fault Classification. Eng. Technol. J. 2013, 41, 915–923. [Google Scholar] [CrossRef]
  18. Kong, L.; Dong, J.; Ge, J.; Li, M.; Pan, J. Efficient frequency domain-based transformers for high-quality image deblurring. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Vancouver, BC, Canada, 17–24 June 2023; pp. 5886–5895. [Google Scholar]
  19. Yi, K.; Zhang, Q.; Fan, W.; Wang, S.; Wang, P.; He, H.; An, N.; Lian, D.; Cao, L.; Niu, Z. Frequency-domain MLPs are more effective learners in time series forecasting. In Proceedings of the 37th Conference on Neural Information Processing Systems (NeurIPS 2023), New Orleans, LA, USA, 10–16 December 2023. [Google Scholar]
  20. Attallah, O. Cervical Cancer Diagnosis Based on Multi-Domain Features Using Deep Learning Enhanced by Handcrafted Descriptors. Appl. Sci. 2023, 13, 1916. [Google Scholar] [CrossRef]
  21. Liu, X.; Gu, Z.; Ding, H.; Zhang, M.; Wang, L. Underwater Image Super-Resolution Using Frequency-Domain Enhanced Attention Network. IEEE Access 2024, 12, 6136–6147. [Google Scholar] [CrossRef]
  22. Sharma, R.; Mahanti, G.K.; Panda, G.; Rath, A.; Dash, S.; Mallik, S.; Hu, R. A Framework for Detecting Thyroid Cancer from Ultrasound and Histopathological Images Using Deep Learning, Meta-Heuristics, and MCDM Algorithms. J. Imaging 2023, 9, 173. [Google Scholar] [CrossRef] [PubMed]
  23. Vo, X.-T.; Nguyen, D.-L.; Priadana, A.; Jo, K.-H. Dynamic Circular Convolution for Image Classification. In Proceedings of the International Workshop on Frontiers of Computer Vision, Yeosu, Republic of Korea, 20–22 February 2023; pp. 43–55. [Google Scholar]
  24. Patro, B.N.; Namboodiri, V.P.; Agneeswaran, V.S. SpectFormer: Frequency and Attention is what you need in a Vision Transformer. arXiv 2023, arXiv:2304.06446. [Google Scholar]
  25. Shen, H.; Wang, C.; Deng, L.; He, L.; Lu, X.; Shao, M.; Meng, D. FDDN: Frequency-guided network for single image dehazing. Neural Comput. Appl. 2023, 35, 18309–18324. [Google Scholar] [CrossRef]
  26. Liu, J.-J.; Hou, Q.; Cheng, M.-M.; Wang, C.; Feng, J. Improving Convolutional Networks With Self-Calibrated Convolutions. In Proceedings of the 2020 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), Seattle, WA, USA, 13–19 June 2020; pp. 10096–10105. [Google Scholar]
  27. Wang, K.N.; Zhuang, S.; Ran, Q.Y.; Zhou, P.; Hua, J.; Zhou, G.Q.; He, X. DLGNet: A dual-branch lesion-aware network with the supervised Gaussian Mixture model for colon lesions classification in colonoscopy images. Med. Image Anal. 2023, 87, 102832. [Google Scholar] [CrossRef] [PubMed]
  28. Jiang, W.; Zhi, L.; Zhang, S.; Zhou, T. A Dual-Branch Framework With Prior Knowledge for Precise Segmentation of Lung Nodules in Challenging CT Scans. IEEE J. Biomed. Health Inf. 2024, 28, 1540–1551. [Google Scholar] [CrossRef] [PubMed]
  29. Liu, Z.; Mao, H.; Wu, C.-Y.; Feichtenhofer, C.; Darrell, T.; Xie, S. A convnet for the 2020s. In Proceedings of the 2022 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), New Orleans, LA, USA, 18–24 June 2022; pp. 11976–11986. [Google Scholar]
  30. Liu, Z.; Lin, Y.; Cao, Y.; Hu, H.; Wei, Y.; Zhang, Z.; Lin, S.; Guo, B. Swin transformer: Hierarchical vision transformer using shifted windows. In Proceedings of the 2021 IEEE/CVF International Conference on Computer Vision (ICCV), Montreal, QC, Canada, 10–17 October 2021; pp. 10012–10022. [Google Scholar]
  31. Guibas, J.; Mardani, M.; Li, Z.; Tao, A.; Anandkumar, A.; Catanzaro, B. Adaptive fourier neural operators: Efficient token mixers for transformers. arXiv 2021, arXiv:2111.13587. [Google Scholar]
  32. Lin, T.-Y.; Dollár, P.; Girshick, R.; He, K.; Hariharan, B.; Belongie, S. Feature pyramid networks for object detection. In Proceedings of the 2017 IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Honolulu, HI, USA, 21–26 July 2017; pp. 2117–2125. [Google Scholar]
  33. Spanhol, F.A.; Oliveira, L.S.; Petitjean, C.; Heutte, L. A Dataset for Breast Cancer Histopathological Image Classification. IEEE Trans. Biomed. Eng. 2016, 63, 1455–1462. [Google Scholar] [CrossRef] [PubMed]
  34. Aresta, G.; Araújo, T.; Kwok, S.; Chennamsetty, S.S.; Safwan, M.; Alex, V.; Marami, B.; Prastawa, M.; Chan, M.; Donovan, M. Bach: Grand challenge on breast cancer histology images. Med. Image Anal. 2019, 56, 122–139. [Google Scholar] [CrossRef]
  35. Golatkar, A.; Anand, D.; Sethi, A. Classification of breast cancer histology using deep learning. In Proceedings of the 15th International Conference on Image Analysis and Recognition ICIAR 2018, Póvoa de Varzim, Portugal, 27–29 June 2018; pp. 837–844. [Google Scholar]
  36. Tatsunami, Y.; Taki, M. Fft-based dynamic token mixer for vision. In Proceedings of the AAAI Conference on Artificial Intelligence, Vancouver, BC, Canada, 20–27 February 2024; pp. 15328–15336. [Google Scholar]
  37. Shi, D. TransNeXt: Robust Foveal Visual Perception for Vision Transformers. In Proceedings of the 2024 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), Seattle, WA, USA, 17–21 June 2024; pp. 17773–17783. [Google Scholar]
  38. Dong, X.; Bao, J.; Chen, D.; Zhang, W.; Yu, N.; Yuan, L.; Chen, D.; Guo, B. Cswin transformer: A general vision transformer backbone with cross-shaped windows. In Proceedings of the 2022 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), New Orleans, LA, USA, 18–24 June 2022; pp. 12124–12134. [Google Scholar]
  39. Yu, W.; Luo, M.; Zhou, P.; Si, C.; Zhou, Y.; Wang, X.; Feng, J.; Yan, S. MetaFormer Is Actually What You Need for Vision. In Proceedings of the 2022 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), New Orleans, LA, USA, 18–24 June 2022; pp. 10819–10829. [Google Scholar]
  40. Zhu, L.; Wang, X.; Ke, Z.; Zhang, W.; Lau, R. BiFormer: Vision Transformer with Bi-Level Routing Attention. In Proceedings of the 2023 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), Vancouver, BC, Canada, 17–24 June 2023. [Google Scholar]
  41. Fu, B.; Zhang, M.; He, J.; Cao, Y.; Guo, Y.; Wang, R. StoHisNet: A hybrid multi-classification model with CNN and Transformer for gastric pathology images. Comput. Methods Programs Biomed. 2022, 221, 106924. [Google Scholar] [CrossRef] [PubMed]
  42. Zhang, T.; Li, L.; Zhou, Y.; Liu, W.; Qian, C.; Ji, X. CAS-ViT: Convolutional Additive Self-attention Vision Transformers for Efficient Mobile Applications. arXiv 2024, arXiv:2408.03703. [Google Scholar]
Figure 1. The hierarchical network architecture of AFFNet.
Figure 1. The hierarchical network architecture of AFFNet.
Sensors 25 00240 g001
Figure 2. The basic architecture in each stage of AFFNet.
Figure 2. The basic architecture in each stage of AFFNet.
Sensors 25 00240 g002
Figure 3. ROC and confusion matrix.
Figure 3. ROC and confusion matrix.
Sensors 25 00240 g003
Figure 4. Visualization of T-SNE of different networks.
Figure 4. Visualization of T-SNE of different networks.
Sensors 25 00240 g004
Table 1. Experimental results under a magnification of 40×. (The best results are in bold.)
Table 1. Experimental results under a magnification of 40×. (The best results are in bold.)
ModelParams (M)Folps (G)Accuracy (%)Precision (%)Recall (%)F1-Score (%)
ConvNeXt-B87.5715.3577.0073.7869.7071.05
Swin Transformer-B86.7515.1685.5083.7683.0483.27
ViT85.8016.8651.7524.2326.7724.22
DFFormer-B36111.9422.1079.7576.9974.4975.51
TransNeXt-base88.9617.8285.0083.0883.0782.71
CSwin-B76.6214.3783.5081.5179.6480.19
PoolFormer-M3655.418.7681.7580.1480.7380.12
BiFormer-B56.049.3690.2589.6789.6289.44
StoHisNet31.137.8191.0091.5586.6188.70
CAS-ViT2.760.5582.2581.2474.8876.68
AFFNet (Ours)68.459.2493.0092.7591.6992.06
Table 2. Experimental results under a magnification of 100×. (The best results are in bold.)
Table 2. Experimental results under a magnification of 100×. (The best results are in bold.)
ModelAccuracy (%)Precision (%)Recall (%)F1-Score (%)
ConvNeXt-B79.5676.1674.0874.79
Swin Transformer-B85.8184.5982.1383.09
ViT56.9743.2133.2632.25
DFFormer-B3681.9778.2278.1477.93
TransNeXt-base87.9885.6386.0185.78
CSwin-B80.5278.2773.6775.65
PoolFormer-M3682.2179.0376.7577.56
BiFormer-B83.8980.1381.6680.40
StoHisNet88.7087.3285.6086.34
CAS-ViT83.6581.2177.8879.34
AFFNet (Ours)90.6290.8287.2988.74
Table 3. Experimental results under a magnification of 200×. (The best results are in bold.)
Table 3. Experimental results under a magnification of 200×. (The best results are in bold.)
ModelAccuracy (%)Precision (%)Recall (%)F1-Score (%)
ConvNeXt-B80.6477.7576.9677.24
Swin Transformer-B83.3779.3380.2779.41
ViT60.5451.0636.5635.88
DFFormer-B3683.3780.1580.1279.82
TransNeXt-base86.8484.9584.9284.83
CSwin-B83.3778.5176.4977.18
PoolFormer-M3682.3882.2077.0079.05
BiFormer-B87.8486.8588.3487.26
StoHisNet87.0986.2385.5985.71
CAS-ViT81.8880.0475.0776.86
AFFNet (Ours)90.3288.8989.4489.09
Table 4. Experimental results under a magnification of 400×. (The best results are in bold.)
Table 4. Experimental results under a magnification of 400×. (The best results are in bold.)
ModelAccuracy (%)Precision (%)Recall (%)F1-Score (%)
ConvNeXt-B84.8984.2081.7782.28
Swin Transformer-B88.7386.2984.9185.33
ViT60.7158.0839.8141.06
DFFormer-B3682.9678.9380.4979.49
TransNeXt-base89.0187.2987.0586.98
CSwin-B74.1771.5765.8468.00
PoolFormer-M3684.0681.6680.9380.67
BiFormer-B87.9184.8086.9485.56
StoHisNet85.7182.8483.3382.89
CAS-ViT83.5181.4377.5878.90
AFFNet (Ours)90.6590.3988.1688.76
Table 5. Experimental results based on the BACH dataset. (The best results are in bold.)
Table 5. Experimental results based on the BACH dataset. (The best results are in bold.)
ModelAccuracy (%)Precision (%)Recall (%)F1-Score (%)
ConvNeXt-B86.2986.4386.2986.31
Swin Transformer-B58.2357.9558.1657.95
ViT42.6145.3842.4340.56
DFFormer-B3685.5485.5585.5185.50
TransNeXt-base91.1791.1791.1591.16
CSwin-B48.8851.1748.8849.18
PoolFormer-M3683.8483.9583.8383.83
BiFormer-B84.4884.4584.4584.44
StoHisNet86.0786.2486.0786.11
CAS-ViT83.5283.5983.5183.49
AFFNet (Ours)93.6293.6693.5993.58
Table 6. The ablation experiment results under a magnification of 40× based on the BreaKHis dataset.
Table 6. The ablation experiment results under a magnification of 40× based on the BreaKHis dataset.
Baseline 1Baseline 2MSCAFFEOAFDFAccuracy (%)Precision (%)Recall
(%)
F1-Score (%)
××××77.0073.7869.7071.05
××××85.7583.3484.8583.88
×××88.5086.2888.3887.09
×××89.7588.4186.9687.31
×××91.0090.2289.3589.58
93.0092.7591.6992.06
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Li, J.; Wang, K.; Jiang, X. Robust Multi-Subtype Identification of Breast Cancer Pathological Images Based on a Dual-Branch Frequency Domain Fusion Network. Sensors 2025, 25, 240. https://doi.org/10.3390/s25010240

AMA Style

Li J, Wang K, Jiang X. Robust Multi-Subtype Identification of Breast Cancer Pathological Images Based on a Dual-Branch Frequency Domain Fusion Network. Sensors. 2025; 25(1):240. https://doi.org/10.3390/s25010240

Chicago/Turabian Style

Li, Jianjun, Kaiyue Wang, and Xiaozhe Jiang. 2025. "Robust Multi-Subtype Identification of Breast Cancer Pathological Images Based on a Dual-Branch Frequency Domain Fusion Network" Sensors 25, no. 1: 240. https://doi.org/10.3390/s25010240

APA Style

Li, J., Wang, K., & Jiang, X. (2025). Robust Multi-Subtype Identification of Breast Cancer Pathological Images Based on a Dual-Branch Frequency Domain Fusion Network. Sensors, 25(1), 240. https://doi.org/10.3390/s25010240

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop