Next Article in Journal
Long-Term Stability of Nanobubbles Generated via Pressure Oscillation-Hydrodynamic Cavitation: A Rapid Assessment by UV–Vis Spectrophotometry
Previous Article in Journal
Development of a Machine Learning Interatomic Potential for Zirconium and Its Verification in Molecular Dynamics
Previous Article in Special Issue
The Superconducting Mechanism in BiS2-Based Superconductors: A Comprehensive Review with Focus on Point-Contact Spectroscopy
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

A Wavelet-Based Bilateral Segmentation Study for Nanowires

School of Computer Science and Technology, Changchun Normal University, Changchun 130032, China
*
Authors to whom correspondence should be addressed.
Nanomaterials 2025, 15(21), 1612; https://doi.org/10.3390/nano15211612
Submission received: 30 September 2025 / Revised: 18 October 2025 / Accepted: 20 October 2025 / Published: 23 October 2025

Abstract

One-dimensional (1D) nanowires represent a critical class of nanomaterials with extensive applications in biosensing, biomedicine, bioelectronics, and energy harvesting. In materials science, accurately extracting their morphological and structural features is essential for effective image segmentation. However, 1D nanowires frequently appear in dispersed or entangled configurations, often with blurred backgrounds and indistinct boundaries, which significantly complicates the segmentation process. Traditional threshold-based methods struggle to segment these structurally complex nanowires with high precision. To address this challenge, we propose a wavelet-based Bilateral Segmentation Network named WaveBiSeNet, to which a Dual Wavelet Convolution Module (DWCM) and a Flexible Upsampling Module (FUM) are introduced to enhance feature representation and improve segmentation accuracy. In this study, we benchmarked WaveBiSeNet against ten segmentation models on a peptide nanowire image dataset. Experimental results demonstrate that WaveBiSeNet achieves, mIoU of 77.59%, an accuracy of 89.95%, an F1 score of 87.22%, and a Kappa coefficient of 74.13%, respectively. Compared to other advanced models, our proposed model achieves better segmentation performance. These findings demonstrate that WaveBiSeNet is an end-to-end deep segmentation network capable of accurately analyzing complex 1D nanowire structures.

1. Introduction

One-dimensional (1D) nanowires [1] are linear nanostructures with cross-sectional dimensions typically ranging from 1 to 100 nanometers, in which the length is much greater than the diameter. Their lengths can extend from several hundred nanometers to several millimeters or even longer, making them a representative class of one-dimensional materials. Depending on their composition and application, 1D nanowires can be classified into several categories. Semiconductor nanowires (e.g., silicon, gallium nitride) are widely used in electronic and optoelectronic devices [2]; metal nanowires (e.g., copper, gold) exhibit outstanding thermal conductivity [3]; oxide nanowires (e.g., zinc oxide) demonstrate excellent performance in photocatalysis and gas sensing [4]; carbon-based nanowires (e.g., carbon nanotubes) combine high strength with superior electrical conductivity [5]; and peptide nanowires, due to their self-assembly capability and biocompatibility, are extensively employed in biomedical and biosensing applications [6,7]. In general, these nanowires possess remarkable electrical conductivity, mechanical flexibility, and tunable physicochemical properties, rendering them highly promising for applications in sensing technologies, energy devices, and biomedical engineering [8].
In materials science, the accurate measurement of morphological parameters of one-dimensional (1D) nanomaterials, such as length, diameter, curvature, and spatial distribution, is essential for understanding material performance, analyzing electrocatalytic activity, and optimizing the design of functional devices. Transmission electron microscopy (TEM) [9] is widely employed to characterize nanowire structures; however, it is inherently limited to providing localized structural information [10]. TEM images of one-dimensional nanowires often exhibit blurred backgrounds, overlapping structures, and adhesion effects, which complicate accurate boundary delineation. Traditional image processing methods, including edge detection, contrast enhancement, and thresholding, are highly sensitive to image quality and frequently rely on manual intervention, rendering them unsuitable for complex structural environments. While semi-automatic annotation tools such as LabelMe [11] offer some assistance, they suffer from limited edge-detection precision and low efficiency for large-scale image datasets. Compared with higher-dimensional objects, segmenting one-dimensional nanowires from the background remains a significant challenge in computer vision [12].
In recent years, deep learning has been widely applied in the field of nanomaterials due to its end-to-end training mechanism and excellent feature extraction capabilities [13,14]. Compared with conventional methods, deep learning offers greater accuracy and robustness in nanowire segmentation. It enables the automatic learning of optimal features, reduces human involvement, and improves generalization across complex image distributions.
This study proposes WaveBiSeNet, a wavelet-based bilateral segmentation network based on the BiSeNetV1 [15] architecture, for the high-precision segmentation of one-dimensional nanowires. The model incorporates a Dual Wavelet Convolution Module (DWCM) to enhance feature extraction and a Flexible Upsampling Module (FUM) to improve edge detail reconstruction. Experiments were conducted on a dataset of 3035 peptide nanowire TEM images, which include both dispersed and networked structures with complex backgrounds and blurred edges. Comparative evaluation against ten semantic segmentation models shows that WaveBiSeNet achieves 77.59% mIoU, 89.95% accuracy, 87.22% F1 score, and 74.13% Kappa, demonstrating accurate segmentation of 1D nanowires under challenging imaging conditions. We hypothesize that, under the same evaluation protocol, the integration of DWCM and FUM enables WaveBiSeNet to improve segmentation in blurred backgrounds and adherent regions compared with BiSeNetV1.
The contributions of this paper are summarized as follows:
(1)
We propose WaveBiSeNet, a wavelet-based bilateral segmentation network, that improves upon BiSeNetV1 [15] for the accurate segmentation of one-dimensional nanowires with complex backgrounds and blurred edges.
(2)
We introduce the Dual Wavelet Convolution Module (DWCM), which enhances feature extraction, and the Flexible Upsampling Module (FUM), which refines fine edge details.
(3)
Experiments on the peptide nanowire dataset demonstrate that WaveBiSeNet outperforms ten existing semantic segmentation models.

2. Related Work

In the early stages, machine learning was applied to predict the properties of nanomaterials and assist in their design, promoting the development of computational approaches in nanomaterial research [16,17]. Several studies have demonstrated the utility of machine learning in this domain. Ghada Dahy et al. [18] proposed an intelligent optimization model capable of classifying different types of nanoparticles. Ajay Vikram Singh et al. [19] applied machine learning techniques to investigate interactions between cells and nanoparticles (NPs). Byoungsang Lee et al. [20] conducted high-precision statistical analysis of large-scale nanoparticle morphologies using machine learning methods. Arkaprava Banerjee et al. [21] effectively predicted the cytotoxicity of TiO2-based multi-component nanomaterials through hyperparameter optimization and structural similarity analysis. Parashuram Bannigidad et al. [22] utilized probabilistic neural networks (PNN) and k-nearest neighbor (K-NN) classifiers to distinguish between boron, iron, and silver nanoparticles. Although machine learning has accelerated research and development in nanomaterials, challenges remain, particularly in terms of data quality and algorithm interpretability [23].
With the advancement of deep learning, its applications in nanomaterials research have become more sophisticated. Recent studies increasingly focused on the image analysis and feature extraction of one-dimensional (1D) nanowire materials. Jacob R. Taylor et al. [24] employed a Vision Transformer [25] to predict Majorana zero modes (MZMs) in 1D superconducting-semiconductor nanowires. Junghyun Lee et al. [26] applied an artificial neural network (ANN) [27] to predict the electromagnetic response of double-layer nanowire gratings. Huitian Bai et al. [28] proposed a bidirectional long short-term memory (LSTM) model combined with a conditional random field layer for the precise localization and orientation detection of nanowires. Linmao Li et al. [29] integrated Mask R-CNN [30] and YOLO [31] to identify and segment nanowire structures in SEM images. Jack D. Kendall et al. [32] developed a neuromorphic network architecture (MN3) based on resistive nanowires, enabling efficient deep neural network training. Xin Hu et al. [33] used convolutional neural networks to perform the super-resolution reconstruction and deblurring of silver nanowires and other nanostructures in blurred optical images.
Despite these advancements, semantic segmentation of one-dimensional (1D) nanowires remains limited. This study proposes a segmentation framework tailored to the structural characteristics of 1D nanowire materials. By integrating deep learning-based semantic segmentation techniques, the framework improves segmentation accuracy and recognition performance for these structures.

3. Materials and Methods

This section presents the baseline BiSeNetV1 [15] and the improved WaveBiSeNet architecture. WaveBiSeNet incorporates two specialized modules: the Dual Wavelet Convolution Module (DWCM), which enhances feature extraction, and the Flexible Upsampling Module (FUM), which refines edge detail reconstruction. Together, these modules enable the accurate segmentation of one-dimensional nanowires, particularly in images with complex backgrounds and blurred edges.

3.1. BiSeNetV1 Model Architecture

The BiSeNetV1 [15] dual-branch real-time semantic segmentation network is designed to strike a balance between segmentation accuracy and computational efficiency. It comprises two parallel paths: a spatial path for preserving fine-grained details and a context path for capturing high-level semantic information. The overall network architecture is illustrated in Figure 1. In the spatial path, three convolutional layers with a kernel size of 3 × 3 and a stride of 2 are employed to progressively downsample the input image, ultimately reducing its resolution to 1/8 of the original size while retaining spatial structure information. In parallel, the contextual path utilizes ResNet-18 [34] as the backbone to extract deep semantic features. This path includes 8× and 16× downsampling operations and integrates a global average pooling module to enhance the model’s global context awareness. To further refine the extracted features, an Attention Refinement Module (ARM) is incorporated, which enhances critical features and suppresses irrelevant information. The structure of the ARM is shown in Figure 2a. Specifically, the ARM first applies global average pooling (GAP) to the input feature map to obtain channel-wise global response statistics. These statistics are then passed through a 1 × 1 convolution layer, followed by normalization and a Sigmoid activation function, to generate channel attention weights. Finally, these weights are multiplied channel-wise with the original feature map, thereby emphasizing task-relevant information while suppressing background noise and irrelevant responses. Furthermore, the features processed by ARM at the 1/16 scale are combined with the global contextual information obtained from the GAP branch and then upsampled to the 1/8 scale. These enriched features are fused with shallow features from earlier layers, effectively injecting global semantic information into higher-resolution representations.
The outputs from both paths are fused by a Feature Fusion Module (FFM). The structure of the FFM is shown in Figure 2b. FFM in BiSeNetV1 integrates the spatial and context pathways at the 1/8 scale. The two feature maps are concatenated along channels and passed through a 3 × 3 convolution, batch normalization (BN), and ReLU to perform local interaction and unify the channel dimensionality to a fixed size. A lightweight channel-attention branch applies global average pooling followed by a 1 × 1 convolution, batch normalization, and a sigmoid to obtain channel-wise coefficients ω ϵ (0, 1). These coefficients recalibrate the main stream via channel-wise multiplication, and a residual addition with the pre-attention feature yields the final output. This design injects high-level semantics while preserving fine-grained spatial details, effectively combining detailed spatial features with rich contextual semantics. Finally, the fused feature map is upsampled to the original resolution using 8× bilinear interpolation, enabling accurate, pixel-level semantic predictions.

3.2. The Structure of the Proposed WaveBiSeNet Model

In BiSeNetV1 [15], stride-2 convolutions in the spatial path discard high frequency edge cues and, together with the shallow depth of this path, limit the receptive field; bilinear upsampling also struggles to recover sharp boundaries. Based on BiSeNetV1 [15], we propose WaveBiSeNet, shown in Figure 3. Using a 3 × 512 × 512 image as input, the three stride-2 convolutions in the spatial path are replaced with Dual Wavelet Convolution Modules (DWCM), that perform wavelet based downsampling to separate low and high frequency subbands, explicitly retaining horizontal, vertical, and diagonal edge responses characteristic of slender one-dimensional nanowires. A second decomposition on the low frequency branch expands the effective receptive field and strengthens contextual modeling. A lightweight convolution then compresses channels. The spatial path sizes evolve as 3 × 512 × 512 → C1 × 256 × 256 → C2 × 128 × 128 → C3 × 64 × 64, with C1 = 64, C2 = 128, and C3 = 256. The context path computes features at scales of 1/4, 1/8, 1/16, and 1/32, that is 128 × 128, 64 × 64, 32 × 32, and 16 × 16, and uses attention refinement modules at 1/16 and 1/32. Immediately after the 1/32 stage we place a Flexible Upsampling Module (FUM) that predicts learnable offsets to dynamically generate sampling points and reconstruct edge details in a lightweight and efficient manner, and it upsamples that branch to 1/8 for fusion. The fused context is then combined with the 1/8 spatial features through the Feature Fusion Module to obtain Cf × 64 × 64. A second Flexible Upsampling Module replaces bilinear interpolation and performs ×8 upsampling back to 512 × 512, producing logits of size 2 × 512 × 512 for two classes. Together, DWCM edge aware downsampling and FUM shape adaptive dynamic upsampling achieve the recovery of blurred details.

3.2.1. Dual Wavelet Convolution Module

Inspired by HWD [35] and WTConv [36], we propose the Dual Wavelet Convolution Module (DWCM), designed to replace the three strided convolution operations with 3   ×   3 kernels in the spatial path of BiSeNetV1 [15], as shown in Figure 4. We stack three DWCMs for two main reasons. First, they one-to-one replace the three stride-2 downsampling operations in the BiSeNetV1 [15] spatial path, preserving the original downsampling schedule so that the resolution halves at each stage and ultimately aligns with the one-eighth fusion scale. Second, they enable a progressive multi-scale analysis that converts each downsampling step into a wavelet-based decomposition: the first stage prioritizes the preservation of high-frequency edges, the second consolidates mid-scale structures, and the third performs a secondary decomposition of the low-frequency component to enlarge the effective receptive field and reduce aliasing. In computer vision, convolutional filters operating in the spatial domain and Fourier transform-based filters in the frequency domain are two widely adopted methods for feature extraction. The Wavelet Transform (WT) adopted in the proposed DWCM is derived from Fourier analysis and harmonic analysis [37]. It enables the extraction of features in both the temporal and frequency domains. This dual-domain representation enhances the model’s ability to capture multi-scale information.
The DWCM consists of two main components: the Wavelet Feature Downsampling Block and the Wavelet Feature Extraction Block. The Wavelet Feature Downsampling Block applies a Haar Wavelet transform (HWT) to decompose the input feature map into four frequency subbands: low frequency, horizontal high frequency, vertical high frequency, and diagonal high frequency, which both downsample the spatial resolution and map the local structure into frequency channels, capturing edge and orientation cues of 1D nanowires. The Wavelet Feature Extraction Block further decomposes the initially obtained subband feature maps. This operation extracts richer multi-scale frequency information. The Wavelet Feature Extraction Block further decomposes the initially decomposed subband feature maps using Discrete Wavelet Transform (DWT). This operation extracts richer multi-scale frequency information. Considering that the low-frequency subband contains richer structural and contextual information, a second wavelet decomposition is performed on it. This operation enhances feature representation and further enlarges the receptive field, facilitating more accurate segmentation of one-dimensional nanowires, especially in cases with complex backgrounds and blurred edges.
Specifically, four wavelet filters are defined with a stride of 2 and a kernel size of 1   ×   1 . The low-frequency filter ( H 0 ) corresponds to the low frequency convolution kernel f L L , and is used to extract the global information of the image. The high-frequency filters ( H 1 ) corresponding to high-frequency convolution kernels ( f L H , f H L , and f H H ), capture detailed information of the local edges. The formulas are presented in Equation (1).
  f L L = 1 1 1 1 ,   f L H = 1 1 1 1 f H L = 1 1 1 1 ,   f L H = 1 1 1 1
In the Wavelet Feature Downsampling Block, the nanowire feature map X i n with dimension C × H × W first processed by the HWT. Convolution operations are performed using the low-pass filter ( H 0 ) and the high-pass filter ( H 1 ) in the horizontal and vertical directions, respectively. The symbol denotes the 2× downsampling operation applied to both the approximation and detail components. This operation decomposes the input into four frequency subbands: the low-frequency subband X L L , the horizontal high-frequency subband X L H , the vertical high-frequency subband X H L , and the diagonal high-frequency subband X H H . These components correspond to the approximation subband (A) and the three detail subbands: vertical (V), horizontal (H), and diagonal (D), respectively. These subbands encode spatial frequency features into separate channels, allowing for resolution reduction while preserving key edge and orientation information. The processed feature map’s size is 4 C × H 2 × W 2 . The formulas are presented in Equation (2).
X L L , X L H , X H L , X H H = H W T ,   X i n
Subsequently, a 1 × 1 convolution, batch normalization (BN), and ReLU activation function are applied to fuse and refine the frequency information across different channels. This operation produces a fused feature map X with dimensions C × H 2 × W 2 ; this operation effectively integrates multi-frequency information. It enhances the model’s ability to perceive both local details and global structures. As a result, the segmentation accuracy is improved. The corresponding formula is presented in (3).
X = R e L U B N C o n v 1 × 1 C o n c a t X L L ,   X L H ,   X H L ,   X H H
In the Wavelet Feature Extraction Block, the downsampled feature map X is further processed using the WT. Through this operation, it is decomposed into four subbands: the low-frequency component X L L ( 1 ) , the horizontal high-frequency component X L H ( 1 ) , the vertical high-frequency component X H L ( 1 ) , and the diagonal high-frequency component X H H ( 1 ) , as shown in Equation (4).
X L L ( 1 ) , X L H ( 1 ) ,   X H L ( 1 )   , X H H ( 1 ) = C o n v 1 × 1 W T ,   X
Because the low-frequency component frequently contains richer and finer-grained frequency information, the low-frequency information X L L ( 1 ) is decomposed into low-frequency information by another WT, yielding four subbands X L L ( 2 ) , X L H ( 2 ) , X H L ( 2 ) , and X H H ( 2 ) . The formula is shown in Equation (5).
X L L ( 2 ) , X L H ( 2 ) ,   X H L ( 2 )   , X H H ( 2 ) = C o n v 1 × 1 W T , X L L ( 1 )
Multilevel feature maps at different frequency scales are finally obtained. The low-frequency component Y L L ( 1 ) , and the high-frequency component Y H ( 1 ) , are generated by fusing the first-level wavelet decomposition results using a 1 × 1 convolution. Similarly, the second-level decomposition results are fused to produce Y L L ( 2 ) and   Y H ( 2 ) . This fusion enhances the low-frequency feature representation, which improves the model’s ability to capture contextual information, as shown in Equation (6).
Y L L ( 1 ) , Y H ( 1 ) = C o n v 1 × 1 W ( 1 ) , X L L ( 2 ) ,   X L H ( 2 ) ,   X H L ( 2 ) ,   X H H ( 2 ) Y L L ( 2 ) , Y H ( 2 ) = C o n v 1 × 1 W ( 2 ) , X L L ( 1 ) ,   X L H ( 1 ) ,   X H L ( 1 ) ,   X H H ( 1 )
The Inverse Wavelet Transform (IWT) was applied to recombine information from different frequency bands, producing an intermediate feature map. Then, a 1 × 1 convolution was employed to adjust the number of output channels. This process yields the final feature map   X o u t with dimensions 4 C × H 2 × W 2 , as described in Equation (7).
X o u t = C o n v 1 × 1 I W T Y L L ( 1 ) + Y L L ( 2 ) ,   Y H ( 1 ) + Y H ( 2 )

3.2.2. Flexible Upsampling Module

Inspired by DySample [38], we propose a lightweight, flexible upsampling module, the Flexible Upsampling Module (FUM). In this module, a convolution layer is added after the DySample [38] grid sampling operation to better capture and refine the features of one-dimensional nanowires. Its architecture is illustrated in Figure 5. By introducing learnable offsets, sampling positions are dynamically generated. This design enables the efficient reconstruction of fine-grained features of one-dimensional nanowires and the recovery of blurred details, while maintaining low computational cost.
Specifically, the input feature map X i n R C × H × W is first used to construct a standard sampling grid X g R 2 × H × W , where each sampling position is represented by two coordinate values. Subsequently, X i n is linearly transformed to produce an offset feature map X R 2 g s 2 × H × W . Here, g denotes the number of channel groups, s represents the upsampling scale factor, and the factor 2 indicates the horizontal and vertical offsets for each coordinate. To constrain the range of the offsets, the offset map is normalized to the interval 0 , 1 using a sigmoid function. It is then multiplied by a learnable dynamic scaling factor 0.5, which is adapted based on the image features, to further adjust the sampling positions. The result is element-wise multiplied with the input feature map X i n , Then, the channel information is remapped to the spatial domain using a pixel_shuffle operation with an upsampling scale factor r, producing the sampling ensemble network X o R 2 g × H × W . The final sampling grid X S is obtained by summing X g and X o . Based on the coordinates of each pixel in X S , bilinear interpolation is performed on X i n using the grid_sample function. The sampled feature maps are sequentially passed through a 1 × 1 convolution layer, followed by Batch Normalization (BN) and a ReLU activation function. This process fuses channel information and introduces nonlinear transformations, which help to mitigate the internal covariate shift. As a result, the reconstruction of the 1D nanowire edges is improved. The output feature map X o u t R C × ( H · r ) × ( W · r ) is obtained. The calculation process is shown in Equation (8).
X i n   R C × H × W X o = pixel _ shuffle ( 0.5 σ w · X i n · w · X i n ) X s   = X g     + X o X o u t   = ReLU ( BN ( Conv 1 × 1 ( grid _ sample ( X i n , X s ) ) ) )

4. Results and Discussion

4.1. Dataset

This study utilized a dataset of Transmission Electron Microscopy (TEM) images of peptide nanowires, synthesized and imaged by Brian Montz in Professor Todd Emrick’s research group at the Department of Polymer Science and Engineering, University of Massachusetts Amherst [39], and made available through the open-access repository Zenodo [40]. The dataset comprised two categories: dispersed and network structures, with 100 images each. To accurately delineate nanowire regions, we first applied a 3 × 3 median filter to the original grayscale TEM images for denoising. The denoised images were then converted to the Hue–Saturation–Value (HSV) color space, and dual thresholds S < 40 and V < 90 were used to obtain initial binary masks. We performed one opening operation with a 3 × 3 structuring element to remove isolated noise, followed by two closing operations to repair small breaks; finally, we dilated the foreground by one pixel to better approximate the physical diameter. To address common issues in thin nanowires—segmented breaks and edge under-segmentation—we adopted SAM-assisted manual correction based on the Segment Anything Model [41]. Annotators used interactive prompts to generate candidate masks and performed pixel-level refinements, after which the results were reviewed and accepted as the final ground truth. In the label maps, nanowire regions are shown in white and the background in black. As shown in Figure 6, “Image” denotes the original TEM image and “Label” the corresponding binary label map. We first split the dataset at the original-image level into training, validation, and test sets in a 7:2:1 ratio, and then cropped each original image into fixed-size patches, yielding a total of 3035 samples: 2135 for training, 600 for validation, and 300 for testing.

4.2. Experimental Setup

The model was trained on a server equipped with an NVIDIA GeForce RTX 4090 (24 GB) GPU. The training environment included Python 3.9.19, PyTorch 1.8.1, and CUDA 12.4. Prior to training, the dataset was normalized to [0, 1], and He initialization was applied to optimize network weights and accelerate convergence. To improve the model’s sensitivity to 1D nanowire boundaries, a combined loss function was employed, consisting of 0.5 times the cross-entropy loss [42] and 0.5 times the weighted Dice loss [43]. Parameters were updated using the Adam optimizer [44], with an initial learning rate of 0.01 and momentum parameters β1 and β2 both set to 0.99. L2 regularization with a weight decay of 0.001 was applied to prevent overfitting. The learning rate was dynamically adjusted via the ReduceLROnPlateau scheduler. When the monitored metric plateaued, the learning rate was reduced by a factor of 0.5. The batch size was set to 16. The model was trained for 300 epochs.

4.3. Evaluation Metric

To quantitatively evaluate the segmentation performance of the WaveBiSeNet model on 1D nanowire materials, we selected four evaluation metrics in our experiments to comprehensively reflect the model’s performance from different aspects: mean intersection over union(mIoU) [45,46,47], accuracy [48,49], F1 score [50,51], and Kappa coefficient (Kappa) [52].
mIoU [45,46,47] is a widely used metric for evaluating image segmentation accuracy. It quantifies the pixel-level overlap between predicted segmentation and ground-truth annotations. The mIoU is computed as the average Intersection over Union (IoU) across all classes, and its formulation is provided in Equation (9).
m I o U = 1 n c i n i i i n i j + j n j i n i i
In the confusion matrix, n i j denotes the number of pixels that belong to class i but are incorrectly predicted as class j . Similarly, n j i represents the number of pixels that belong to class j but are misclassified as class i . The value n i i represents the number of pixels correctly predicted as class i . The variable n c refers to the total number of classes in the image.
Accuracy [48,49] is a fundamental metric for measuring the performance of a classification model, as it represents the proportion of correctly classified samples among the total samples. The formula is shown in Equation (10).
A c c u r a c y = T P + T N T P + T N + F P + F N
Here, TP refers to the number of foreground pixels correctly classified as foreground. TN denotes background pixels correctly classified as background. FP represents background pixels incorrectly classified as foreground, while FN indicates foreground pixels misclassified as background.
We present the confusion matrix for the improved model:
Prediction
ReferenceTP: 19,502,345FN: 1,636,842
FP: 5,798,124TN: 51,705,679
F1 score [50,51] is a composite metric that balances precision [53,54] and recall [53,54], providing a more comprehensive assessment of the model’s ability to detect foreground regions. The formula is shown in Equation (11).
F 1   s c o r e = 2 × P r e c i s i o n × R e c a l l P r e c i s i o n + R e c a l
The Kappa coefficient [52] is used to assess the classification consistency of the model in segmentation tasks. It quantifies the improvement of the segmentation results over those expected from random classification. In addition, it effectively corrects for the influence of class imbalance on performance evaluation. The coefficient reflects the actual classification performance by comparing the observed agreement p o with the expected agreement by chance p e . The calculation formula is provided in Equation (12).
K a p p a = p o p e 1 p e
where p o denotes the actual proportion of all pixels correctly classified by the model. It refers to the ratio of correctly predicted foreground and background pixels to the total number of pixels. In contrast, p e denotes the expected proportion of correctly classified pixels under random category assignment. It is estimated based on the prior distribution of foreground and background classes.

4.4. Ablation Experiments

To validate the effectiveness of the proposed modules, ablation experiments were conducted on the peptide nanowire dataset. The experimental results are presented in Table 1. The Dual Wavelet Convolution Module (DWCM) preserved edge details while reducing spatial resolution through efficient downsampling and receptive field expansion. This resulted in a model mIoU of 77.29%. The Flexible Upsampling Module (FUM) also demonstrated good performance in recovering fine-grained features, achieving an mIoU of 76.32%. When both modules were applied simultaneously, the mIoU increased to 77.59%. Compared with the baseline model BiSeNetV1 [15], the proposed modules led to improvements in mIoU, accuracy, F1 score, and Kappa. These results indicate the effectiveness of the proposed design in the 1D nanowire segmentation task.
We conducted visual ablation experiments to evaluate the effectiveness of the proposed modules. The results are presented in Figure 7. This figure illustrates the segmentation performance of the BiSeNetV1 [15] semantic segmentation model with different module combinations. The “Image” refers to the original input, and the “Label” denotes the corresponding ground truth. Key segmentation regions are highlighted with red dashed boxes for comparison. Compared to the baseline BiSeNetV1 [15], the DWCM improves edge detection and enhances the recognition of local structures. The FUM significantly improves detail recovery, resulting in smoother boundaries. When both modules are integrated, the model achieves the best performance in segmenting nanowire edges. Structural adhesion is effectively reduced, and boundary separation becomes clearer.

4.5. Model Comparison and Analysis

To comprehensively evaluate the segmentation performance of WaveBiSeNet, comparative experiments were conducted using the same allocation ratio of the peptide nanowire dataset. We used different random seeds to train each model five times and reported the accuracy results as the mean ± standard deviation. The results were compared with a series of state-of-the-art real-time segmentation algorithms, including BiSeNetV1 [15], FSSNet [55], EDA-Net [56], ESNet [57], CANet [58], AGLNet [59], BiSeNetV2 [60], Mobile-Unet [61], EGE-Unet [57], and LETNet [62]. As shown in Table 2, WaveBiSeNet achieved competitive performance across multiple evaluation metrics when ResNet18 [34] was used as the backbone network. Specifically, it achieved an mIoU of 77.59%, an accuracy of 89.95%, an F1 score of 87.22%, and a Kappa coefficient of 74.13%. Compared with the baseline model BiSeNetV1 [15], WaveBiSeNet improved these metrics by 2.12%, 1.26%, 1.56%, and 2.62%, respectively. EGE-Unet [57] attains 75.84% mIoU, which is slightly lower than WaveBiSeNet. FSSNet [55]; EDA-Net [56], ESNet [57], and CANet [58] achieved mIoUs between 68% and 72%, making them more suitable for resource-constrained scenarios. BiSeNetV2 [60] achieved an mIoU of 65.50%, offering a lightweight and efficient solution but with slightly reduced precision. AGLNet [59], LETNet [62], and Mobile-Unet [61] reported mIoUs between 60% and 66%, with most Kappa scores below 80%. These models exhibited limited capability in capturing fine details and delineating boundaries, which makes the segmentation of nanowire images particularly challenging, especially when the precise recognition of tiny structures is needed. Overall, WaveBiSeNet effectively preserves fine details while capturing broader contextual information, improving the segmentation of one-dimensional nanowires.
To further evaluate model performance, the Loss and mIoU curves over 300 training epochs are plotted in Figure 8. The Loss curve is shown in Figure 8a, and the mIoU curve is shown in Figure 8b.
As shown in Figure 8a, all models exhibit rapid loss reduction within the first 30 epochs, followed by convergence. ESNet [57] and CANet [58] start with the highest losses (>0.6) and show unstable training. BiSeNetV1 [15], BiSeNetV2 [60], and AGLNet [59] converge more quickly to 0.18, with moderate oscillations. Mobile-Unet [61] and EGE-Unet [57] remain less stable, ending between 0.20 and 0.25, while FSSNet [55] and EDA-Net [56] converge slowly around 0.22. LETNet [62] achieves a stable loss of 0.15 but converges slower than WaveBiSeNet. In contrast, WaveBiSeNet converges faster, stabilizing at 0.13 after epoch 210, which reflects faster convergence, lower final loss, and improved training stability.
From Figure 8b, all models show rising mIoU in the first 30 epochs before stabilization. BiSeNetV1 [15], BiSeNetV2 [60], EGE-Unet [57], and AGLNet [59] reach 0.73–0.75, but fluctuate in later stages. ESNet [56], CANet [57], and EDA-Net [56] remain limited (0.66–0.69), while Mobile-Unet [61] and LETNet [62] perform the worst (0.55–0.60), with Mobile-Unet [61] converging prematurely. FSSNet [55] stabilized at 0.69. Compared with other models, WaveBiSeNet achieves faster convergence (within 60 epochs) and a higher final mIoU of 0.77. These results suggest that the proposed improvements enhance its effectiveness for one-dimensional nanowire segmentation, particularly under challenging background conditions.
We conducted a statistical significance analysis of BiSeNetV1 and WaveBiSeNet using the bootstrap method with 10,000 resampling iterations, and the results are presented in Figure 9. Figure 9a illustrates the mIoU distributions of the two models. The histogram of BiSeNetV1, shown in red, peaks at approximately 75.5%, indicating that most of its mIoU values are concentrated at this level. The distribution of WaveBiSeNet, shown in green, peaks at around 77.5% and is clearly shifted to the right, reflecting better overall performance. The 95% confidence intervals, marked by dashed lines, are approximately 75.0–75.8% for BiSeNetV1 and 77.0–78.0% for WaveBiSeNet. Since these intervals do not overlap, the results provide an initial indication that WaveBiSeNet achieves significantly higher mIoU than BiSeNetV1. Figure 9b shows the distribution of the mIoU differences between the two models. The light blue histogram represents the differences obtained after each bootstrap resampling. The black solid line indicates the baseline of no difference (Difference = 0), while the orange dashed lines and shaded area represent the 95% confidence interval of the difference, approximately ranging from 1.5 to 2.5. This interval lies entirely to the right of zero and does not intersect the baseline, demonstrating that the observed improvement is not due to random variation but is statistically significant at the 95% confidence level.
In summary, the bootstrap analysis confirms that WaveBiSeNet achieves a significantly higher mIoU than BiSeNetV1. This performance improvement is statistically significant (p < 0.05), supported by the non-overlapping confidence intervals and the strictly positive difference distribution.

4.6. Visual Comparison of Segmentation Performance

We randomly selected four images from the test set to visualize and compare the prediction results of each model, as shown in Figure 10. From top to bottom, the first row presents the original images (Image), the second row shows the ground truth labels (Label), the third row displays the predictions of the proposed WaveBiSeNet, and the subsequent rows correspond to the predictions of other compared models. Key segmentation regions are highlighted with red dashed boxes to illustrate performance differences among models. In the dispersed state, 1D nanowires show slight overlap, while in complex states, extensive interweaving and background blurring are observed. Such structural complexity often leads to inaccuracies and adhesion in edge segmentation. The visualization results indicate that WaveBiSeNet provides clearer edge restoration, with an improved distinction of fine structures and reduced adhesion between regions. This model excels in preserving local features, such as elongated linear structures and branches. In contrast, BiSeNetV1 [15], FSSNet [55], CANet [58], and LETNet [62] segment large regions reasonably well but are less accurate at fine edges. EDA-Net [56], ESNet [57], AGLNet [59], BiSeNetV2 [60], and EGE-Unet [57] exhibit edge adhesion problems, indicating limited ability to handle complex boundaries. Mobile-Unet [61] performs the worst, with evident segmentation errors and detail omissions, failing to capture complex structures accurately. In summary, WaveBiSeNet demonstrates robust performance in segmentation accuracy, edge restoration, and fine-detail recognition for 1D nanowire images.

5. Conclusions

To address the key challenges in one-dimensional nanowire segmentation, including blurred backgrounds, indistinct edges, and low image contrast, we propose a wavelet-based bilateral segmentation network (WaveBiSeNet). The model was trained and validated on a peptide nanowire dataset. Experimental results show that WaveBiSeNet achieves a mIoU of 77.59%, an Accuracy of 89.95%, an F1 score of 87.22%, and a Kappa coefficient of 74.13%. Under the same data split and evaluation protocol with ResNet18 as the backbone, WaveBiSeNet improves over BiSeNetV1 by 2.12 percentage points in mIoU, 1.26 percentage points in accuracy, 1.56 percentage points in F1 score, and 2.62 percentage points in Kappa. Qualitative analyses indicate that the method effectively handles adherent regions in one-dimensional nanowire materials and exhibits strong robustness under ambiguous backgrounds. Built upon the BiSeNetV1 architecture, the network introduces a Dual Wavelet Convolution Module and a Flexible Upsampling Module. While reducing spatial resolution, it maximally preserves structural information at the edges of the nanowires and enhances context awareness, thereby recovering blurred edge details more faithfully. Overall, WaveBiSeNet provides an accurate and efficient solution for 1D nanowire segmentation. Future work will focus on further optimizing the network architecture to improve segmentation performance and adaptability. Efforts will also be made to enhance the model’s compatibility with various nanomaterial types. Additionally, cross-modal transfer learning and few-shot learning strategies will be explored to facilitate the practical application of deep learning-based semantic segmentation in nanomaterials research.

Author Contributions

Conceptualization, Y.Z.; methodology, Y.H.; investigation, Y.H. and F.L.; software, Y.H. and F.L.; validation, F.L.; resources, Y.Z. and G.L.; funding acquisition, Y.Z.; supervision, Y.Z. and G.L.; writing—original draft, Y.H.; writing—review & editing, Y.H. and Y.Z. All authors have read and agreed to the published version of the manuscript.

Funding

This research was funded by the National Natural Science Foundation of China, grant number: 61604019. The funders had no role in the study design, data collection and analysis, decision to publish, or preparation of the manuscript.

Data Availability Statement

The raw data supporting the conclusions of this article will be made available by the authors on request.

Conflicts of Interest

The authors declare no conflicts of interest.

References

  1. Garnett, E.; Mai, L.; Yang, P. Introduction: 1D Nanomaterials/Nanowires. Chem. Rev. 2019, 119, 8955–8957. [Google Scholar] [CrossRef]
  2. Dasgupta, N.P.; Sun, J.; Liu, C.; Brittman, S.; Andrews, S.C.; Lim, J.; Gao, H.; Yan, R.; Yang, P. ChemInform Abstract: 25th Anniversary Article: Semiconductor Nanowires—Synthesis, Characterization, and Applications. ChemInform 2014, 45. [Google Scholar] [CrossRef]
  3. Jiu, J.; Suganuma, K. Metallic Nanowires and Their Application. IEEE Trans. Compon. Packag. Manuf. Technol. 2016, 6, 1733–1751. [Google Scholar] [CrossRef]
  4. Comini, E.; Sberveglieri, G. Metal oxide nanowires as chemical sensors. Mater. Today 2010, 13, 36–44. [Google Scholar] [CrossRef]
  5. Shi, C.; Owusu, K.A.; Xu, X.; Zhu, T.; Zhang, G.; Yang, W.; Mai, L. 1D Carbon-Based Nanocomposites for Electrochemical Energy Storage. Small 2019, 15, e1902348. [Google Scholar] [CrossRef]
  6. Liu, X.; Gao, H.; Ward, J.E.; Liu, X.; Yin, B.; Fu, T.; Chen, J.; Lovley, D.R.; Yao, J. Power generation from ambient humidity using protein nanowires. Nature 2020, 578, 550–554. [Google Scholar] [CrossRef]
  7. Creasey, R.C.G.; Mostert, A.B.; Solemanifar, A.; Nguyen, T.A.H.; Virdis, B.; Freguia, S.; Laycock, B. Biomimetic Peptide Nanowires Designed for Conductivity. ACS Omega 2019, 4, 1748–1756. [Google Scholar] [CrossRef]
  8. Azam, Z.; Singh, A. Various applications of nanowires. In Innovative Applications of Nanowires for Circuit Design; IGI Global Scientific Publishing: Hershey, PA, USA, 2021; pp. 17–53. [Google Scholar]
  9. Tang, C.Y.; Yang, Z. Transmission electron microscopy (TEM). In Membrane Characterization; Hilal, N., Ismail, A.F., Matsuura, T., Oatley-Radcliffe, D., Eds.; Elsevier: Pokfulam, Hong Kong, 2017; pp. 145–159. [Google Scholar]
  10. Yao, L.; Chen, Q. Machine learning in nanomaterial electron microscopy data analysis. In Intelligent Nanotechnology; Zheng, Y., Wu, Z., Eds.; Elsevier: Urbana, IL, USA, 2023; pp. 279–305. [Google Scholar]
  11. Russell, B.C.; Torralba, A.; Murphy, K.P.; Freeman, W.T. LabelMe: A Database and Web-Based Tool for Image Annotation. Int. J. Comput. Vis. 2007, 77, 157–173. [Google Scholar] [CrossRef]
  12. Shahab, W.; Al-Otum, H.; Al-Ghoul, F. A modified 2D chain code algorithm for object segmentation and contour tracing. Int. Arab J. Inf. Technol. 2009, 6, 250–257. [Google Scholar]
  13. Bai, H.; Wu, S. Deep-learning-based nanowire detection in AFM images for automated nanomanipulation. Nanotechnol. Precis. Eng. 2021, 4, 013002. [Google Scholar] [CrossRef]
  14. Lin, B.; Emami, N.; Santos, D.A.; Luo, Y.; Banerjee, S.; Xu, B.-X. A deep learned nanowire segmentation model using synthetic data augmentation. Npj Comput. Mater. 2022, 8, 1–12. [Google Scholar] [CrossRef]
  15. Yu, C.; Wang, J.; Peng, C.; Gao, C.; Yu, G.; Sang, N. Bisenet: Bilateral segmentation network for real-time semantic seg-mentation. In Proceedings of the European Conference on Computer Vision (ECCV), Munich, Germany, 8–14 September 2018; pp. 325–341. [Google Scholar]
  16. Liu, W.; Wu, Y.; Hong, Y.; Zhang, Z.; Yue, Y.; Zhang, J. Applications of machine learning in computational nanotechnology. Nanotechnology 2022, 33, 162501. [Google Scholar] [CrossRef]
  17. Masson, J.-F.; Biggins, J.S.; Ringe, E. Machine learning for nanoplasmonics. Nat. Nanotechnol. 2023, 18, 111–123. [Google Scholar] [CrossRef]
  18. Dahy, G.; Soliman, M.M.; Alshater, H.; Slowik, A.; Hassanien, A.E. Optimized deep networks for the classification of nanoparticles in scanning electron microscopy imaging. Comput. Mater. Sci. 2023, 223. [Google Scholar] [CrossRef]
  19. Singh, A.V.; Maharjan, R.-S.; Kanase, A.; Siewert, K.; Rosenkranz, D.; Singh, R.; Laux, P.; Luch, A. Machine-Learning-Based Approach to Decode the Influence of Nanomaterial Properties on Their Interaction with Cells. ACS Appl. Mater. Interfaces 2020, 13, 1943–1955. [Google Scholar] [CrossRef]
  20. Lee, B.; Yoon, S.; Lee, J.W.; Kim, Y.; Chang, J.; Yun, J.; Ro, J.C.; Lee, J.-S.; Lee, J.H. Statistical Characterization of the Morphologies of Nanoparticles through Machine Learning Based Electron Microscopy Image Analysis. ACS Nano 2020, 14, 17125–17133. [Google Scholar] [CrossRef]
  21. Banerjee, A.; Kar, S.; Pore, S.; Roy, K. Efficient predictions of cytotoxicity of TiO2-based multi-component nanoparticles using a machine learning-based q-RASAR approach. Nanotoxicology 2023, 17, 78–93. [Google Scholar] [CrossRef] [PubMed]
  22. Bannigidad, P.; Potraj, N.; Gurubasavaraj, P. Metal and Metal Oxide Nanoparticle Image Analysis Using Machine Learning Algorithm. In Proceedings of the International Conference on Big Data Innovation for Sustainable Cognitive Computing, Coimbatore, India, 16–17 December 2022; pp. 27–38. [Google Scholar]
  23. Jia, Y.; Hou, X.; Wang, Z.; Hu, X. Machine Learning Boosts the Design and Discovery of Nanomaterials. ACS Sustain. Chem. Eng. 2021, 9, 6130–6147. [Google Scholar] [CrossRef]
  24. Taylor, J.R.; Das Sarma, S. Vision transformer based deep learning of topological indicators in Majorana nanowires. Phys. Rev. B 2025, 111. [Google Scholar] [CrossRef]
  25. Dosovitskiy, A.; Beyer, L.; Kolesnikov, A.; Weissenborn, D.; Zhai, X.; Unterthiner, T.; Dehghani, M.; Minderer, M.; Heigold, G.; Gelly, S. An image is worth 16×16 words: Transformers for image recognition at scale. arXiv 2020, arXiv:2010.11929. [Google Scholar] [CrossRef]
  26. Lee, J.; Oh, J.; Chi, H.; Lee, M.; Hwang, J.; Jeong, S.; Kang, S.; Jee, H.; Bae, H.; Hyun, J.; et al. Deep Learning-Assisted Design of Bilayer Nanowire Gratings for High-Performance MWIR Polarizers. Adv. Mater. Technol. 2024, 9, 2302176. [Google Scholar] [CrossRef]
  27. Guo, G.; Wang, H.; Bell, D.; Bi, Y.; Greer, K. KNN model-based approach in classification. In OTM Confederated International Conferences” On The Move to Meaningful Internet Systems; Springer: Berlin/Heidelberg, Germany, 2003; Volume 2888, pp. 986–996. [Google Scholar]
  28. Bai, H.; Wu, S. Nanowire Detection in AFM Images Using Deep Learning. Microsc. Microanal. 2020, 27, 54–64. [Google Scholar] [CrossRef]
  29. Li, L. Zno SEM Image Segmentation Based on Deep Learning. In Proceedings of the IOP Conference Series Materials Science and Engineering, Qingdao, China, 28–29 December 2019. [Google Scholar] [CrossRef]
  30. He, K.; Gkioxari, G.; Dollár, P.; Girshick, R. Mask r-cnn. In Proceedings of the 2017 IEEE International Conference on Computer Vision (ICCV), Venice, Italy, 22–29 October 2017; pp. 2961–2969. [Google Scholar]
  31. Redmon, J.; Divvala, S.; Girshick, R.; Farhadi, A. You Only Look Once: Unified, Real-Time Object Detection. In Proceedings of the 2016 IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Las Vegas, NV, USA, 27–30 June 2016; pp. 779–788. [Google Scholar] [CrossRef]
  32. Kendall, J.D.; Pantone, R.D.; Nino, J.C. Deep learning in memristive nanowire networks. arXiv 2020, arXiv:2003.02642. [Google Scholar] [CrossRef]
  33. Hu, X.; Jia, X.; Zhang, K.; Lo, T.W.; Fan, Y.; Liu, D.; Wen, J.; Yong, H.; Rahmani, M.; Zhang, L.; et al. Deep-learning-augmented microscopy for super-resolution imaging of nanoparticles. Opt. Express 2023, 32, 879–890. [Google Scholar] [CrossRef] [PubMed]
  34. He, K.; Zhang, X.; Ren, S.; Sun, J. Deep Residual Learning for Image Recognition. arXiv 2015, arXiv:1512.03385. [Google Scholar] [CrossRef]
  35. Xu, G.; Liao, W.; Zhang, X.; Li, C.; He, X.; Wu, X. Haar wavelet downsampling: A simple but effective downsampling module for semantic segmentation. Pattern Recognit. 2023, 143, 109819. [Google Scholar] [CrossRef]
  36. Finder, S.E.; Amoyal, R.; Treister, E.; Freifeld, O. Wavelet convolutions for large receptive fields. In Proceedings of the European Conference on Computer Vision, Milan, Italy, 29 September–4 October 2024; pp. 363–380. [Google Scholar]
  37. AMS NOTICES. Bull. Am. Meteorol. Soc. 1997, 78, 2696–2697. [CrossRef]
  38. Liu, W.; Lu, H.; Fu, H.; Cao, Z. Learning to upsample by learning to sample. In Proceedings of the 2023 IEEE/CVF International Conference on Computer Vision (ICCV), Paris, France, 1–6 October 2023; pp. 6027–6037. [Google Scholar]
  39. Lu, S.; Montz, B.; Emrick, T.; Jayaraman, A. Semi-supervised machine learning workflow for analysis of nanowire morphologies from transmission electron microscopy images. Digit. Discov. 2022, 1, 816–833. [Google Scholar] [CrossRef]
  40. Lu, S.; Montz, B.; Emrick, T.; Jayaraman, A. Transmission Electron Microscopy (TEM) Image Datasets of Peptide/Protein Nanowire Morphologies. Digit. Discov. 2022, 1, 816–833. [Google Scholar] [CrossRef]
  41. Kirillov, A.; Mintun, E.; Ravi, N.; Mao, H.; Rolland, C.; Gustafson, L.; Xiao, T.; Whitehead, S.; Berg, A.C.; Lo, W.-Y. Segment anything. In Proceedings of the 2023 IEEE/CVF International Conference on Computer Vision (ICCV), Paris, France, 1–6 October 2023; pp. 4015–4026. [Google Scholar]
  42. Mao, A.; Mohri, M.; Zhong, Y. Cross-entropy loss functions: Theoretical analysis and applications. In Proceedings of the International conference on Machine learning, Vienna, Austria, 21–27 July 2024. [Google Scholar]
  43. Sudre, C.H.; Li, W.; Vercauteren, T.; Ourselin, S.; Cardoso, M.J. Generalised Dice overlap as a deep learning loss function for highly unbalanced segmentations. arXiv 2017, arXiv:1707.03237. [Google Scholar] [CrossRef]
  44. Kingma, D.P.; Ba, J. Adam: A method for stochastic optimization. arXiv 2014. [Google Scholar] [CrossRef]
  45. Lateef, F.; Ruichek, Y. Survey on semantic segmentation using deep learning techniques. Neurocomputing 2019, 338, 321–348. [Google Scholar] [CrossRef]
  46. Minaee, S.; Boykov, Y.Y.; Porikli, F.; Plaza, A.J.; Kehtarnavaz, N.; Terzopoulos, D. Image Segmentation Using Deep Learning: A Survey. IEEE Trans. Pattern Anal. Mach. Intell. 2021, 44, 3523–3542. [Google Scholar] [CrossRef] [PubMed]
  47. Takos, G. A survey on deep learning methods for semantic image segmentation in real-time. arXiv 2009, arXiv:2009.12942. [Google Scholar] [CrossRef]
  48. Arefin, S.; Chowdhury; Parvez, R.; Ahmed, T.; Abrar, A.S.; Sumaiya, F. Understanding APT detection using Machine learning algorithms: Is superior accuracy a thing? In Proceedings of the 2024 IEEE International Conference on Electro Information Technology (eIT), Eau Claire, WI, USA, 30 May–1 June 2024; pp. 532–537. [Google Scholar]
  49. Lee, C.-Y.; Hung, C.-H.; Le, T.-A. Intelligent Fault Diagnosis for BLDC with Incorporating Accuracy and False Negative Rate in Feature Selection Optimization. IEEE Access 2022, 10, 69939–69949. [Google Scholar] [CrossRef]
  50. Tan, S.C.; Zhu, S. Binary search of the optimal cut-point value in ROC analysis using the F1 score. In Proceedings of the 5th International Conference on Computer Science and Application Engineering (CSAE 2022), Sanya, China, 25–27 October 2022; p. 012002. [Google Scholar]
  51. Lam, K.F.Y.; Gopal, V.; Qian, J. Confidence Intervals for the F1 Score: A Comparison of Four Methods. arXiv 2023, arXiv:2309.14621. [Google Scholar] [CrossRef]
  52. McHugh, M.L. Interrater reliability: The kappa statistic. Biochem. Med. 2012, 22, 276–282. [Google Scholar] [CrossRef]
  53. Shang, H.; Langlois, J.-M.; Tsioutsiouliklis, K.; Kang, C. Precision/recall on imbalanced test data. In Proceedings of the 26th International Conference on Artificial Intelligence and Statistics (AISTATS 2023), Valencia, Spain, 25–27 April 2023; pp. 9879–9891. [Google Scholar]
  54. Miao, J.; Zhu, W. Precision–recall curve (PRC) classification trees. Evol. Intell. 2021, 15, 1545–1569. [Google Scholar] [CrossRef]
  55. Wang, J.; Zhuang, Y.; Liu, Y. FSS-Net: A Fast Search Structure for 3D Point Clouds in Deep Learning. Int. J. Netw. Dyn. Intell. 2023, 2, 100005. [Google Scholar] [CrossRef]
  56. Yang, C.; Gao, F. EDA-Net: Dense aggregation of deep and shallow information achieves quantitative photoacoustic blood oxygenation imaging deep in human breast. In Proceedings of the Medical Image Computing and Computer Assisted In-tervention–MICCAI 2019: 22nd International Conference, Shenzhen, China, 13–17 October 2019; pp. 246–254. [Google Scholar]
  57. Wang, Y.; Zhou, Q.; Xiong, J.; Wu, X.; Jin, X. ESNet: An efficient symmetric network for real-time semantic segmentation. In Proceedings of the Pattern Recognition and Computer Vision: Second Chinese Conference, PRCV 2019, Xi’an, China, 8–11 November 2019; pp. 41–52. [Google Scholar]
  58. Zhang, C.; Lin, G.; Liu, F.; Yao, R.; Shen, C. CANet: Class-Agnostic Segmentation Networks with Iterative Refinement and Attentive Few-Shot Learning. In Proceedings of the 2019 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), Long Beach, CA, USA, 15–20 June 2019; pp. 5212–5221. [Google Scholar]
  59. Zhou, Q.; Wang, Y.; Fan, Y.; Wu, X.; Zhang, S.; Kang, B.; Latecki, L.J. AGLNet: Towards real-time semantic segmentation of self-driving images via attention-guided lightweight network. Appl. Soft Comput. 2020, 96, 106682. [Google Scholar] [CrossRef]
  60. Yu, C.; Gao, C.; Wang, J.; Yu, G.; Shen, C.; Sang, N. BiSeNet V2: Bilateral Network with Guided Aggregation for Real-Time Semantic Segmentation. Int. J. Comput. Vis. 2021, 129, 3051–3068. [Google Scholar] [CrossRef]
  61. Jing, J.; Wang, Z.; Rätsch, M.; Zhang, H. Mobile-Unet: An efficient convolutional neural network for fabric defect detection. Text. Res. J. 2020, 92, 30–42. [Google Scholar] [CrossRef]
  62. Ta, N.; Chen, H.; Liu, X.; Jin, N. LET-Net: Locally enhanced transformer network for medical image segmentation. Multimed. Syst. 2023, 29, 3847–3861. [Google Scholar] [CrossRef]
Figure 1. Structure of the BiSeNetV1 network.
Figure 1. Structure of the BiSeNetV1 network.
Nanomaterials 15 01612 g001
Figure 2. Structure of the Attention Refinement Module and the Feature Fusion Module. (a) ARM; (b) FFM.
Figure 2. Structure of the Attention Refinement Module and the Feature Fusion Module. (a) ARM; (b) FFM.
Nanomaterials 15 01612 g002
Figure 3. Structure of the WaveBiSeNet network.
Figure 3. Structure of the WaveBiSeNet network.
Nanomaterials 15 01612 g003
Figure 4. Structure of the DWCM.
Figure 4. Structure of the DWCM.
Nanomaterials 15 01612 g004
Figure 5. Structure of the FUM.
Figure 5. Structure of the FUM.
Nanomaterials 15 01612 g005
Figure 6. Peptide nanowire dataset. (a) original image; (b) label image.
Figure 6. Peptide nanowire dataset. (a) original image; (b) label image.
Nanomaterials 15 01612 g006
Figure 7. Model ablation experiment results. (a) BiSeNetV1; (b) BiSeNetV1+FUM; (c) BiSeNetV1 + DWCM; (d) BiSeNetV1 + FUM + DWCM.
Figure 7. Model ablation experiment results. (a) BiSeNetV1; (b) BiSeNetV1+FUM; (c) BiSeNetV1 + DWCM; (d) BiSeNetV1 + FUM + DWCM.
Nanomaterials 15 01612 g007
Figure 8. Curves of different semantic segmentation networks. (a) Loss curve; (b) mIoU curve.
Figure 8. Curves of different semantic segmentation networks. (a) Loss curve; (b) mIoU curve.
Nanomaterials 15 01612 g008
Figure 9. Bootstrap analysis of mIoU between BiSeNetV1 and WaveBiSeNet. (a) mIoU distributions with 95% confidence intervals; (b) mIoU difference distribution with 95% confidence intervals.
Figure 9. Bootstrap analysis of mIoU between BiSeNetV1 and WaveBiSeNet. (a) mIoU distributions with 95% confidence intervals; (b) mIoU difference distribution with 95% confidence intervals.
Nanomaterials 15 01612 g009
Figure 10. Model prediction results. (a) WaveBiSeNet; (b) BiSeNetV1; (c) FSSNet; (d) EDA-Net; (e) ESNet; (f) CANet; (g) AGLNet; (h) BiSeNetV2; (i) Mobile-Unet; (j) EGE-Unet; (k) LETNet.
Figure 10. Model prediction results. (a) WaveBiSeNet; (b) BiSeNetV1; (c) FSSNet; (d) EDA-Net; (e) ESNet; (f) CANet; (g) AGLNet; (h) BiSeNetV2; (i) Mobile-Unet; (j) EGE-Unet; (k) LETNet.
Nanomaterials 15 01612 g010aNanomaterials 15 01612 g010b
Table 1. Comparison of ablation experiments on the peptide nanowire dataset.
Table 1. Comparison of ablation experiments on the peptide nanowire dataset.
MethodmIoU (%)Accuracy (%)F1 (%)Kappa (%)
BiSeNetV175.47% ± 0.3188.69% ± 0.2685.66% ± 0.4071.51% ± 0.62
BiSeNetV1 + FUM76.32% ± 0.3589.34% ± 0.2486.44% ± 0.3472.40% ± 0.46
BiSeNetV1 + DWCM77.29% ± 0.3889.75% ± 0.2586.73% ± 0.3576.62% ± 0.49
BiSeNetV1 + FUM + DWCM77.59% ± 0.4289.95% ± 0.2387.22% ± 0.3074.13% ± 0.33
Table 2. Network model test results.
Table 2. Network model test results.
ModelsBackboneMIoU (%)Accuracy (%)F1 (%)Kappa (%)
WaveBiSeNetResNet1877.59% ± 0.4289.95% ± 0.2387.22% ± 0.3074.13% ± 0.33
BiSeNetV1ResNet1875.47% ± 0.3188.69% ± 0.2685.66% ± 0.4071.51% ± 0.62
FSSNet-69.53% ± 0.6985.22% ± 0.5581.47% ± 0.6063.43% ± 0.53
EDA-Net-68.18% ± 0.6684.51% ± 0.4580.45% ± 0.6961.14% ± 0.64
ESNet-71.21% ± 0.3186.13% ± 0.5382.48% ± 0.5565.34% ± 0.63
CANetMobilenetV271.28% ± 0.6286.57% ± 0.6782.81% ± 0.6065.63% ± 0.51
AGLNet-62.50% ± 0.4579.95% ± 0.5675.76% ± 0.5852.13% ± 0.43
BiSeNetV2-65.50% ± 0.3582.46% ± 0.3478.31% ± 0.4057.21% ± 0.35
Mobile-UnetMobileNet54.37% ± 0.4675.22% ± 0.5869.51% ± 0.3038.64% ± 0.57
EGE-Unet-75.84% ± 0.4088.91% ± 0.6285.84% ± 0.4671.75% ± 0.38
LETNet-62.55% ± 0.3580.44% ± 0.3576.01% ± 0.3552.20% ± 0.45
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Hou, Y.; Zhang, Y.; Liang, F.; Liu, G. A Wavelet-Based Bilateral Segmentation Study for Nanowires. Nanomaterials 2025, 15, 1612. https://doi.org/10.3390/nano15211612

AMA Style

Hou Y, Zhang Y, Liang F, Liu G. A Wavelet-Based Bilateral Segmentation Study for Nanowires. Nanomaterials. 2025; 15(21):1612. https://doi.org/10.3390/nano15211612

Chicago/Turabian Style

Hou, Yuting, Yu Zhang, Fengfeng Liang, and Guangjie Liu. 2025. "A Wavelet-Based Bilateral Segmentation Study for Nanowires" Nanomaterials 15, no. 21: 1612. https://doi.org/10.3390/nano15211612

APA Style

Hou, Y., Zhang, Y., Liang, F., & Liu, G. (2025). A Wavelet-Based Bilateral Segmentation Study for Nanowires. Nanomaterials, 15(21), 1612. https://doi.org/10.3390/nano15211612

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop