Next Article in Journal
Multimodality Imaging Approach in the Diagnosis of Constrictive Pericarditis
Next Article in Special Issue
Generating Findings for Jaw Cysts in Dental Panoramic Radiographs Using a GPT-Based VLM: A Preliminary Study on Building a Two-Stage Self-Correction Loop with a Structured Output (SLSO) Framework
Previous Article in Journal
Reliability and Agreement of a Dual-Method Radiographic Standard vs. Clinical Goniometry for Shank–Forefoot Alignment: A GRRAS-Compliant Study
Previous Article in Special Issue
Segmentation-Guided Preprocessing Improves Deep Learning Diagnostic Accuracy and Confidence of Ameloblastoma and Odontogenic Keratocyst in Cone Beam CT Images—A Preliminary Study
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

SBMN: Similarity-Based Memory Network for the Diagnosis of Vertical Root Fracture in Dental Imaging

1
School of Electronic Science and Engineering, Nanjing University, Nanjing 210023, China
2
Department of Dentomaxillofacial Radiology, Nanjing Stomatological Hospital, Medical School of Nanjing University, Nanjing 210008, China
*
Authors to whom correspondence should be addressed.
Diagnostics 2026, 16(5), 710; https://doi.org/10.3390/diagnostics16050710
Submission received: 12 January 2026 / Revised: 30 January 2026 / Accepted: 12 February 2026 / Published: 27 February 2026
(This article belongs to the Special Issue Application of Artificial Intelligence to Oral Diseases)

Abstract

Background/Objectives: Medical image analysis of vertical root fractures (VRFs) is challenged by limited annotated data, class imbalance, and subtle inter-class differences. To address these issues, we propose an SBMN: a Similarity-Based Memory Network that integrates Category Memory with the Basic SBMN Module and a similarity-based classifier. Methods: An SBMN stores representative features for each class and leverages similarity-based gating to enhance feature discrimination. Experiments were conducted on a CBCT dataset of fractured and non-fractured teeth to evaluate performance. Results: The SBMN achieved up to 97.1% and 99.7% classification accuracy on automatically and manually segmented images, respectively. Memory manipulation experiments confirm the critical role of Category Memory in controlling classification outcomes. Conclusions: These results indicate that SBMNs offer an effective and interpretable approach for small-sample medical image classification and diagnosis.

1. Introduction

Vertical root fracture (VRF) is a longitudinal fracture that begins at the apical root and extends coronally, often detectable on facial or lingual surfaces [1,2,3,4]. Classified by the American Association of Endodontists as the most severe type of cracked tooth, VRF accounts for 2–5% of tooth fractures [5] and ranks as the third leading cause of tooth loss [6]. It occurs in both endodontically and non-endodontically treated teeth, though the prevalence is considerably higher in the former [7,8,9]. Interestingly, most cases in untreated teeth have been reported in Chinese populations, likely reflecting dietary differences [10]. Despite its clinical significance, early diagnosis of VRF remains challenging because clinical manifestations are often subtle, nonspecific, or completely asymptomatic until substantial periapical destruction occurs. This delay frequently leads to misdiagnosis, inappropriate treatment, and eventual extraction [11,12]. Moreover, clinical presentations often mimic periodontal disease [13], and conventional diagnostic tools—including illumination, staining, bite tests, inspection, surgical exploration, and radiography—are limited in reliability [14]. Radiographs in particular may fail when the X-ray beam is not aligned with the fracture plane, leaving surgical inspection as the only definitive method, which is invasive and potentially damaging [15]. Collectively, these limitations not only compromise diagnostic accuracy but also increase clinical workload, highlighting the need for more reliable, reproducible, and non-invasive decision-support strategies.
Cone-beam computed tomography (CBCT), introduced in the late 1990s to reduce radiation exposure [16], has become a major advancement in dentomaxillofacial imaging and is widely applied to root canal visualization and VRF detection [17,18]. However, when fracture lines approximate the voxel size of CBCT, they may appear blurred, making diagnosis heavily dependent on radiologist expertise [16,19]. Consequently, although CBCT provides three-dimensional information, accurate interpretation still depends heavily on expert experience and careful slice-by-slice inspection. This manual review process is time-consuming in routine practice, particularly when screening large CBCT volumes, and may lead to inter-observer variability across clinicians. In recent years, artificial intelligence (AI) has increasingly been adopted as an assistive tool in medical image analysis. Convolutional neural networks (CNNs) remain dominant in computer vision and have achieved substantial success in diverse imaging tasks [20,21,22]. More recently, Transformer-based models [23], including Vision Transformer (ViT) [24] and Swin Transformer [25], have further improved image classification and segmentation performance. Nevertheless, these models typically require large-scale pretraining to generalize effectively, whereas VRF datasets—like many biomedical datasets—are small, noisy, and imbalanced. Such characteristics hinder stable clinical deployment of conventional networks and frequently result in overfitting or poor generalization across institutions. In this context, memory-augmented neural networks offer a promising solution. By incorporating an external memory module that can be selectively read from and written to during training and testing, these networks can retain task-relevant information and mitigate overfitting. This mechanism enhances generalization on small and heterogeneous datasets, making memory networks particularly suitable for biomedical applications such as VRF diagnosis, where data scarcity is a major challenge.
Additionally, our dataset was collected from a single center using a single CBCT device, which may limit generalizability. Despite this limitation, from a translational perspective, an AI-assisted VRF diagnostic system can be integrated into routine clinical workflows. In practice, CBCT volumes can first undergo automatic tooth segmentation, followed by per-tooth classification to generate fracture probability scores. These predictions can serve as a rapid screening or decision-support mechanism, highlighting suspicious regions for radiologists or endodontists, particularly in early-stage fractures, subtle fracture lines, or anatomically complex cases.
In this paper, we propose a novel memory network based on the calculation of similarity (SBMN) to classify VRF CBCT images. A key innovation of our approach is the introduction of category memory, which is stored independently of the network architecture and preserves the distinctive feature representations of each image category. Inspired by the attention mechanism in Transformer models [23], we design a similarity computation strategy that measures the relationship between stored category memory and features extracted from new input images. Within the SBMN module, category memory and input features are mutually updated: the stored memory guides the refinement of incoming features, while the new inputs simultaneously update and enrich the category memory. This bidirectional interaction enables more robust feature learning and enhances classification performance on small-scale medical datasets.
The contributions of our method can be summarized into three folds:
  • We propose a novel classification framework that integrates similarity-based computation with category memory, enabling automatic diagnosis of VRFs.
  • To ensure effective categorization of memory, we design a dedicated similarity loss function. By accurately aligning category memory with the corresponding class, the memory can better enhance the features extracted from input images, thereby improving classification—and ultimately diagnostic—accuracy.
  • We introduce a bidirectional update mechanism between category memory and input features during training. On the one hand, category memory is continuously refined by assimilating the characteristics of images from its corresponding class. On the other hand, each input image retrieves relevant memory components for fusion and feature enhancement, leading to more robust representation learning.

2. Related Work

2.1. Convolutional Neural Network (CNN)

Convolutional Neural Networks (CNNs) [26] have played a foundational role in computer vision, with their theoretical roots in the concept of the receptive field. In 1980, Fukushima introduced a neural architecture incorporating convolutional and pooling operations [27], which later inspired LeNet-5 [28], the first convolutional network successfully trained using backpropagation. Despite these early contributions, CNNs gained widespread recognition only after the success of AlexNet [21], which demonstrated the effectiveness of deep convolutional models on large-scale visual recognition tasks. This breakthrough accelerated the development of deeper and more structured architectures, including VGG [29], GoogLeNet [30], ResNet [22], DenseNet [31], and EfficientNet [32]. Recent surveys have systematically reviewed the evolution of CNN architectures and their applications in image classification, segmentation, and object detection, highlighting ongoing innovations in network design and efficiency [33,34,35]. Although Transformer-based models [23] have shown strong performance in various vision tasks, CNNs have not been supplanted but instead refined and modernized. Notably, ConvNeXt [36] demonstrates that well-designed CNNs can achieve performance comparable to, or even exceeding, that of vision transformers on multiple benchmarks. Motivated by these developments, we adopt CNNs as the backbone of the proposed SBMN framework to extract discriminative features from VRF CBCT images.

2.2. Transformer and Similarity

The emergence of the Transformer architecture [23] has profoundly impacted artificial intelligence, rapidly establishing itself as a leading framework for numerous natural language processing (NLP) tasks [37,38,39,40,41]. Beyond NLP, Transformers have also been successfully applied to core computer vision (CV) tasks, including segmentation [42,43,44,45], object detection [46,47,48], and image classification [24,25,49,50], as well as other vision-related applications. The key to their success lies in the self-attention mechanism, which captures correlations among embedded tokens. In NLP, higher attention weights reflect stronger semantic or contextual relationships. In vision, feature representations are more spatially grounded, and CNN-based architectures remain highly effective, particularly in extracting and hierarchically analyzing discriminative features for classification and segmentation. Recent surveys demonstrate that Transformers and their hybrid CNN–Transformer models continue to evolve across image and video tasks, addressing the trade-off between local and global dependencies [51,52,53,54]. These studies provide a systematic overview of Vision Transformers, their variants, and comparative analyses with CNNs, highlighting ongoing innovations in architecture design and task-specific adaptations.

2.3. Memory Networks

The concept of memory in neural networks [55,56] dates back to the late 20th century, with recurrent neural networks (RNNs) [57] and long short-term memory (LSTM) [58] designed to capture long-range dependencies in sequential data. However, the “memory” in these architectures is limited to internal hidden states, which are inherently transient and unstable over extended sequences. To address tasks requiring retention of earlier information—such as document-based question answering—memory networks with external memory modules were introduced [59,60,61]. In recent years, memory networks have gained increasing attention in computer vision, with applications in video object segmentation [62,63,64], domain adaptation [65], anomaly detection [66,67,68,69], and image classification [70]. Unlike conventional memory networks that store and retrieve past inputs as static memory, modern approaches leverage similarity-based interactions between inputs and stored memory. Recent advances in similarity-based and Hopfield memory mechanisms provide enhanced capacity and retrieval dynamics for vision tasks, including anomaly detection and spatiotemporal segmentation [71,72,73]. This design allows simultaneous refinement of both input features and memory content, enabling dynamic updates that improve overall network performance.

3. Materials and Methods

3.1. Materials

3.1.1. Patients and Datasets

We used a dataset composed of 216 patients (126 males and 90 females, with an age range between 19–86 years). All CBCT images were collected by a NewTom VG scanner (QR SRL, Verona, Italy) with a voxel size of 0.15 mm, 110 kV, 3.6–3.7 mA, a field of view of 12 × 8 cm, and an acquisition time of 5.4 s. The selection criteria for CBCT images of VRF teeth were as follows: (a) non-endodontically treated teeth; (b) images exhibiting high quality and free from any artifacts, such as motion or beam hardening artifacts. For CBCT images of non-VRF teeth, three types of teeth were selected: (a) healthy teeth; (b) apical periodontitis teeth caused by caries; (c) periodontitis teeth with horizontal bone loss. Of all CBCT images from these patients, 139 VRF and 139 non-VRF teeth were independently annotated by two experienced radiologists, each with more than 10 years of clinical experience. Inter-examiner agreement between the two radiologists was almost perfect (Kappa = almost perfect), while intra-examiner agreement for each radiologist was also almost perfect. In addition, a third radiologist, blinded to the study design and patient information, performed a repeat evaluation three months later, showing substantial agreement with the consensus annotation (Kappa = 0.711). This procedure ensured high-quality and reliable labeling of VRF and non-VRF teeth. The study was approved by the Ethics Committee of Nanjing Stomatological Hospital, Medical School of Nanjing University [refEthics2018NL044], and the requirement for written informed consent was waived by the Ethics Committee. While the dataset is de-identified, we acknowledge that AI models may introduce biases, especially when applied to vulnerable populations or groups underrepresented in the dataset. Future work should carefully monitor potential biases and ensure equitable application of AI-based diagnostic tools [74].

3.1.2. Image Preprocessing

The collected CBCT images were segmented using two methods: auto segmentation and manual segmentation. The auto segmentation method, consisting of five steps, was introduced by Hu et al. [75]. The effectiveness and reliability of this automatic tooth segmentation approach were independently validated in the same study, which reported accurate and stable performance on CBCT images. Furthermore, all automatically segmented results in our study were visually reviewed by experienced radiologists to ensure anatomical plausibility before further analysis. The manual segmentation method involved experienced radiologists manually segmenting VRF and non-VRF teeth from the same raw dentition images. Table 1 shows the specific segmentation results of our study. After segmentation, the segmented slices were resized to 224 × 224 to match the fixed input size required by the network, and gray-level transformation was applied to the images. Considering that this is a medical dataset with relatively small volume and potential biases, networks are prone to overfitting. Therefore, we applied several data augmentation methods. Specifically, the input images were randomly flipped horizontally and vertically, color jitter was applied (brightness = 0.5, contrast = 0.5, saturation = 0.5, hue = 0.3), and images were randomly rotated within the range of −90° to 90°. All augmentation methods were applied only to the training dataset, while all segmented images were normalized before being fed into the network.

3.2. Methods

In this section, we first provide a concise overview of our similarity-based memory network architecture for the classification and diagnosis of vertical root fractures (Section 3.2.1). We then describe each component of the network in detail (Section 3.2.2), followed by a comprehensive explanation of the associated loss functions (Section 3.2.3).

3.2.1. Overview

The overall architecture of an SBMN is illustrated in Figure 1. The central component is the SBMN module, highlighted in light blue. This module takes two types of inputs: (i) the image feature representations extracted by the embedding network (implemented with ResNet50 in this study), and (ii) the category memory. It produces three outputs, namely remake_ x i , similarity matrix, and modified category memory. To generate the final classification result, we construct a similarity-based classifier using the embedding network output together with the modified category memory, replacing the conventional MLP classifier. The remake_ x i and similarity matrix are used to compute the Remake Loss and Similarity Loss, respectively, while the modified category memory is iteratively fed back to replace the original memory for memory refinement. Within the SBMN module, multiple basic SBMN units are stacked to perform a sequence of operations—primarily similarity computation—on the two input branches, ultimately producing the aforementioned outputs. Additional architectural details are provided in the following subsections.

3.2.2. Network Architecture

Embedding Network
The embedding network is responsible for extracting features from the input images. The SBMN module is designed as a universal architecture that can be coupled with various embedding networks, provided that their output dimensionality is fixed to H (e.g., H = 2048 , which is used by default). Both CNN-based architectures, such as ResNet and DenseNet, and Transformer-based architectures, such as ViT, can serve as embedding networks. All embedding networks are initialized with parameters pre-trained on ImageNet classification.
Category Memory
Category Memory ( C M ) is a memory collection that stores k sets of memory corresponding to each category, where C M R M × k × H refers to the size of each memory set (default = 16), k is the number of categories to be classified depending on the task, and H denotes the dimension of each slice of memory, which is set to be the same as the output dimension of the embedding network. Similarly to [76], our C M consists of feature representations of input images, which significantly reduces memory requirements. The C M is randomly initialized with mean 0 and variance 1. To address the issue of low consistency reported in [76], we apply a multi-head encoder to the memory before subsequent calculations, instead of using the raw memory directly. Furthermore, m can be set to a reasonable size, and the C M is updated after every training step.
Basic SBMN Module
The main module, the SBMN Module, consists of k Basic SBMN Modules. Each Basic SBMN Module has the same structure but does not share parameters. The Basic SBMN Module has four heads structure, trying to capture different useful information. All heads within a Basic SBMN Module are independent. For clarity, only one representative Basic SBMN Module, including its multi-head structure, is shown in Figure 2.
Encoder. The encoder takes both feature representation and t-th category memory as inputs, where t 1 , 2 , , k . We denote by x i R B × H and m t R M × H for feature representation and t-th category memory. Here, B denotes the batch size of the current input. The encoder consists of two different linear layers, E K and E V to generate Key and Value respectively, denoted by K x R B × H / 4 , V x R B × H / 2 , K m R M × H / 4 and V m R M × H / 2 . We depict the process as follows.
K x , K m = E K ( x i , m t )
V x , V m = E V ( x i , m t )
Similarity Calculation and Weighting. After acquiring K x , K m , drawing on the idea of contrastive learning, we decide to determine the degree of modification of Category Memory and x i through the calculation of similarity. We can express this process with the following formula:
s i m = K x K m T
W m = l a y e r _ s o f t m a x ( s i m )
W x = b a t c h _ s o f t m a x ( s i m )
W V m = W m V m
W V x = W x T V x
where ⊙ denotes the matrix multiplication and T means matrix transpose. We denote by s i m R B × H the similarity of x i and m t . l a y e r _ s o f t m a x and b a t c h _ s o f t m a x are both softmax normalization operations, the only difference between them is that l a y e r _ s o f t m a x performs softmax on dimension M, while b a t c h _ s o f t m a x acts on dimension B. Then we calculate the weighted sum of V m and V x according to the normalized s i m and denote them by W V m R B × H / 2 and W V x R M × H / 2 . Before outputting the similarity vector s i m ¯ R B × 1 , which measures how similar x i is to this certain category of memory m t , sigmoid activation is required and then averaged on dimension M:
s i g s i m = σ ( s i m )
s i m ¯ = m e a n ( s i g s i m )
here σ and m e a n ( ) refers to the sigmoid function and the average operation on dimension M.
Decoder. We use calculated weighted sum W V x to modify m t and W V m to modify x i . We concatenate W V x with V M and W V m with V x , denoted by m o d m ˜ and m o d x ˜ . Then we put four heads of m o d m ˜ and m o d x ˜ into decoder. The decoder consists of two parts. The first part is a mix module consisting of a two-layer MLP, functioning as mixing m o d m ˜ or m o d x ˜ . We denote this module as f m i x . The second part is f d e c to combine separated multi-heads together. The total process is shown as follows:
m o d x = f d e c f m i x m o d x 1 ˜ , f m i x m o d x 2 ˜ , f m i x m o d x 3 ˜ , f m i x m o d x 4 ˜
m o d m = f d e c f m i x m o d m 1 ˜ , f m i x m o d m 2 ˜ , f m i x m o d m 3 ˜ , f m i x m o d m 4 ˜
where m o d x R B × H , [ , ] means concatenate operation.
Remake Decoder. The remake decoder is designed to validate the effectiveness of the stored memory and to exploit it appropriately. In the similarity calculation and weighting stage, we assume that the input representation can extract relevant information from the memory bank, and expect this information to be meaningful and capable of reconstructing the input representation. To achieve this, the decoder concatenates the weighted sums of category memories across all heads and passes the result through a feed-forward layer to obtain r e m a k e _ x i , formulated as follows:
r e m a k e _ x i = f r e _ d e c W V m 1 , W V m 2 , W V m 3 , W V m 4
Similarity-Based Classifier
To fully exploit the representational capability of the category memory within our network, we construct it as a classifier, as illustrated in Figure 3. Specifically, the model generates its final predictions by computing the similarity between the input feature representation and the category memory, with the resulting similarity scores serving directly as the classification output. Under this formulation, when the feature of an input sample is more similar to the memory of a particular class, the model naturally assigns a higher likelihood to that class.
A key detail is that the category memories used for similarity computation differ between the training and testing phases. During training, the model utilizes the class memories updated by the SBMN module. This enables the network to continuously refine and maintain an up-to-date and accurate memory bank through loss optimization and backpropagation. However, during testing, the category memories are kept unchanged because the updates generated by the SBMN module are no longer constrained by gradient optimization.
The similarity computation itself can be considered a gated similarity mechanism. Its formulation may either follow the same similarity-gating strategy used in the primary SBMN module or adopt a different similarity function. This flexibility allows the similarity-based classification module to be adapted to different datasets or specific application needs.

3.2.3. Training Loss Functions

Here the overall training loss is a weighted sum of three different losses:
L t o t a l = λ c l s L c l s + λ s i m L s i m + λ r e m a k e L r e m a k e
where we denote by L c l s , L s i m and L r e m a k e classification, similarity and remake losses, respectively. λ is weighted factor ranged between 0 to 1. The classification loss ensures the accuracy of classification to achieve the purpose of diagnose VRFs, the similarity loss helps maintain the validity of each set of memory in Category Memory, while remake loss allows us to generate feature representations of a certain category from stored Category Memory. The detailed descriptions of each loss are presented in the following.
Classification Loss. The final classification is performed using a similarity-based classifier. Given an input image, the embedding network produces a feature representation x i , which is further processed by the SBMN Module to obtain the modified category memory m o d m R M × K H . The feature representation x i and the modified category memory m o d m are jointly fed into the classifier to generate the final prediction:
p r e d c l s = f c l s ( x i , m o d m ) ,
where f c l s denotes the classifier and p r e d c l s R B × K represents the predicted label probabilities of the input images. We employ the cross-entropy loss function to compute the classification loss:
L c l s = i = 1 K l a b e l i log ( p r e d c l s i ) ,
where l a b e l i = 1 if the input sample belongs to the i-th category, and l a b e l i = 0 otherwise.
Similarity Loss. The similarity loss is designed to help maintain the validity of Category Memory. As we all know, all inputs will modify Category Memory. The degree of modification depends on the similarity between their feature representations and a certain category of memory. The more similar it will be, the more it will be modified. To reduce the excessive memory modification resulting in one category of memory being modified to another, we design and use similarity loss function to restrict this. The loss function is shown below:
L s i m = 1 K i = 1 K j N l a b e l j i l o g ( s i m ¯ j i ) + 1 l a b e l j i l o g 1 s i m ¯ j i
where N denotes the total num of inputs and the l a b e l here is in one-hot form.
Remake Loss. The remake loss is originally designed to explore the use of memory. We wish that under some certain kinds of stimulus, typical feature representations of a certain category can be restored from memory. Here we use r e m a k e _ x i , denoted as r x i , together with x i to calculate remake loss.
L r e m a k e = M S E ( r x i , x i ) = r x i x i 2 2

4. Results

We choose to use SGD as our optimizer. The SGD weight decay is set as 0.0004 and the SGD momentum is 0.9. We use a mini-batch size of 32 and an initial learning rate of 0.005. The learning rate adjustment strategy is to multiply the current learning rate by 0.95 after every 5 epochs, the minimum learning rate is 0.00001. We train our model on Tesla V100 GPU with 16G memory. These hyperparameters were selected based on prior studies in small-sample medical imaging tasks and preliminary experiments on our dataset, which indicated stable convergence and high classification performance.

4.1. Evaluation Metrics

In order to evaluate the performance of our model quantitatively, we chose to calculate diagnostic accuracy, sensitivity, specificity, positive predictive value (PPV) and F1 score of both auto-segmentation dataset and manual segmentation dataset. Since we only need to distinguish the input images with or without fracture, we can easily have access to the confusion matrix of the classification results, then we can calculate the metrics mentioned above. In addition, the random division of datasets may bring contingency and uncertainty to the classification results, thus causing errors in the evaluation of network performance. To avoid this condition, we conducted five-folds cross validation and calculated the average metrics as final results.

4.2. Results of SBMN

An SBMN exhibits strong flexibility and can be readily adjusted by employing various embedding networks. Considering that automatic segmentation better reflects realistic clinical workflows, the auto-segmentation results are regarded as the primary evaluation, whereas manual segmentation is used only as a controlled reference. In this study, we empirically experiment with VGG19, ResNet50, DenseNet169, ResNeXt, and ViT-B as embedding networks of SBMNs for binary classification. The corresponding results are presented in Table 2.
As shown in Table 2, within the auto-segmentation group, the ResNet50-based SBMN achieves the highest accuracy and F1 score, whereas the ResNeXt-based SBMN attains the best PPV and specificity. Specifically, the ResNet50-based SBMN reaches an accuracy of 97.1%, a sensitivity of 98.6%, and an F1 score of 97.1%, which are 2.5%, 4.9%, and 2.6% higher than the second-best results, respectively. Meanwhile, the ResNeXt-based SBMN achieves the highest specificity (97.1%) and PPV (97.0%), outperforming the ResNet50-based model by 1.4% and 1.2%. In contrast, the Inception_v3-based SBMN shows inferior performance compared with other backbones, while the remaining networks maintain similar and relatively high results.
In the manual-segmentation group, all models exhibit higher performance overall. This improvement is expected because experienced radiologists manually exclude low-quality or ambiguous regions, which reduces background interference and may lead to performance inflation. Therefore, these results should be interpreted as an upper-bound reference under ideal conditions rather than a realistic clinical scenario. Since clinical diagnosis prioritizes the detection of all potential vertical root fractures, sensitivity remains an important metric. The ResNet50-, VGG19-, ResNeXt-, and ViT-B-based SBMNs all reach 100.0% sensitivity; however, this may reflect the specific characteristics and limited size of our datasets rather than definitive model superiority. The DenseNet169-based SBMN achieves the highest specificity and PPV. Overall, the ViT-based SBMN demonstrates the highest accuracy and F1 score (99.7%), but these results should be interpreted with caution given the sample size and dataset characteristics. Considering the limited sample size, only descriptive statistics including accuracy, sensitivity, specificity, PPV, and F1 score are reported. Inferential statistics such as confidence intervals or p-values were not calculated, as they may not be reliable with the current dataset. Future studies with larger sample sizes will allow more robust statistical analyses.

4.3. Comparisons with Other Networks

To further validate the effectiveness of the proposed model, the ResNet50-based SBMN is compared with several representative convolutional neural networks (CNNs), including VGG19, ResNet50, DenseNet169, Inception_v3, and ResNeXt, as well as Transformer-based architectures such as ViT-B, Swin-T, and Mobile-ViT, alongside the SCCNN model. For fair comparison, the compared models are equipped with multilayer perceptron (MLP) classifiers of identical input dimensions to minimize the potential influence of parameter quantity on the evaluation metrics. In addition, radiologists are invited to diagnose the manually segmented dataset to provide reference metrics. The comparative results are summarized in Table 3.
As shown in Table 3, the proposed SBMN consistently outperforms all other networks across most evaluation metrics. In the auto-segmentation group, the SBMN achieves the highest accuracy 97.1% and sensitivity 98.6%, which are 3.2% and 3.6% higher than the second-best model, respectively. In the manual-segmentation group, the SBMN ranks first in accuracy, sensitivity, and F1 score, while its specificity and PPV remain within 1.4% of the highest results. Furthermore, stable performance under automatic segmentation demonstrates the feasibility of deep learning-based assistance in realistic clinical scenarios.

4.4. Visualization

To gain a more intuitive understanding of the Category Memory mechanism and the effect of the SBMN module on the feature extraction capability of the embedding network, several visualization experiments are conducted to analyze its influence on the feature representations. Specifically, two images are randomly selected from the datasets with vertical root fractures (VRF) and without fractures (non-VRF), forming three comparison groups: (a) non-VRF vs. non-VRF, (b) non-VRF vs. VRF, and (c) VRF vs. VRF.
As illustrated in Figure 4, three-dimensional scatter plots are generated to demonstrate the feature distributions under different network configurations, corresponding to the Untrained ResNet50, Trained ResNet50, and SBMN (based on the ResNet50 embedding network) models. After feeding each image pair into the networks, the embedding feature representations are obtained as x i R H (where H = 2048 ). In the 3D scatter plots, each point represents one feature channel, with the horizontal and vertical axes indicating the feature values of the same channel from the two images, respectively. These visualizations intuitively reveal how the introduction of the SBMN module influences the structure and discriminative characteristics of the feature distributions.
In the untrained ResNet50, the scatter points are widely dispersed and lack clear patterns, indicating that the extracted features across channels exhibit weak discriminative consistency. After training on the tooth fracture dataset, the ResNet50 shows more organized feature distributions: the scatter points of same-class image pairs (non-VRF–non-VRF or VRF–VRF) are distributed closer to the 45° diagonal line, reflecting higher inter-channel similarity, whereas those of different-class pairs (non-VRF–VRF) are located nearer to the coordinate axes, suggesting greater inter-class differences. When the SBMN is employed, these differences become even more pronounced. Scatter points for same-class pairs are densely distributed along and around the 45° diagonal line, while those for different-class pairs are mainly concentrated near the coordinate axes. This observation indicates that the SBMN module significantly enhances the embedding network’s discriminative feature extraction capability, enabling it to more effectively distinguish between intra-class and inter-class inputs, thereby improving the overall feature representation performance.
To further investigate how the SBMN module affects the network’s focus on diagnostically relevant regions, we employ visualization techniques to compare feature localization between ResNet50 and the SBMN (based on the ResNet50 embedding network). Figure 5 presents the results obtained using Grad-CAM and guided backpropagation. Three representative CBCT images with vertical root fractures (VRF) and three non-VRF images are selected for visualization. Figure 5a–c depict fractured teeth, where yellow rectangles indicate the specific locations of root fractures, while Figure 5d–f show non-fractured teeth.
As illustrated, ResNet50 exhibits relatively diffuse activation regions that are not precisely aligned with the actual fracture sites. In contrast, the SBMN produces heatmaps and activation distributions that are more concentrated and better correspond to the clinically relevant fracture areas. This observation indicates that the incorporation of Category Memory enhances the network’s attention to diagnostically important regions, thereby improving interpretability and diagnostic reliability. Moreover, the guided backpropagation results of the SBMN also reveal finer texture details within the fracture zones, further confirming its superior capability in identifying subtle structural variations.

4.5. Ablation Study

To further investigate the effect of different similarity calculation strategies within the proposed SBMN framework, four commonly used methods are compared: dot product, cosine similarity, Pearson correlation, and Euclidean distance. Table 4 summarizes the evaluation metrics obtained from each approach, including accuracy, sensitivity, specificity, PPV, and F1 score.
Among all tested methods, the dot product achieves the best overall performance, yielding the highest accuracy (97.1%) and F1 score (97.1%), as well as superior sensitivity (98.6%). This result suggests that the dot product provides a more effective measure of feature consistency in the latent representation space, facilitating more discriminative learning for vertical root fracture (VRF) detection. Cosine similarity shows comparable results, with slightly higher specificity (98.2%) and PPV (96.4%), indicating its robustness in distinguishing negative samples. In contrast, both Pearson correlation and Euclidean distance perform relatively worse across all metrics, implying that linear correlation and distance-based similarity are less capable of capturing subtle semantic relationships in feature embeddings. In conclusion, the results demonstrate that different similarity-gated computation methods can all be effectively integrated within the SBMN framework. Depending on specific application scenarios and computational requirements, the similarity function can be flexibly adjusted to achieve an optimal balance between discriminative performance and efficiency.

5. Discussion

This study evaluated the clinical feasibility and diagnostic performance of the proposed Similarity-Based Gated Memory Network (SBMN) for detecting vertical root fractures (VRFs) on CBCT images using a relatively small real-world dataset. The primary aim was to improve diagnostic reliability under small-sample conditions by incorporating a memory mechanism that stabilizes feature learning and reduces overfitting. The main contributions and findings of this work can be summarized as follows:
  • Introduction of Category Memory: We introduce a category-level memory mechanism that preserves representative patterns of fractured and non-fractured teeth. By retaining stable class-specific information across training samples, the model can better handle limited and heterogeneous medical data, which are common in clinical studies.
  • Design of the Basic SBMN Module: This similarity-guided design enables the network to selectively emphasize clinically relevant features while suppressing noise and irrelevant structures frequently observed in CBCT images. By integrating similarity computation with gating operations and a similarity loss function, this module ensures the stability of the Category Memory repository while maintaining discriminative and robust feature representations.
  • Proposal of the Similarity-based Classifier: Instead of relying solely on conventional fully connected classifiers, predictions are generated based on similarity between current inputs and stored category representations. This strategy provides a more transparent decision process and improves interpretability, which is desirable for clinical adoption.
  • Experimental Validation on Small Medical Datasets: On a small dataset of vertical root fractures (VRF), the SBMN achieved high diagnostic accuracy on both automatically and manually segmented images, indicating that the memory mechanism can effectively capture subtle structural differences associated with fractures. Importantly, competitive performance was maintained under automatic segmentation, which better reflects realistic clinical scenarios where manual annotation is not feasible.
The proposed framework can be seamlessly integrated into routine clinical workflows through automatic tooth segmentation and subsequent tooth-level fracture probability estimation, thereby serving as a rapid screening or decision-support tool rather than replacing expert judgment. Misclassified cases predominantly involved teeth with very thin fracture lines, severe imaging artifacts, or complex periodontal destruction. Although analysis of these errors provides a general understanding of clinically challenging scenarios, the specific features underlying the network’s decisions remain inherently opaque. Accordingly, such error analysis should be regarded as a high-level descriptive assessment rather than a definitive explanation of the model’s decision-making process. This study is limited by the use of a single-center dataset acquired from a single CBCT device, which may restrict generalizability. In addition, the relatively limited sample size may reduce statistical power, requiring larger cohorts to more reliably confirm the stability and reproducibility of the reported performance. While cross-validation demonstrates stable results and supports the robustness of the category memory mechanism across folds, the risk of overfitting cannot be entirely excluded. In practical applications, additional training or fine-tuning may be necessary for different CBCT scanners, populations, or imaging protocols to ensure consistent performance. Future multi-center and prospective studies are therefore warranted to further validate the robustness and generalizability of the proposed approach.
Overall, the results suggest that the SBMN is a promising decision-support approach for VRF detection, although further large-scale and multi-center validation is required before routine clinical deployment.
To further validate the effectiveness of the memory mechanism, we conduct experiments to analyze the relationships between the hidden Category Memory and the feature representations, both obtained through the encoder within the SBMN module. The experiments focus on the channel-level interactions between these two representations, aiming to demonstrate how Category Memory participates in and strengthens the feature extraction process of the embedding network.
As illustrated in Figure 6, we select one randomly chosen non-VRF image and plot the scatter distributions of 2048 channel pairs, where the x-axis represents the feature value from the image representation x _ K , while the y-axis denotes the corresponding value from the memory representation h i d _ K , both obtained through the encoder within the SBMN module. Here, the notations 1 and 2 indicate that the representations are derived from two different Basic SBMN modules.
In Figure 6a, where the image feature and the memory representation belong to different classes, the scatter points are roughly distributed along the inverse diagonal Y = X , indicating weak or negative correlation. In contrast, Figure 6b shows the case where the image and memory belong to the same class. Here, the scatter points align closely along the diagonal Y = X , reflecting strong positive correlation. This pattern is consistent with our expectations: the similarity-based gating computation in the SBMN performs a dot product between corresponding feature and memory vectors across all channels. Points along Y = X yield a large positive contribution to the similarity score, whereas points along Y = X produce a small negative contribution. These results demonstrate that the current similarity computation effectively captures the relationship between feature representations and their corresponding Category Memory.
Furthermore, Figure 6c,d illustrate the relationships between two different Basic SBMN modules. The scatter points in these plots are randomly distributed around the origin, showing no apparent linear correlation. This inter-module decorrelation confirms that different SBMN modules independently encode distinct category-specific knowledge, avoiding feature redundancy and enhancing the overall discriminative capacity and interpretability of the network.
To further validate the effectiveness of the Category Memory, we perform memory manipulation experiments in which the stored memories are modified in various ways, including replacing all memories with zero vectors, using a single-class memory, random initialization, or swapping memories between different classes. These modified memories are then used as input to the network for classification prediction. The results show that when memories are replaced with zero vectors or a single-class memory, inter-class similarity scores become indistinguishable. Under these conditions, the network defaults to selecting the first indexed class, resulting in perfect accuracy for one class while completely misclassifying the other. In the randomized memory experiments, prediction performance fluctuates significantly, with most trials exhibiting a substantial drop in accuracy and only occasional instances of relatively high accuracy. In the memory swapping experiments, correct classification is achieved only when each memory unit contains the corresponding class-specific memory.
Collectively, these findings demonstrate that the network’s classification performance critically depends on the integrity and class-specific content of the Category Memory. Moreover, they confirm that modifying the memory can directly influence the network’s classification behavior, highlighting the central role of Category Memory within the SBMN framework.

6. Conclusions

To improve the classification and diagnosis of vertical root fractures (VRFs), we propose a novel Similarity-Based Memory Network (SBMN) that leverages Category Memory and the Basic SBMN Module to enhance feature representation and classification. The network stores representative features for each class in a structured memory, while the Basic SBMN Module regulates the interaction between input features and memory, ensuring stability and effective feature enhancement. Experiments on a small-scale CBCT dataset indicate that SBMNs achieve improved classification accuracy and feature discriminability compared with conventional networks. Channel-level analyses show that intra-class features align closely with their corresponding memory, whereas inter-class features are decorrelated, demonstrating that the network effectively captures class-specific information. Memory manipulation experiments further confirm the critical role of Category Memory, as classification performance depends on the integrity and specificity of the stored memories. Overall, SBMNs provide a robust framework for small-sample medical image analysis, delivering reliable diagnostic performance, enhanced interpretability, and consistent class-specific feature representation. Incorporating structured memory with the Basic SBMN Module offers an effective strategy for automated diagnosis in scenarios with limited data. However, given the current single-center dataset and lack of extensive multi-center validation, these findings primarily demonstrate methodological potential rather than immediate clinical readiness. Future studies with diverse scanners and populations are needed to confirm clinical applicability.

Author Contributions

Conceptualization, Y.C.; Methodology, J.W., Y.F.Z. and X.Y.J.; Software, Y.F.Z.; Validation, J.W. and Y.F.Z.; Formal analysis, Y.F.Z. and Z.T.L.; Investigation, J.W. and Y.F.Z.; Resources, Z.T.L.; Data curation, Z.T.L. and Y.F.Z.; Writing—original draft preparation, J.W.; Writing—review and editing, J.W. and Y.C.; Visualization, J.W.; Supervision, Y.C. and J.Y.; Project administration, Y.C.; Funding acquisition, Y.C. All authors have read and agreed to the published version of the manuscript.

Funding

This work was supported by the Nanjing University Integrated Research Platform of the Ministry of Education—Top Talents Program under Grant 0210-14380255.

Institutional Review Board Statement

The study was conducted in accordance with the Declaration of Helsinki and approved by the Ethics Committee of Nanjing Stomatological Hospital, Medical School of Nanjing University (protocol code Ethics2018NL044, approval date 16 May 2018).

Informed Consent Statement

Patient consent was waived by the Ethics Committee due to the retrospective nature of the study.

Data Availability Statement

The data that support the findings of this study are not publicly available due to privacy and ethical restrictions. Data may be available from the corresponding author upon reasonable request.

Conflicts of Interest

The authors declare no competing interests.

References

  1. Rivera, E.M.; Walton, R.E. Longitudinal tooth fractures: Findings that contribute to complex endodontic diagnoses. Endod. Top. 2010, 16, 82–111. [Google Scholar] [CrossRef]
  2. Huang, C.C.; Lee, B.S. Diagnosis of vertical root fracture in endodontically treated teeth using computed tomography. J. Dent. Sci. 2015, 10, 227–232. [Google Scholar] [CrossRef][Green Version]
  3. Diagnostic issues dealing with the management of teeth with vertical root fractures: A narrative review. G. Ital. Endod. 2014, 28, 91–96. [CrossRef][Green Version]
  4. Alsani, A.; Balhaddad, A.A.; Nazir, M.A. Vertical root fracture: A case report and review of the literature. G. Ital. Endod. 2017, 31, 21–28. [Google Scholar] [CrossRef]
  5. Duak, K.; Kundabala, M.; Bhat, K. Endodontic Miscellany: 1. An unusual vertical root fracture. Endodontology 2004, 16, 23–26. [Google Scholar]
  6. Silva, E.J.N.L.D.; Santos, G.R.D.; Krebs, R.L.; Coutinho-Filho, T.D.S. Surgical Alternative for Treatment of Vertical Root fracture: A Case Report. Iran. Endod. J. 2012, 7, 40. [Google Scholar]
  7. Hsiao, L.T.; Ho, J.C.; Huang, C.F.; Hung, W.C.; Chang, C.W. Analysis of clinical associated factors of vertical root fracture cases found in endodontic surgery. J. Dent. Sci. 2020, 15, 200–206. [Google Scholar] [CrossRef] [PubMed]
  8. Aviad, T.; Tamse, Z.; Fuss, J.; Lustig, J.; Joseph, J. An evaluation of endodontically treated vertically fractured teeth. J. Endod. 1999, 25, 506–508. [Google Scholar] [CrossRef] [PubMed]
  9. Okaguchi, M.; Kuo, T.; Ho, Y.C. Successful treatment of vertical root fracture through intentional replantation and root fragment bonding with 4-META/MMA-TBB resin. J. Formos. Med. Assoc. 2018, 118, 671–678. [Google Scholar] [CrossRef]
  10. Wang, P.; Su, L. Clinical observation in 2 representative cases of vertical root fracture in nonendodontically treated teeth. Oral Surg. Oral Med. Oral Pathol. Oral Radiol. Endod. 2009, 107, e39–e42. [Google Scholar] [CrossRef]
  11. Liao, W.C.; Chen, C.H.; Pan, Y.H.; Chang, M.C.; Jeng, J.H. Vertical Root Fracture in Non-Endodontically and Endodontically Treated Teeth: Current Understanding and Future Challenge. J. Pers. Med. 2021, 11, 1375. [Google Scholar] [CrossRef]
  12. Hasan, S.; Singh, K.; Salati, N. Cracked tooth syndrome: Overview of literature. Int. J. Appl. Basic Med. Res. 2015, 5, 164–168. [Google Scholar] [CrossRef]
  13. Taschieri, S.; Tamse, A.; Fabbro, M.D.; Rosano, G.; Tsesis, I. A new surgical technique for preservation of endodontically treated teeth with coronally located vertical root fractures: A prospective case series. Oral Surg. Oral Med. Oral Pathol. Oral Radiol. Endod. 2010, 110, 45–52. [Google Scholar] [CrossRef]
  14. Kimura, Y.; Tanabe, M.; Amano, Y.; Kinoshita, J.I.; Yamada, Y.; Masuda, Y. Basic study of the use of laser on detection of vertical root fracture. J. Dent. 2009, 37, 909–912. [Google Scholar] [CrossRef]
  15. Pitts, D.L.; Natkin, E. Diagnosis and treatment of vertical root fractures. J. Endod. 1983, 9, 338–346. [Google Scholar] [CrossRef] [PubMed]
  16. Corbella, S.; Del Fabbro, M.; Tamse, A.; Rosen, E.; Tsesis, I.; Taschieri, S. Cone beam computed tomography for the diagnosis of vertical root fractures: A systematic review of the literature and meta-analysis. Oral Surg. Oral Med. Oral Pathol. Oral Radiol. 2014, 118, 593–602. [Google Scholar] [CrossRef] [PubMed]
  17. Ferreira, R.I.; Bahrami, G.; Isidor, F.; Wenzel, A.; Haiter-Neto, F.; Groppo, F.C. Detection of vertical root fractures by cone-beam computerized tomography in endodontically treated teeth with fiber-resin and titanium posts: An in vitro study. Oral Surg. Oral Med. Oral Pathol. Oral Radiol. 2013, 115, E49–E57. [Google Scholar] [CrossRef]
  18. Kajan, Z.D.; Taromsari, M. Value of cone beam CT in detection of dental root fractures. Dentomaxillofac. Radiol. 2012, 41, 3–10. [Google Scholar] [CrossRef] [PubMed]
  19. Makeeva, I.M.; Byakova, S.F.; Novozhilova, N.E.; Adzhieva, E.K.; Golubeva, G.I.; Grachev, V.I.; Kasatkina, I.V. Detection of artificially induced vertical root fractures of different widths by CBCT in vitro and in vivo. Int. Endod. J. 2015, 48, 628–635. [Google Scholar] [CrossRef]
  20. Lecun, Y.; Boser, B.; Denker, J.; Henderson, D.; Howard, R.; Hubbard, W.; Jackel, L. Backpropagation Applied to Handwritten Zip Code Recognition. Neural Comput. 1989, 1, 541–551. [Google Scholar] [CrossRef]
  21. Krizhevsky, A.; Sutskever, I.; Hinton, G. ImageNet Classification with Deep Convolutional Neural Networks. In Proceedings of the NIPS; Curran Associates, Inc.: Red Hook, NY, USA, 2012. [Google Scholar] [CrossRef]
  22. He, K.; Zhang, X.; Ren, S.; Sun, J. Deep Residual Learning for Image Recognition. In Proceedings of the 2016 IEEE Conference on Computer Vision and Pattern Recognition (CVPR); IEEE: New York, NY, USA, 2016. [Google Scholar] [CrossRef]
  23. Vaswani, A.; Shazeer, N.; Parmar, N.; Uszkoreit, J.; Jones, L.; Gomez, A.N.; Kaiser, Ł.; Polosukhin, I. Attention is All You Need. In Proceedings of the 30st Conference on Neural Information Processing Systems (NIPS 2017); NeurIPS Proceedings: New York, NY, USA, 2017. [Google Scholar]
  24. Dosovitskiy, A.; Beyer, L.; Kolesnikov, A.; Weissenborn, D.; Zhai, X.; Unterthiner, T.; Dehghani, M.; Minderer, M.; Heigold, G.; Gelly, S. An image is worth 16x16 words: Transformers for image recognition at scale. arXiv 2020, arXiv:2010.11929. [Google Scholar]
  25. Liu, Z.; Lin, Y.; Cao, Y.; Hu, H.; Wei, Y.; Zhang, Z.; Lin, S.; Guo, B. Swin Transformer: Hierarchical Vision Transformer using Shifted Windows. arXiv 2021, arXiv:2103.14030. [Google Scholar] [CrossRef]
  26. Hubel, D.H.; Wiesel, T.N. Receptive fields, binocular interaction and functional architecture in the cat’s visual cortex. J. Physiol. 1962, 160, 106–154. [Google Scholar] [CrossRef]
  27. Fukushima, K. Neocognitron: A self-organizing neural network model for a mechanism of pattern recognition unaffected by shift in position. Biol. Cybern. 1980, 36, 193–202. [Google Scholar] [CrossRef] [PubMed]
  28. Lecun, Y.; Bottou, L. Gradient-based learning applied to document recognition. Proc. IEEE 1998, 86, 2278–2324. [Google Scholar] [CrossRef]
  29. Simonyan, K.; Zisserman, A. Very Deep Convolutional Networks for Large-Scale Image Recognition. Comput. Sci. 2014, abs/1409.1556. [Google Scholar]
  30. Szegedy, C.; Liu, W.; Jia, Y.; Sermanet, P.; Rabinovich, A. Going Deeper with Convolutions. In Proceedings of the 2015 IEEE Conference on Computer Vision and Pattern Recognition (CVPR); IEEE: New York, NY, USA, 2014. [Google Scholar] [CrossRef]
  31. Huang, G.; Liu, Z.; Laurens, V.D.M.; Weinberger, K.Q. Densely Connected Convolutional Networks. In Proceedings of the 2017 IEEE Conference on Computer Vision and Pattern Recognition (CVPR); IEEE: New York, NY, USA, 2016. [Google Scholar] [CrossRef]
  32. Tan, M.; Le, Q.V. EfficientNet: Rethinking Model Scaling for Convolutional Neural Networks. arXiv 2019, arXiv:1905.11946. [Google Scholar] [CrossRef]
  33. Zhao, X.; Wang, L.; Zhang, Y.; Han, X.; Deveci, M.; Parmar, M. A review of convolutional neural networks in computer vision. Artif. Intell. Rev. 2024, 57, 43. [Google Scholar] [CrossRef]
  34. Krichen, M. Convolutional Neural Networks: A Survey. Computers 2023, 12, 151. [Google Scholar] [CrossRef]
  35. Liu, Y.; Xue, J.; Li, D.; Zhang, W.; Kiang, C.T.; Xu, Z. Image recognition based on lightweight convolutional neural network. Image Vis. Comput. 2024, 146, 105037. [Google Scholar] [CrossRef]
  36. Liu, Z.; Mao, H.; Wu, C.Y.; Feichtenhofer, C.; Darrell, T.; Xie, S. A ConvNet for the 2020s. arXiv 2022, arXiv:2201.03545. [Google Scholar] [CrossRef]
  37. Devlin, J.; Chang, M.W.; Lee, K.; Toutanova, K. BERT: Pre-training of Deep Bidirectional Transformers for Language Understanding. arXiv 2019, arXiv:1810.04805. [Google Scholar] [CrossRef]
  38. Kitaev, N.; Kaiser, U.; Levskaya, A. Reformer: The Efficient Transformer. arXiv 2020, arXiv:2001.04451. [Google Scholar] [CrossRef]
  39. Wang, S.; Li, B.Z.; Khabsa, M.; Fang, H.; Ma, H. Linformer: Self-Attention with Linear Complexity. arXiv 2020, arXiv:2006.04768. [Google Scholar] [CrossRef]
  40. Zaheer, M.; Guruganesh, G.; Dubey, A.; Ainslie, J.; Alberti, C.; Ontanon, S.; Pham, P.; Ravula, A.; Wang, Q.; Yang, L. Big Bird: Transformers for Longer Sequences. arXiv 2020, arXiv:2007.14062. [Google Scholar] [CrossRef]
  41. Ainslie, J.; Ontanon, S.; Alberti, C.; Pham, P.; Ravula, A.; Sanghai, S. ETC: Encoding Long and Structured Data in Transformers. arXiv 2020, arXiv:2004.08483. [Google Scholar] [CrossRef]
  42. Chen, J.; Lu, Y.; Yu, Q.; Luo, X.; Zhou, Y. TransUNet: Transformers Make Strong Encoders for Medical Image Segmentation. arXiv 2021, arXiv:2102.04306. [Google Scholar] [CrossRef]
  43. Cheng, B.; Misra, I.; Schwing, A.G.; Kirillov, A.; Girdhar, R. Masked-attention Mask Transformer for Universal Image Segmentation. In Proceedings of the 2022 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR); IEEE: New York, NY, USA, 2022; pp. 1280–1289. [Google Scholar] [CrossRef]
  44. Fu, J.; Liu, J.; Tian, H.; Li, Y.; Bao, Y.; Fang, Z.; Lu, H. Dual Attention Network for Scene Segmentation. In Proceedings of the 2019 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR); IEEE: New York, NY, USA, 2020. [Google Scholar] [CrossRef]
  45. Huang, Z.; Wang, X.; Wei, Y.; Huang, L.; Huang, T.S. CCNet: Criss-Cross Attention for Semantic Segmentation. IEEE Trans. Pattern Anal. Mach. Intell. 2020, 45, 6896–6908. [Google Scholar] [CrossRef] [PubMed]
  46. Hu, H.; Gu, J.; Zhang, Z.; Dai, J.; Wei, Y. Relation Networks for Object Detection. arXiv 2017, arXiv:1711.11575. [Google Scholar] [CrossRef]
  47. Carion, N.; Massa, F.; Synnaeve, G.; Usunier, N.; Kirillov, A.; Zagoruyko, S. End-to-End Object Detection with Transformers. arXiv 2020, arXiv:2005.12872. [Google Scholar] [CrossRef]
  48. Zhu, X.; Su, W.; Lu, L.; Li, B.; Dai, J. Deformable DETR: Deformable Transformers for End-to-End Object Detection. arXiv 2020, arXiv:2010.04159. [Google Scholar] [CrossRef]
  49. Bello, I.; Zoph, B.; Le, Q.; Vaswani, A.; Shlens, J. Attention Augmented Convolutional Networks. In Proceedings of the 2019 IEEE/CVF International Conference on Computer Vision (ICCV); IEEE: New York, NY, USA, 2020. [Google Scholar] [CrossRef]
  50. Yu, Q.; Xia, Y.; Bai, Y.; Lu, Y.; Shen, W. Glance-and-Gaze Vision Transformer. arXiv 2021, arXiv:2106.02277. [Google Scholar] [CrossRef]
  51. Khan, A.; Rauf, Z.; Sohail, A.; Khan, A.R.; Asif, H.; Asif, A.; Farooq, U. A survey of the vision transformers and their CNN-transformer based variants. Artif. Intell. Rev. 2023, 56, 2917–2970. [Google Scholar] [CrossRef]
  52. Maurício, J.; Domingues, I.; Bernardino, J. Comparing Vision Transformers and Convolutional Neural Networks for Image Classification: A Literature Review. Appl. Sci. 2023, 13, 5521. [Google Scholar] [CrossRef]
  53. Wang, Y.; Deng, Y.; Zheng, Y.; Chattopadhyay, P.; Wang, L. Vision Transformers for Image Classification: A Comparative Survey. Technologies 2025, 13, 32. [Google Scholar] [CrossRef]
  54. Le, D.P.C.; Wang, D.; Le, V.T. A Comprehensive Survey of Recent Transformers in Image, Video and Diffusion Models. Comput. Mater. Contin. 2024, 80, 37–60. [Google Scholar] [CrossRef]
  55. Atkeson, C.G.; Schaal, S. Memory-based neural networks for robot learning. Neurocomputing 1995, 9, 243–269. [Google Scholar] [CrossRef]
  56. Das, S.; Giles, C.L.; Sun, G.Z. Learning Context-free Grammars: Capabilities and Limitations of a Recurrent Neural Network with an External Stack Memory. arXiv 1992. [Google Scholar]
  57. Graves, A. Generating Sequences with Recurrent Neural Networks. arXiv 2013, arXiv:1308.0850. [Google Scholar] [CrossRef]
  58. Hochreiter, S.; Schmidhuber, J. Long Short-Term Memory. Neural Comput. 1997, 9, 1735–1780. [Google Scholar] [CrossRef] [PubMed]
  59. Sukhbaatar, S.; Szlam, A.; Weston, J.; Fergus, R. End-To-End Memory Networks. arXiv 2015, arXiv:1503.08895. [Google Scholar] [CrossRef]
  60. Weston, J.; Chopra, S.; Bordes, A. Memory networks. arXiv 2014, arXiv:1305.4807. [Google Scholar] [CrossRef]
  61. Kumar, A.; Irsoy, O.; Ondruska, P.; Iyyer, M.; Bradbury, J.; Gulrajani, I.; Zhong, V.; Paulus, R.; Socher, R. Ask Me Anything: Dynamic Memory Networks for Natural Language Processing. J. Mach. Learn. Res. 2015, arXiv:1506.07285. [Google Scholar] [CrossRef]
  62. Oh, S.W.; Lee, J.Y.; Xu, N.; Kim, S.J. Video Object Segmentation using Space-Time Memory Networks. arXiv 2019, arXiv:1904.00607. [Google Scholar] [CrossRef]
  63. Seong, H.; Oh, S.W.; Lee, J.Y.; Lee, S.; Lee, S.; Kim, E. Hierarchical Memory Matching Network for Video Object Segmentation. arXiv 2021, arXiv:2109.11404. [Google Scholar] [CrossRef]
  64. Liang, S.; Shen, X.; Huang, J.; Hua, X.S. Video Object Segmentation with Dynamic Memory Networks and Adaptive Object Alignment. In Proceedings of the 2021 IEEE/CVF International Conference on Computer Vision (ICCV); IEEE: New York, NY, USA, 2021. [Google Scholar] [CrossRef]
  65. Vibashan, V.S.; Oza, P.; Sindagi, V.A.; Gupta, V.; Patel, V.M. MeGA-CDA: Memory Guided Attention for Category-Aware Unsupervised Domain Adaptive Object Detection. arXiv 2021, arXiv:2103.04224. [Google Scholar] [CrossRef]
  66. Chen, Y.; Wang, Y.; Pan, Y.; Yao, T.; Tian, X.; Mei, T. A Style and Semantic Memory Mechanism for Domain Generalization. arXiv 2021, arXiv:2112.07517. [Google Scholar] [CrossRef]
  67. Hou, J.; Zhang, Y.; Zhong, Q.; Xie, D.; Zhou, H. Divide-and-Assemble: Learning Block-wise Memory for Unsupervised Anomaly Detection. arXiv 2021, arXiv:2107.13118. [Google Scholar] [CrossRef]
  68. Gong, D.; Liu, L.; Le, V.; Saha, B.; Mansour, M.R.; Venkatesh, S.; Van den Hengel, A. Memorizing Normality to Detect Anomaly: Memory-Augmented Deep Autoencoder for Unsupervised Anomaly Detection. arXiv 2020, arXiv:1904.02639. [Google Scholar] [CrossRef]
  69. Park, H.; Noh, J.; Ham, B. Learning Memory-guided Normality for Anomaly Detection. arXiv 2020, arXiv:2003.13228. [Google Scholar] [CrossRef]
  70. Sandler, M.; Zhmoginov, A.; Vladymyrov, M.; Jackson, A. Fine-tuning Image Transformers using Learnable Memory. arXiv 2022, arXiv:2203.15243. [Google Scholar] [CrossRef]
  71. Wu, D.; Hu, Y.C.; Hsiao, T.Y.; Liu, H. Uniform Memory Retrieval with Larger Capacity for Modern Hopfield Models. arXiv 2024, arXiv:2404.03827. [Google Scholar] [CrossRef]
  72. Tran, C.D.; Pham, L.H.; Tran, N.N.; Ho, P.N.; Jeon, J.W. Dual Memory Networks Guided Reverse Distillation for Unsupervised Anomaly Detection; Lecture Notes in Computer Science; Springer: Berlin/Heidelberg, Germany, 2025; pp. 361–378. [Google Scholar] [CrossRef]
  73. Lu, H.; Guo, Z.; Zuo, W. Modulated Memory Network for Video Object Segmentation. Mathematics 2024, 12, 863. [Google Scholar] [CrossRef]
  74. McGrath, C.; Chau, C.W.R.; Molina, G.F. Monitoring oral health remotely: Ethical considerations when using AI among vulnerable populations. Front. Oral Health 2025, 6, 1587630. [Google Scholar] [CrossRef]
  75. Hu, Z.; Cao, D.; Hu, Y.; Wang, B.; Zhang, Y.; Tang, R.; Zhuang, J.; Gao, A.; Chen, Y.; Lin, Z. Diagnosis of in vivo vertical root fracture using deep learning on cone-beam CT images. BMC Oral Health 2022, 22, 382. [Google Scholar] [CrossRef]
  76. Wu, Z.; Xiong, Y.; Yu, S.X.; Lin, D. Unsupervised Feature Learning via Non-parametric Instance Discrimination. In Proceedings of the 2018 IEEE/CVF Conference on Computer Vision and Pattern Recognition; IEEE: New York, NY, USA, 2018. [Google Scholar] [CrossRef]
Figure 1. Overall framework of the Similarity-Based Memory Network (SBMN). The architecture consists of two main components: (a) an embedding network implemented with ResNet50, and (b) the SBMN module. Black arrows indicate the general data flow, while blue arrows represent the process in which the embedding features and the modified category memory are combined and fed into a similarity-based classifier to generate the final prediction.
Figure 1. Overall framework of the Similarity-Based Memory Network (SBMN). The architecture consists of two main components: (a) an embedding network implemented with ResNet50, and (b) the SBMN module. Black arrows indicate the general data flow, while blue arrows represent the process in which the embedding features and the modified category memory are combined and fed into a similarity-based classifier to generate the final prediction.
Diagnostics 16 00710 g001
Figure 2. Schematic illustration of the basic SBMN module. Each Basic SBMN Module has a four-head design and independent parameters. Only one module is illustrated for clarity.
Figure 2. Schematic illustration of the basic SBMN module. Each Basic SBMN Module has a four-head design and independent parameters. Only one module is illustrated for clarity.
Diagnostics 16 00710 g002
Figure 3. The c architecture of the similarity-based classifier. Different arrow styles and colors indicate the data flows: blue arrows represent the training data path, red dashed arrows indicate the testing data path, gray arrows denote shared inputs, and green arrows represent the output leading to the classifier.
Figure 3. The c architecture of the similarity-based classifier. Different arrow styles and colors indicate the data flows: blue arrows represent the training data path, red dashed arrows indicate the testing data path, gray arrows denote shared inputs, and green arrows represent the output leading to the classifier.
Diagnostics 16 00710 g003
Figure 4. Visualization of feature distributions under different network configurations. (a) non-VRF vs. non-VRF, (b) non-VRF vs. VRF, and (c) VRF vs. VRF. Three models are compared: Untrained ResNet50, Trained ResNet50, and the SBMN (based on the ResNet50 embedding network). Each point represents one feature channel, where the horizontal and vertical coordinates indicate the corresponding feature values from two images.
Figure 4. Visualization of feature distributions under different network configurations. (a) non-VRF vs. non-VRF, (b) non-VRF vs. VRF, and (c) VRF vs. VRF. Three models are compared: Untrained ResNet50, Trained ResNet50, and the SBMN (based on the ResNet50 embedding network). Each point represents one feature channel, where the horizontal and vertical coordinates indicate the corresponding feature values from two images.
Diagnostics 16 00710 g004
Figure 5. Visualization comparison between ResNet50 and the SBMN using Grad-CAM and guided backpropagation methods. Columns from left to right show: original CBCT images, ResNet50 heatmaps, ResNet50 Grad-CAM, ResNet50 grayscale guided backpropagation, SBMN heatmaps, SBMN Grad-CAM, and SBMN guided backpropagation. Rows (ac) correspond to VRF teeth, and rows (df) correspond to non-VRF teeth. Yellow boxes mark the fracture regions in VRF cases.
Figure 5. Visualization comparison between ResNet50 and the SBMN using Grad-CAM and guided backpropagation methods. Columns from left to right show: original CBCT images, ResNet50 heatmaps, ResNet50 Grad-CAM, ResNet50 grayscale guided backpropagation, SBMN heatmaps, SBMN Grad-CAM, and SBMN guided backpropagation. Rows (ac) correspond to VRF teeth, and rows (df) correspond to non-VRF teeth. Yellow boxes mark the fracture regions in VRF cases.
Diagnostics 16 00710 g005
Figure 6. Channel-wise scatter distributions between the image features ( x _ K ) and hidden Category Memory ( h i d _ K ) obtained from the encoder within the SBMN module for a randomly selected non-VRF CBCT image. Notations 1 and 2 indicate two different Basic SBMN modules. Panels (a,b) show intra-class and inter-class correlations within the same module, while panels (c,d) illustrate inter-module decorrelation.
Figure 6. Channel-wise scatter distributions between the image features ( x _ K ) and hidden Category Memory ( h i d _ K ) obtained from the encoder within the SBMN module for a randomly selected non-VRF CBCT image. Notations 1 and 2 indicate two different Basic SBMN modules. Panels (a,b) show intra-class and inter-class correlations within the same module, while panels (c,d) illustrate inter-module decorrelation.
Diagnostics 16 00710 g006
Table 1. Dataset distribution for different segmentation methods and categories (number of teeth).
Table 1. Dataset distribution for different segmentation methods and categories (number of teeth).
Segmentation CategoryTotal (Train/Test)
Auto SegmentationVRF555 (416/139)
Non-VRF563 (424/139)
Manual SegmentationVRF276 (208/68)
Non-VRF276 (208/68)
Table 2. Comparison of SBMN with different backbones using auto and manual segmentation.
Table 2. Comparison of SBMN with different backbones using auto and manual segmentation.
Auto SegmentationAccuracySensitivitySpecificityPPVF1 Score
SBMN (ResNet50)97.198.695.795.897.1
SBMN (DenseNet169)93.592.195.094.893.4
SBMN (VGG19)94.693.595.795.694.5
SBMN (Inception_v3)78.991.466.273.081.2
SBMN (ResNeXt)94.692.197.197.094.5
SBMN (ViT)92.493.791.291.292.4
Manual SegmentationAccuracySensitivitySpecificityPPVF1 Score
SBMN (ResNet50)99.1100.097.197.198.5
SBMN (DenseNet169)98.497.1100.0100.098.5
SBMN (VGG19)98.5100.097.197.198.6
SBMN (Inception_v3)80.180.979.479.780.3
SBMN (ResNeXt)97.1100.094.194.497.1
SBMN (ViT)99.7100.099.599.899.7
Table 3. Performance comparison of the SBMN and other representative CNN- and Transformer-based networks using auto and manual segmentation.
Table 3. Performance comparison of the SBMN and other representative CNN- and Transformer-based networks using auto and manual segmentation.
Auto
Segmentation
AccuracySensitivitySpecificityPPVF1 Score
ResNet5091.492.190.790.891.4
DenseNet16987.180.693.592.686.6
VGG1987.889.286.386.787.7
Inception_v387.184.989.288.786.8
ResNeXt92.492.892.192.192.5
SCCNN93.995.092.993.094.0
ViT-B90.690.191.190.990.6
Swin-T91.592.890.290.491.5
Mobile-ViT87.986.589.388.987.9
SBMN (ResNet50)97.198.695.795.897.1
Manual
Segmentation
AccuracySensitivitySpecificityPPVF1 Score
ResNet5097.897.098.598.597.8
DenseNet16996.394.198.598.596.2
VGG1994.992.797.096.994.8
Inception_v387.886.389.288.987.6
ResNeXt97.897.198.598.597.8
SCCNN98.098.396.297.897.2
ViT-B98.298.198.398.298.2
Swin-T98.698.498.298.398.3
Mobile-ViT97.496.498.298.197.3
SBMN (ResNet50)99.1100.097.197.198.5
Table 4. Evaluation metrics of different similarity calculation methods.
Table 4. Evaluation metrics of different similarity calculation methods.
Similarity
Calculation Method
AccuracySensitivitySpecificityPPVF1 Score
Dot Product97.198.695.795.897.1
Cosine Similarity97.097.698.296.497.0
Pearson Correlation96.896.597.196.796.6
Euclidean Distance94.794.994.695.195.0
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Wang, J.; Jin, X.Y.; Zhang, Y.F.; Yuan, J.; Lin, Z.T.; Chen, Y. SBMN: Similarity-Based Memory Network for the Diagnosis of Vertical Root Fracture in Dental Imaging. Diagnostics 2026, 16, 710. https://doi.org/10.3390/diagnostics16050710

AMA Style

Wang J, Jin XY, Zhang YF, Yuan J, Lin ZT, Chen Y. SBMN: Similarity-Based Memory Network for the Diagnosis of Vertical Root Fracture in Dental Imaging. Diagnostics. 2026; 16(5):710. https://doi.org/10.3390/diagnostics16050710

Chicago/Turabian Style

Wang, Jie, Xin Yan Jin, Yi Fan Zhang, Jie Yuan, Zi Tong Lin, and Ying Chen. 2026. "SBMN: Similarity-Based Memory Network for the Diagnosis of Vertical Root Fracture in Dental Imaging" Diagnostics 16, no. 5: 710. https://doi.org/10.3390/diagnostics16050710

APA Style

Wang, J., Jin, X. Y., Zhang, Y. F., Yuan, J., Lin, Z. T., & Chen, Y. (2026). SBMN: Similarity-Based Memory Network for the Diagnosis of Vertical Root Fracture in Dental Imaging. Diagnostics, 16(5), 710. https://doi.org/10.3390/diagnostics16050710

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop