Next Article in Journal
Active Damped PI Speed Loop Design for Motor Direct-Drive Operating Mechanism for High-Voltage Circuit Breakers
Previous Article in Journal
A Memory-Efficient Compensation Algorithm for Vertical Crosstalk in 8K LCD Panels
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Towards Robust Scene Text Recognition: A Dual Correction Mechanism with Deformable Alignment

1
School of Microelectronics, Tianjin University, Tianjin 300072, China
2
School of Electrical and Information Engineering, Tianjin University, Tianjin 300072, China
*
Authors to whom correspondence should be addressed.
Electronics 2025, 14(19), 3968; https://doi.org/10.3390/electronics14193968
Submission received: 9 September 2025 / Revised: 2 October 2025 / Accepted: 7 October 2025 / Published: 9 October 2025

Abstract

Scene Text Recognition (STR) faces significant challenges under complex degradation conditions, such as distortion, occlusion, and semantic ambiguity. Most existing methods rely heavily on language priors for correction, but effectively constructing language rules remains a complex problem. This paper addresses two key challenges: (1) The over-correction behavior of language models, particularly on semantically deficient input, can result in both recognition errors and loss of critical information. (2) Character misalignment in visual features, which affects recognition accuracy. To address these problems, we propose a Deformable-Alignment-based Dual Correction Mechanism (DADCM) for STR. Our method includes the following key components: (1) We propose a visually guided and language-assisted correction strategy. A dynamic confidence threshold is used to control the degree of language model intervention. (2) We designed a visual backbone network called SCRTNet. The net enhances key text regions through a channel attention module (SENet) and applies deformable convolution (DCNv4) in deep layers to better model distorted or curved text. (3) We propose a deformable alignment module (DAM). The module combines Gumbel-Softmax-based anchor sampling and geometry-aware self-attention to improve character alignment. Experiments on multiple benchmark datasets demonstrate the superiority of our approach. Especially on the Union14M-Benchmark, where the recognition accuracy surpasses previous methods by 1.1%, 1.6%, 3.0%, and 1.3% on the Curved, Multi-Oriented, Contextless, and General subsets, respectively.

1. Introduction

Text recognition is an important research area in computer vision and natural language processing. In particular, STR focuses on accurately extracting text information from images. This task plays a key role in many AI applications, such as automated document processing, license plate recognition, intelligent surveillance, and autonomous driving [1,2,3]. In recent years, with the rapid development of deep learning, STR has achieved remarkable progress across various fields [4,5,6].
However, in complex natural scenes, text images often present various challenging features. As shown in Figure 1, these include complex fonts, blurriness, occlusion, distortion, and lack of semantic content. These factors make scene text recognition difficult. As shown in Figure 2, existing STR methods can be divided into two main categories: visual feature learning [7,8,9,10,11,12] and semantic understanding [13,14,15,16,17,18,19,20].
Visual methods mainly use preprocessing to correct distortions and introduce some spatial variations to enhance the image feature extraction capability [9,13,21]. However, these methods often fail to accurately model semantic dependencies between characters when the text is blurry or structurally degraded.
On the other hand, semantic understanding methods [14,15,22,23,24] introduce semantic information implicitly or explicitly to address the shortcomings of visual models in the presence of blurry or incomplete visual inputs. These methods significantly improve the accuracy of scene text recognition.
Despite these advances, existing methods still have the following limitations: (1) Over-integration of language models makes it difficult to distinguish between visual and language signals, leading to poor interpretability. (2) Over-correction by language models introduces too much semantic information from model architecture or dataset corpus, which can cause the model to predict text that follows grammatical rules, even for non-semantic text, such as abbreviations or random combinations of letters, numbers, and symbols. (3) The lack of targeted correction in language models: some methods’ iterative corrections negatively impact inference speed.
To address the aforementioned challenges, we propose DADCM, a method designed from three key perspectives. The overall workflow of our approach is illustrated in Figure 3. Specifically, DADCM aims to achieve the following: (1) Visual feature extraction is more accurate, improving the visual model’s fine-grained ability to model complex text and enhancing its robustness to blurry and distorted characters. (2) Language correction becomes more intelligent, with visual information as the primary guide and the language model providing auxiliary correction, avoiding misjudgment of non-semantic text caused by excessive reliance on language priors. (3) Features and characters are better aligned to handle various forms of text images.
Therefore, the main contributions of this paper are as follows:
(1) We propose a novel visually guided and language-assisted framework that balances visual predictions and linguistic priors in a structured manner. This framework improves recognition robustness under noisy, ambiguous, or incomplete visual inputs, and its design is compatible with potential extensions to more complex STR systems.
(2) To enhance feature extraction, we designed SCRTNet, a lightweight backbone that integrates channel attention with deformable convolution (DCNv4) [25]. This task-driven combination strengthens feature discrimination in distorted, curved, or perspective-deformed text regions, allowing the model to capture complex text patterns more effectively.
(3) We also introduce the Dynamic Anchor-based Alignment Module (DAM), which combines Gumbel-Softmax-based dynamic anchor sampling, self-attention refinement, and Gaussian-weighted feature extraction. This module explicitly addresses character–feature misalignment and significantly improves recognition performance for irregular text shapes.
(4) Extensive experiments across multiple STR benchmarks demonstrate the practical effectiveness of our approach. Especially on the Union14M-Benchmark, recognition accuracy surpasses previous methods by 1.1%, 1.6%, 3.0%, and 1.3% on the Curved, Multi-Oriented, Contextless, and General subsets, respectively, highlighting both the conceptual novelty and empirical impact of our method.

2. Methods

The architecture of DADCM is shown in Figure 4. It consists of four main components: the Visual Feature Extraction Backbone Network, the Language Correction Module, the Character Inference Feature Module, and the Vision–Language Joint Optimization.

2.1. Visual Model

We propose a novel visual feature extraction backbone, named SCRTNet, which integrates local and global information. The detailed structure is shown in Figure 5. SCRTNet is built upon a hybrid architecture that combines ResNet [20,26] and Transformer-based modules [22,23,24,27,28], enabling more powerful and robust visual feature representation.
In the early stage of feature extraction, specifically after the second stage of ResNet, we introduce a channel attention module (Squeeze-and-Excitation Module, SENet) [29]. This module models the importance of each channel and reweights the responses accordingly. It enhances the responses related to text regions while suppressing redundant background information.
Furthermore, in the fourth stage of ResNet, we apply DCNv4 in the second convolutional layer of each residual block. Unlike standard convolutions, DCNv4 samples features with learnable offsets, allowing the receptive field to adapt to irregular or distorted text patterns. This significantly improves the network’s ability to handle common text variations, such as rotation, curvature, and perspective distortion in natural scenes.
The Transformer unit has two layers, each incorporating positional encoding, multi-head self-attention (MHSA), and a feed-forward network (FFN). This design enhances the model’s ability to capture global spatial structures.
By combining the above modules, we construct a visual feature extractor that integrates local precision and global context awareness. The detailed formulation of the extracted visual feature F v i , j is as follows:
F v i , j = T R ( X ) R B × P × C
where X R H × W × 3 denotes the input text image, H and W represent the height and width, respectively. R denotes the feature representation extracted by SCResNet-46, and T refers to the output of the Transformer encoder. B is the batch size, and C is the number of channels. The length of the resulting visual feature sequence is defined P = H 32 × W 8 . The final visual features F v i , j subsequently fed into the recognition module for inference.

2.2. Recognition and Inference Module

To address alignment challenges caused by text deformation, irregular layout, and uneven character spacing in complex scenes, we propose a Deformable Alignment Module (DAM). This module is integrated with parallel visual attention (PVA) [14,24] to enhance the recognition process. The overall structure is illustrated in Figure 4. DAM derives alignment-enhanced features from F v i , j by employing Gumbel-Softmax [30] differentiable sampling, offset refinement, and Gaussian kernel-based feature aggregation. The detailed process is illustrated in Figure 6.
First, at each spatial location of the feature map, we apply a three-layer 1 × 1 convolutional network to predict N d candidate anchor points x i , y i , along with their confidence scores p i . Then, we apply Gumbel-Softmax sampling to select the final anchor point in a differentiable manner. Next, we compute character-level geometric relations using a self-attention-based offset generation module, which dynamically predicts position shifts and yields the refined anchor locations Δ x , Δ y . Finally, we extract the aligned feature at x , y . Using a Gaussian kernel-weighted aggregation approach as follows:
( x * , y * ) = i = 1 N d p i · ( x i , y i )
Δ = Softmax Q K T d k V · W offset
( x , y ) = ( x * + Δ x , y * + Δ y )
G ( x , y ) = exp ( x x ) 2 + ( y y ) 2 2 σ 2
G t = G ( x , y ) · F v i , j
where Q, K, and V are obtained through linear projections of the feature map. d k is a scaling factor, and W o f f s e t is a learnable parameter matrix. The parameter σ adaptively controls the spatial extent of the weighting function. In addition, we obtain the attention map using the Parallel Visual Attention module as follows:
p t = Att t T F v i , j
Att = Softmax G F v i , j
G F v i , j = W 1 tanh W 2 O c + W 3 F v i , j
where A t t R h w × N denotes the attention map, h and w are the height and width of the feature map, respectively, and d is the feature dimension. N represents the maximum decoding length. O c R T × C is the position encoding of character sequences [31]. W 1 , W 2 , and W 3 learnable weights, and t refers to the current decoding time step.
Finally, we use the output of DAM as the Q and take the attention-enhanced features from PVA as the K and V. The final fusion is performed as follows:
F a = Softmax Q K T d V
The result is passed through a linear layer to produce the visual prediction probability distribution P v y i . Figure 7 shows the visual attention maps generated by the visual module for recognition.

2.3. Language Module

In this work, we decouple the visual and language models by introducing an independent language model as a correction module. In our architecture, the context characters y 1 : i 1 are directly derived from visual features. To eliminate potential bias during backpropagation, we apply Blocking Gradient Flow (BGF) [22,23] to isolate learning between different modalities. This ensures the independence of visual and linguistic representations. Within the correction module, the visual model provides predictions P ( y i | y 1 : i 1 ) for each character. The visual confidence reflects the reliability of character recognition. Low-confidence predictions may be caused by blur, occlusion, or noise, while high-confidence predictions are typically accurate. Based on these confidence scores, we generate a binary mask that determines whether a character should be corrected. If a character’s visual confidence falls below a predefined threshold, it is masked and refined by the language model using contextual information. The attention mechanism within the multi-head blocks is described as follows:
M y i = 0 if P ( y i ) τ otherwise
F mha = softmax Q K T d + M V
where Q R t × d denotes either the positional encoding of character sequences from the first layer or the output of the final layer. The K and V are derived from the visual character probabilities P ( y i ) . M R t × t is the attention mask used to prevent attention to low-confidence positions. By stacking multiple BCN layers, we obtain a deep architecture that models the bidirectional representation of text F l [22,23]. In this design, visual confidence is used to dynamically generate the attention mask. The language model performs correction only when necessary, serving as a flexible auxiliary module in scene text recognition.

2.4. Vision-Language Fusion

Scene text contains rich semantic information. Incorporating language models to guide and fuse with visual models can significantly improve recognition accuracy. However, in STR, a significant challenge of introducing language models is preventing the over-correction of originally correct character sequences, especially in cases with weak or no semantic context. To address this issue, we propose a confidence-based correction mechanism. Specifically, we combine outputs from both the visual and language models, and use a predefined confidence threshold τ to dynamically determine which characters should be corrected by the language model. In this way, only low-confidence predictions are refined by the language model. Furthermore, the fused features are used to assess the reliability of the correction.
To integrate features from both visual and linguistic modalities, we adopt a gating mechanism [14,16,22] to fuse the visual features with the language-corrected representations. The gating mechanism employs learnable weights to balance the contributions from each modality. The fusion process is defined as follows:
G = σ [ F a , F l ] W f
F f = G F a + ( 1 G ) F l
where W f is a learnable parameter. F a , F l , and F f represent the visual features, language-corrected features, and fused features, respectively. σ denotes the sigmoid activation function. G R t × d is the gated feature, dynamically selected from F a and F l .

2.5. Training Objective

The final objective function is defined as follows:
L = λ v L v + λ l L l + λ f L f
L * = 1 N t = 1 N log y t g t
L v = L main + λ L aux
L cos = 1 cos ( G t , F v i , j ) , if q = 1 max ( 0 , cos ( G t , F v i , j ) m ) , if q = 1
where L v , L l , and L f represent the losses from the visual module, the language module, and the final fusion module, respectively. y t and g t represent the prediction and ground truth. λ v , λ l , and λ f are balancing coefficients. Both L l and L f adopt the cross-entropy loss as defined in Equation (16). The visual loss L v follows a stage-wise hybrid training strategy. The auxiliary loss ( L a u x ) uses cosine embedding loss L cos to align the features extracted by PVA and DAM. The main visual loss L m a i n employs the cross-entropy loss from L * .

3. Experiment

3.1. Datasets

To enable a more comprehensive comparison with existing STR methods, we train our model on two synthetic datasets, SynthText (ST) [32] and SynthText90K (90K) [33], as well as a real-world training set, Union14M-L [19]. We evaluate DADCM on multiple benchmark datasets that cover a wide range of STR scenarios. These include (1) six commonly used STR benchmarks—ICDAR 2013 (IC13) [34], Street View Text (SVT) [35], IIIT5K-Words (IIIT5K) [36], ICDAR 2015 (IC15) [37], Street View Text-Perspective (SVTP) [38], and CUTE80 [39]. Specifically, we use the version of IC13 that contains 857 images. For IC15, the standard test set with 1811 images is used. (2) The real-world benchmark Union14M-L, which contains seven challenging subsets: Curved, Multi-Oriented (MO), Artistic, Contextless (Cless), Salient, Multi-Word (MW), and General. These subsets present more complex and diverse scene text conditions.

3.2. Implementation Details

The parameter settings used in our experiments are as follows: (1) The feature dimension d of the recognition model is set to 512. (2) The number of candidate points N d in DAM is set to 5. (3) The offset prediction module uses three layers, each with four self-attention heads. (4) The language model consists of 3 layers, with six attention heads in each layer. (5) The balance factors λ v , λ l , and λ f are set to 0.3, 0.2, and 0.5, respectively.
We resize all input images to 64 × 256, and apply the data augmentation strategy from CDisNet [28]. The model is trained to recognize 94 character classes, including 10 digits, 52 case-sensitive letters, 31 punctuation marks, and one special “END” token. We use case-insensitive word accuracy as the evaluation metric. All STR models, including our DADCM, are trained on the same datasets. The training is conducted in two stages using four NVIDIA RTX 4090 GPUs (NVIDIA, Santa Clara, CA, USA). The operating system was Ubuntu (Canonical, London, UK), the deep learning framework used was PyTorch (v2.2.0, Meta Platforms, Menlo Park, CA, USA), and CUDA (v11.8, NVIDIA, Santa Clara, CA, USA) was employed for GPU acceleration.
In the pretraining stage, we train the visual module separately. We use the AdamW optimizer [40] with a weight decay of 0.05 to pre-train on SynthText (ST) and SynthText90K (90 K) for 20 epochs. The batch size is set to 384, and the initial learning rate is 1.5 × 10 4 . We adopt a cosine learning rate scheduler with two epochs of linear warm-up. Additionally, during the first 6000 iterations, the DAM is frozen, and only the backbone network is trained to ensure stable feature extraction.
In the fine-tuning stage, we incorporate a pre-trained language model [22] into the training process. To keep consistency with the pretraining setup, we continue to use the AdamW optimizer with a weight decay of 0.01. The model is fine-tuned on the Union14M-L training set for 10 epochs.The initial learning rate is set to 1 × 10 4 , and we use a cosine scheduler without warm-up. The batch size is set to 256.
All experiments are conducted on large-scale datasets including ST, 90K, and Union14M-L. Due to the sufficient size of these datasets, training results are stable and not sensitive to random initialization. In addition, all baseline models were retrained under the same training settings to ensure a fair comparison, rather than directly adopting results from the literature.

3.3. Ablation Study on Visual Model

3.3.1. Effectiveness of SCRTNet

To evaluate the impact of different backbone architectures on recognition accuracy, we compare SCRTNet with ResNet+TFs, ConvNeXtV2 [41], and ViT [42] under the same experimental settings as pretraining. For a fair comparison, we disable both the language and alignment modules in all models. The inference process and the overall framework remain the same. We conduct two sets of experiments: one trained only on synthetic datasets and another trained only on Union14M-L. Word accuracy on the challenging IC15 and Union14M-L benchmark subsets is used as the evaluation metric.
Moreover, the results in Table 1 indicate the promising effectiveness of the proposed SCRTNet. We choose SCRTNet as the visual backbone for its balance between feature extraction capability and computational efficiency. The combination of channel attention and DCNv4 allows the network to better capture multi-scale and deformable text features, which is crucial for downstream DAM-based alignment and recognition. Compared to general-purpose encoders such as Swin Transformer, SCRTNet achieves competitive performance while being more lightweight and task-specific. Compared with the other three feature extraction backbones, SCRTNet achieves higher recognition accuracy on each benchmark dataset. When trained on synthetic data, real-world data, and a combination of both, SCRTNet achieves recognition accuracy improvements of 0.8%, 3.3%, and 2.2%, respectively, compared to using ResNet+TFs as the backbone. These results indicate that SCRTNet exhibits better adaptability in recognizing curved and multi-oriented text.
While our primary comparisons are with ResNet, ConvNeXtV2, and ViT, SCRTNet was selected for its superior trade-off between recognition accuracy and efficiency. Future work could include comparisons with more recent lightweight CNN-Transformer hybrids or Swin-based encoders.

3.3.2. Effectiveness of DAM

DAM is designed to address the challenges posed by the diversity of scene text data. Therefore, it is expected to bring improvements across different types of datasets. To ensure a fair comparison with SCRTNet, we apply the same training settings from Table 1 to the visual backbone enhanced with DAM, using both synthetic datasets and Union14M-L. As shown in Figure 8, the word-level average accuracy improves by 2.3% after integrating DAM. The overall trend of the results clearly demonstrates the effectiveness of DAM.
Interestingly, the gains from DAM vary across subsets. It provides larger improvements on curved and multi-oriented text because these cases suffer from significant character–feature misalignment, which DAM is specifically designed to correct. In contrast, for contextless text, the main challenge is the lack of linguistic cues rather than visual misalignment, while DAM still improves alignment, its effect is limited in low-context scenarios. These observations highlight that DAM is particularly effective for addressing spatial alignment issues in irregular text.

3.4. Vision-Language Fusion

3.4.1. Ablation Study on Vision–Language Fusion

The language model is introduced to complement and correct characters that are difficult to extract through visual features alone. The fusion between language and vision also plays a key role in model performance. To evaluate the effectiveness of our dynamic vision–language fusion, we conduct three comparison experiments: (1) visual-only inference, (2) language-only inference, and (3) fused inference. The results are shown in Table 2. To ensure consistency in evaluation, all three models undergo a two-stage training scheme, including pre-training and fine-tuning.
To further investigate the contribution of the Deformable Alignment Module (DAM), we conducted an ablation study by integrating it into the model with VM, LM, and VLM. As shown in Table 2, adding DAM improves recognition accuracy on the irregular text subsets, including IC15, Curve, Multi-Oriented, and Artistic, demonstrating its effectiveness in aligning visual features for challenging text shapes. The performance on Contextless and general text remains stable, indicating that DAM does not negatively affect simpler or context-free cases. Overall, the inclusion of DAM yields a slight increase in average accuracy, confirming its complementary role in enhancing the model’s robustness for complex scene text recognition.
As shown in Table 2, the Visual–Language Fusion Module (VLM) improves performance over using only the Vision Module, with gains of 2.9%, 2.5%, 5.6%, 3.2%, 2.0%, 3.6%, 0.8%, and 6.1% on IC15, Curve, Multi-Oriented, Artistic, Contextless, Salient, Multi-Words, and General datasets, respectively, and an average improvement of 3.3%. These results demonstrate that the model effectively integrates visual and language information. To better understand the contribution of the Language Module itself, we compare it with the Vision Module alone. The Language Module improves recognition accuracy on most test subsets, with gains of 1.7%, 2.3%, 3.5%, 2.7%, 0.4%, and 2.8% for IC15, Curve, Multi-Oriented, Artistic, Salient, and Multi-Words, respectively. However, a slight decrease of 1.2% is observed on the Contextless subset. This is likely because the Language Module may “over-correct” text that was already correctly recognized, relying on semantic context, while it effectively aids in correcting semantically meaningful text and complements the visual model on difficult cases, scene text often contains visually challenging content without semantic cues. Therefore, the application of the Language Module requires careful consideration.
Therefore, we should not rely solely on language correction, as some scene texts lack semantic content. On the other hand, visual predictions may also be unreliable. A robust vision–language fusion strategy is essential for handling complex text in real-world scenarios. In Table 2, compared to using the visual model alone, the fusion module brings further improvements of 1.2%, 0.2%, 2.1%, 0.5%, 3.2%, 1.5%, 0.4%, and 3.3% on each test subset.
Figure 9 shows the character-level confidence predictions from the visual, linguistic, and fusion modules in DADCM. As each module is added, the confidence scores for character predictions increase to varying degrees. This indicates that the modules complement each other and help the model better understand the text images. The fusion module, in particular, enhances the interaction between visual features and semantic information. As a result, the overall recognition becomes more stable and accurate. These results demonstrate the effectiveness of DADCM in handling complex scene text recognition tasks.

3.4.2. Sensitivity to Confidence Threshold

To further investigate the effect of the language correction confidence threshold τ , we conducted a sensitivity analysis on the Union14M-Benchmark sub-datasets, as shown in Table 3. The results indicate that recognition accuracy remains stable across a wide range of τ values, with fluctuations within ±0.2% for most sub-datasets. This demonstrates that the dynamic vision–language fusion mechanism consistently enhances performance without relying on a specific threshold.
Notably, the Contextless subset exhibits slightly lower accuracy and minimal variation across different τ values, which is consistent with the observation in Table 2 that the language module may over-correct text lacking semantic content. This suggests that the model successfully avoids excessive correction on contextless text through the confidence-based selection.
When τ takes moderate values near 0.5, the resulting performance is similar to that of the model employing both visual and language branches. Incorporating the complete visual-linguistic fusion module provides an additional gain in accuracy. These findings indicate that the proposed dynamic fusion strategy effectively leverages complementary visual and linguistic information, improving recognition accuracy for semantically meaningful and visually challenging text while mitigating potential errors on contextless text.

3.5. Comparison with SOTA

We evaluate DADCM on six commonly used STR benchmarks and the seven subsets of the Union14M-L benchmark. The results are compared with three different types of STR methods. Table 3 shows the accuracy comparison across all 13 test datasets. All methods are trained under the same settings as our model.
As shown in Table 4, the comparison results with various STR methods suggest that attention-based approaches and models incorporating language information show better generalization in challenging scenarios. This demonstrates the effectiveness of attention mechanisms and language models for complex scene text recognition.
In addition, existing STR methods have already achieved strong performance on the six common benchmarks. Therefore, improvements on these datasets are relatively small. Our method achieves a 0.4% average gain, with the highest improvement of 0.7% on the challenging IC15 dataset. On the Union14M-L benchmark, our method performs better, especially on subsets with irregular layouts, varying orientations, and limited semantic context. Specifically, accuracy improves by 0.7% on IC15, 1.1% on Curve, 1.6% on Multi-Oriented, and 3% on Contextless.
However, our method still faces limitations in scenarios involving multi-line or multi-word text. This may be attributed to the tendency of the offset predictions in the alignment module to align with dominant visual features. In addition, the model may focus more on character-level alignment within single words, sometimes treating multiple words as a single unit.
As shown in Figure 10, the prediction results of the DADCM model from the visual branch, the linguistic branch, and the visual–linguistic fusion branch are presented. By comparing the outputs of different modules, we observe that DADCM performs robustly and consistently in complex and challenging scene text recognition tasks.
For blurry, distorted, or irregularly shaped text images, the visual branch alone can produce accurate predictions. This suggests that the proposed visual modeling component has strong representational capacity in capturing visual cues. In cases where some characters are missing but the overall semantic content remains clear, the linguistic branch helps correct and complete the text based on context. For example, in the case of “Chang,” the language model successfully infers the correct result. Moreover, For contextless text images, such as proper names (“Evelyn”) or short words (“mei”), language-based correction can actually reduce recognition accuracy. In our experiments, these cases benefit from the hybrid model, which leverages visual alignment and attention mechanisms to achieve more accurate predictions than purely language-assisted methods.
However, some failure cases remain. When severe occlusion occurs, the model sometimes produces incorrect predictions, even though the output confidence scores are high across all modules. For instance, the prediction of “CAP” under occlusion could not be corrected. This indicates that the current model still has limited robustness in such conditions. Future work could focus on improving the model’s ability to reason over occluded regions. Enhancing semantic modeling and context understanding in these cases may help further improve the generalization performance of the system.

3.6. Analysis of Inference Speed

As shown in Table 5, we compare the average inference time and FLOPs of the proposed DADCM with several representative scene text recognition models on the seven subsets of Union14M. The proposed DADCM achieves an average inference time of 19.2 ms and requires 2.65 G FLOPs. Compared with MAERec, which requires 91.0 ms and 2.97 G FLOPs, as well as SRN, ABINet++, and BUSNet, DADCM provides faster or comparable inference while maintaining high recognition accuracy. Although slightly slower than CRNN, which runs at 6.5 milliseconds, and VisionLAN at 17.8 milliseconds, DADCM demonstrates superior robustness in handling complex scenes and challenging text samples. Overall, these results indicate that DADCM achieves a favorable balance between recognition accuracy and computational efficiency, making it well suited for practical applications.

4. Discussion

This paper presents a scene text recognition method based on a dual-alignment correction mechanism with deformable alignment (DADCM). The proposed approach integrates SCRTNet, the DAM, and feature fusion to address challenges such as irregular layouts, occlusion, and lack of semantic information.
Extensive experiments across multiple benchmark datasets demonstrate the effectiveness of our method. In particular, ablation studies highlight the importance of establishing robust visual–language fusion rules in improving recognition accuracy, confirming and extending findings from previous STR research.
Limitations and Future Work: Although DADCM achieves competitive performance on benchmarks such as ICDAR2015, Total-Text, and CTW1500, challenges remain under general, lexicon-free settings. Recognition stability can fluctuate in images with high content randomness, complex text layouts, or significant visual distractions. The model also shows limitations in feature preservation and sequence alignment for long character sequences.
Future research should focus on enhancing model generalization in complex environments, improving adaptability to multi-scale texts, and achieving a better trade-off between accuracy and inference efficiency. Exploring model compression or other optimization techniques may help improve inference speed while maintaining high accuracy.

5. Conclusions

In this work, we proposed DADCM, a scene text recognition method based on a dual-alignment correction mechanism with deformable alignment. By integrating SCRTNet, the DAM, and feature fusion, the method effectively addresses challenges such as irregular text layouts, occlusion, and limited semantic information. Extensive experiments on multiple benchmark datasets confirm the effectiveness of DADCM, and ablation studies highlight the importance of robust visual–language fusion for improving recognition accuracy. These findings provide insights for future development of more accurate and generalizable scene text recognition models.

Author Contributions

Conceptualization, Y.F. and C.L.; methodology, Y.F.; software, Y.F.; validation, Y.F. and C.L.; formal analysis, Y.F.; investigation, Y.F.; resources, Y.F.; data curation, Y.F.; writing—original draft preparation, Y.F.; writing—review and editing, C.L.; visualization, Y.F.; supervision, C.L.; project administration, Y.F.; funding acquisition, C.L. All authors have read and agreed to the published version of the manuscript.

Funding

This research was funded by the National Key Research and Development Program of China, grant number 2023YFF0905603.

Data Availability Statement

The datasets used in this study, including IIIT5K, SVT, IC13, IC15, SVTP, CUTE, are publicly available at their respective repositories. The additional data generated during the experiments, such as model predictions and intermediate feature maps, are available from the corresponding author on reasonable request.

Conflicts of Interest

The authors declare no conflicts of interest.

References

  1. Wan, Z.; Zhang, J.; Zhang, L.; Luo, J.; Yao, C. On Vocabulary Reliance in Scene Text Recognition. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Seattle, WA, USA, 14–19 June 2020; pp. 11425–11434. [Google Scholar]
  2. Long, S.; He, X.; Yao, C. Scene Text Detection and Recognition: The Deep Learning Era. Int. J. Comput. Vis. 2021, 129, 1–24. [Google Scholar] [CrossRef]
  3. Wang, Y.; Xie, H.; Zha, Z.; Tian, Y.; Fu, Z.; Zhang, Y. R-net: A Relationship Network for Efficient and Accurate Scene Text Detection. IEEE Trans. Multimed. 2021, 23, 1316–1329. [Google Scholar] [CrossRef]
  4. Zhang, C.; Tao, Y.; Du, K.; Ding, W.; Wang, B.; Liu, J.; Wang, W. Character-Level Street View Text Spotting Based on Deep Multisegmentation Network for Smarter Autonomous Driving. IEEE Trans. Artif. Intell. 2021, 3, 297–308. [Google Scholar] [CrossRef]
  5. Ouali, I.; Ben Halima, M.; Ali, W. Augmented Reality for Scene Text Recognition, Visualization and Reading to Assist Visually Impaired People. Procedia Comput. Sci. 2022, 207, 158–167. [Google Scholar] [CrossRef]
  6. Mei, Q.; Hu, Q.; Yang, C.; Zheng, H.; Hu, Z. Port Recommendation System for Alternative Container Port Destinations Using a Novel Neural Language-Based Algorithm. IEEE Access 2020, 8, 199970–199979. [Google Scholar] [CrossRef]
  7. Liu, W.; Chen, C.; Wong, K.-Y.K.; Su, Z.; Han, J. STAR-Net: A Spatial Attention Residue Network for Scene Text Recognition. In Proceedings of the British Machine Vision Conference, York, UK, 19–22 September 2016. [Google Scholar]
  8. Shi, B.; Wang, X.; Lyu, P.; Yao, C.; Bai, X. Robust Scene Text Recognition with Automatic Rectification. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Las Vegas, NV, USA, 27–30 June 2016; pp. 4168–4176. [Google Scholar]
  9. Zhan, F.; Lu, S. ESIR: End-to-End Scene Text Recognition via Iterative Image Rectification. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Salt Lake City, UT, USA, 18–22 June 2018; pp. 2059–2068. [Google Scholar]
  10. Luo, C.; Jin, L.; Sun, Z. MORAN: A Multi-Object Rectified Attention Network for Scene Text Recognition. Pattern Recognit. 2019, 90, 109–118. [Google Scholar] [CrossRef]
  11. Li, H.; Wang, P.; Shen, C.; Zhang, G. Show, Attend and Read: A Simple and Strong Baseline for Irregular Text Recognition. In Proceedings of the AAAI Conference on Artificial Intelligence, Honolulu, HI, USA, 27 January–1 February 2019; pp. 8610–8617. [Google Scholar]
  12. Yang, X.; He, D.; Zhou, Z.; Kifer, D.; Giles, C.L. Learning to Read Irregular Text with Attention Mechanisms. In Proceedings of the International Joint Conference on Artificial Intelligence, Melbourne, Australia, 19–25 August 2017; pp. 3280–3286. [Google Scholar]
  13. Shi, B.; Yang, M.; Wang, X.; Lyu, P.; Yao, C.; Bai, X. ASTER: An Attentional Scene Text Recognizer with Flexible Rectification. IEEE Trans. Pattern Anal. Mach. Intell. 2018, 41, 2035–2048. [Google Scholar] [CrossRef] [PubMed]
  14. Yu, D.; Li, X.; Zhang, C.; Liu, T.; Han, J.; Liu, J.; Ding, E. Towards Accurate Scene Text Recognition with Semantic Reasoning Networks. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Seattle, WA, USA, 13–19 June 2020; pp. 12113–12122. [Google Scholar]
  15. Zhao, Z.; Tang, J.; Lin, C.; Wu, B.; Huang, C.; Liu, H.; Tan, X.; Zhang, Z.; Xie, Y. Multi-Modal In-Context Learning Makes an Ego-Evolving Scene Text Recognizer. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Seattle, WA, USA, 16–22 June 2024; pp. 15567–15576. [Google Scholar]
  16. Yue, X.; Kuang, Z.; Lin, C.; Sun, H.; Zhang, W. RobustScanner: Dynamically Enhancing Positional Clues for Robust Text Recognition. In Proceedings of the European Conference on Computer Vision, Virtual, 23–28 August 2020; pp. 135–151. [Google Scholar]
  17. Baek, J.; Kim, G.; Lee, J.; Park, S.; Han, D.; Yun, S.; Seong, J.O.; Lee, H. What Is Wrong with Scene Text Recognition Model Comparisons? Dataset and Model Analysis. In Proceedings of the IEEE/CVF International Conference on Computer Vision, Seoul, Republic of Korea, 27 October–2 November 2019; pp. 4715–4723. [Google Scholar]
  18. Bautista, D.; Atienza, R. Scene Text Recognition with Permuted Autoregressive Sequence Models. In Proceedings of the European Conference on Computer Vision, Tel Aviv, Israel, 23–27 October 2022; pp. 178–196. [Google Scholar]
  19. Jiang, Q.; Wang, J.; Peng, D.; Liu, C.; Jin, L. Revisiting Scene Text Recognition: A Data Perspective. In Proceedings of the IEEE/CVF International Conference on Computer Vision, Paris, France, 2–6 October 2023; pp. 20543–20554. [Google Scholar]
  20. Qiao, Z.; Zhou, Y.; Yang, D.; Zhou, Y.; Wang, W. SEED: Semantics Enhanced Encoder-Decoder Framework for Scene Text Recognition. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Virtual, 14–19 June 2020; pp. 13528–13537. [Google Scholar]
  21. Shi, B.; Bai, X.; Yao, C. An End-to-End Trainable Neural Network for Image-Based Sequence Recognition and Its Application to Scene Text Recognition. IEEE Trans. Pattern Anal. Mach. Intell. 2016, 39, 2298–2304. [Google Scholar] [CrossRef] [PubMed]
  22. Fang, S.; Xie, H.; Wang, Y.; Mao, Z.; Zhang, Y. Read Like Humans: Autonomous, Bidirectional and Iterative Language Modeling for Scene Text Recognition. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Virtual, 19–25 June 2021; pp. 7098–7107. [Google Scholar]
  23. Fang, S.; Mao, Z.; Xie, H.; Wang, Y.; Yan, C.; Zhang, Y. ABINet++: Autonomous, Bidirectional and Iterative Language Modeling for Scene Text Spotting. IEEE Trans. Pattern Anal. Mach. Intell. 2023, 45, 7123–7141. [Google Scholar] [CrossRef] [PubMed]
  24. Wang, Y.; Xie, H.; Fang, S.; Wang, J.; Zhu, S.; Zhang, Y. From Two to One: A New Scene Text Recognizer with Visual Language Modeling Network. In Proceedings of the IEEE/CVF International Conference on Computer Vision, Montreal, QC, Canada, 10–17 October 2021; pp. 14194–14203. [Google Scholar]
  25. Xiong, Y.; Li, Z.; Chen, Y.; Wang, F.; Zhu, X.; Luo, J.; Wang, W.; Lu, T.; Li, H.; Qiao, Y.; et al. Efficient Deformable ConvNets: Rethinking Dynamic and Sparse Operator for Vision Applications. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Seattle, WA, USA, 16–22 June 2024; pp. 5652–5661. [Google Scholar]
  26. Wang, T.; Zhu, Y.; Jin, L.; Luo, C.; Chen, X.; Wu, Y.; Wang, Q.; Cai, M. Decoupled Attention Network for Text Recognition. In Proceedings of the AAAI Conference on Artificial Intelligence, New York, NY, USA, 7–12 February 2020; pp. 12216–12224. [Google Scholar]
  27. Na, B.; Kim, Y.; Park, S. Multi-Modal Text Recognition Networks: Interactive Enhancements between Visual and Semantic Features. In Proceedings of the European Conference on Computer Vision, Tel Aviv, Israel, 23–27 October 2022; pp. 446–463. [Google Scholar]
  28. Zheng, T.; Chen, Z.; Fang, S.; Xie, H.; Jiang, Y. CDistNet: Perceiving Multi-Domain Character Distance for Robust Text Recognition. Int. J. Comput. Vis. 2024, 132, 300–318. [Google Scholar] [CrossRef]
  29. Hu, J.; Li, S.; Sun, G. Squeeze-and-Excitation Network. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Salt Lake City, UT, USA, 18–22 June 2018; pp. 7132–7141. [Google Scholar]
  30. Jang, E.; Gu, S.; Poole, B. Categorical Reparameterization with Gumbel-Softmax. In Proceedings of the International Conference on Learning Representations, Toulon, France, 24–26 April 2017. [Google Scholar]
  31. Vaswani, A.; Shazeer, N.; Parmar, N.; Uszkoreit, J.; Jones, L.; N-Gomez, A.; Kaiser, Ł.; Polosukhin, I. Attention Is All You Need. In Proceedings of the Advances in Neural Information Processing Systems, Long Beach, CA, USA, 4–9 December 2017. [Google Scholar]
  32. Gupta, A.; Vedaldi, A.; Zisserman, A. Synthetic Data for Text Localisation in Natural Images. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Las Vegas, NV, USA, 26 June–1 July 2016; pp. 2315–2324. [Google Scholar]
  33. Jaderberg, M.; Simonyan, K.; Vedaldi, A.; Zisserman, A. Synthetic Data and Artificial Neural Networks for Natural Scene Text Recognition. In Proceedings of the Conference on Neural Information Processing Systems, Montreal, QC, Canada, 8–13 December 2014. [Google Scholar]
  34. Karatzas, D.; Shafait, F.; Uchida, S. ICDAR 2013 Robust Reading Competition. In Proceedings of the International Conference on Document Analysis and Recognition, Washington, DC, USA, 25–28 August 2013; pp. 1484–1493. [Google Scholar]
  35. Wang, K.; Babenko, B.; Belongie, S. End-to-End Scene Text Recognition. In Proceedings of the International Conference on Computer Vision, Barcelona, Spain, 6–13 November 2011; pp. 1457–1464. [Google Scholar]
  36. Mishra, A.; Karteek, A.; Jawahar, C.V. Scene Text Recognition Using Higher Order Language Priors. In Proceedings of the British Machine Vision Conference, Swansea, UK, 7–10 September 2015; pp. 1–11. [Google Scholar]
  37. Karatzas, D.; Gomez-Bigorda, L.; Nicolaou, A.; Ghosh, S.; Bagdanov, A.; Iwamura, M.; Matas, J.; Neumann, L.; Chandrasekhar, V.; Lu, S.; et al. ICDAR 2015 Competition on Robust Reading. In Proceedings of the International Conference on Document Analysis and Recognition, Tunis, Tunisia, 23–26 August 2015; pp. 1156–1160. [Google Scholar]
  38. Phan, T.-Q.; Shivakumara, P.; Tian, S.; Tan, C.-L. Recognizing Text with Perspective Distortion in Natural Scenes. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Portland, OR, USA, 23–27 June 2013; pp. 569–576. [Google Scholar]
  39. Anhar, R.; Palaiahnakote, S.; Chan, C.-S.; Tan, C.-L. A Robust Arbitrary Text Detection System for Natural Scene Images. Expert Syst. Appl. 2014, 41, 8027–8048. [Google Scholar] [CrossRef]
  40. Loshchilov, I.; Hutter, F. Fixing Weight Decay Regularization in Adam. In Proceedings of the International Conference on Learning Representations, Toulon, France, 24–26 April 2017. [Google Scholar]
  41. Yang, M.; Yang, B.; Liao, M.; Zhu, Y.; Bai, X. Class-Aware Mask-Guided Feature Refinement for Scene Text Recognition. Pattern Recognit. 2024, 149, 110244. [Google Scholar] [CrossRef]
  42. Atienza, R. Vision Transformer for Fast and Efficient Scene Text Recognition. In Proceedings of the International Conference on Document Analysis and Recognition, Lausanne, Switzerland, 5–10 September 2021; pp. 319–334. [Google Scholar]
  43. Lee, J.; Park, S.; Baek, J.; Seong, J.-O.; Kim, S.; Lee, H. On Recognizing Texts of Arbitrary Shapes with 2D Self-Attention. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Seattle, WA, USA, 14–19 June 2020; pp. 546–547. [Google Scholar]
  44. Wang, P.; Da, C.; Yao, C. Multi-Granularity Prediction for Scene Text Recognition. In Proceedings of the European Conference on Computer Vision, Tel Aviv, Israel, 23–27 October 2022; pp. 1484–1493. [Google Scholar]
  45. Cheng, C.; Wang, P.; Da, C.; Zheng, Q.; Yao, C. LISTER: Neighbor Decoding for Length-Insensitive Scene Text Recognition. In Proceedings of the International Conference on Computer Vision, Paris, France, 2–6 October 2023; pp. 19484–19494. [Google Scholar]
  46. Wei, J.; Zhan, H.; Lu, Y.; Tu, X.; Yin, B.; Liu, C.; Pal, U. Image as a Language: Revisiting Scene Text Recognition via Balanced, Unified and Synchronized Vision-Language Reasoning Network. In Proceedings of the AAAI Conference on Artificial Intelligence, Vancouver, BC, Canada, 20–27 February 2024; pp. 5885–5893. [Google Scholar]
Figure 1. Various types of scene text images.
Figure 1. Various types of scene text images.
Electronics 14 03968 g001
Figure 2. Different types of STR methods. From left to right: the pipelines of a pre-processing visual model, a model with separate visual and linguistic modules, and a vision–language fusion model.
Figure 2. Different types of STR methods. From left to right: the pipelines of a pre-processing visual model, a model with separate visual and linguistic modules, and a vision–language fusion model.
Electronics 14 03968 g002
Figure 3. DADCM. The pipeline of our DADCM, which integrates a deformable alignment module for bidirectional correction.
Figure 3. DADCM. The pipeline of our DADCM, which integrates a deformable alignment module for bidirectional correction.
Electronics 14 03968 g003
Figure 4. The architecture of the proposed DADCM. The Offset Prediction module consists of 1 × 1 convolutions, self-attention mechanism, and a fully connected (FC) layer.
Figure 4. The architecture of the proposed DADCM. The Offset Prediction module consists of 1 × 1 convolutions, self-attention mechanism, and a fully connected (FC) layer.
Electronics 14 03968 g004
Figure 5. Structure of the Visual Module. Each ResBlock contains Conv-BN-ReLU layers. For clarity, batch normalization and activation functions are omitted in the diagram.
Figure 5. Structure of the Visual Module. Each ResBlock contains Conv-BN-ReLU layers. For clarity, batch normalization and activation functions are omitted in the diagram.
Electronics 14 03968 g005
Figure 6. Architecture of the DAM.
Figure 6. Architecture of the DAM.
Electronics 14 03968 g006
Figure 7. The visual attention maps generated by the visual module for recognition.
Figure 7. The visual attention maps generated by the visual module for recognition.
Electronics 14 03968 g007
Figure 8. Effectiveness of DAM. Word accuracy comparison between models with and without DAM, trained using both synthetic datasets and Union14M-L10.
Figure 8. Effectiveness of DAM. Word accuracy comparison between models with and without DAM, trained using both synthetic datasets and Union14M-L10.
Electronics 14 03968 g008
Figure 9. Character prediction results from the visual module, linguistic module, and the final DADCM output (from left to right). Confidence scores are in the range [0, 1]. [E] denotes the end-of-sequence (“END”) token. Red indicates incorrect predictions; blue indicates corrected and accurate characters.
Figure 9. Character prediction results from the visual module, linguistic module, and the final DADCM output (from left to right). Confidence scores are in the range [0, 1]. [E] denotes the end-of-sequence (“END”) token. Red indicates incorrect predictions; blue indicates corrected and accurate characters.
Electronics 14 03968 g009
Figure 10. Prediction results of DADCM from the visual, linguistic, and fusion modules. From left to right and bottom to top, the characters represent: ground-truth labels (GT), visual (VM), linguistic (LM), and fusion predictions (FM). Red: Incorrect characters, Blue: Corrected and accurate characters.
Figure 10. Prediction results of DADCM from the visual, linguistic, and fusion modules. From left to right and bottom to top, the characters represent: ground-truth labels (GT), visual (VM), linguistic (LM), and fusion predictions (FM). Red: Incorrect characters, Blue: Corrected and accurate characters.
Electronics 14 03968 g010
Table 1. Ablation study of different feature extraction networks trained with synthetic datasets and Union14M-L. The Union14M-Benchmark is divided into seven subsets from left to right: Curve, Multi-Oriented, Artistic, Contextless, Salient, Multi-Words, and General. S denotes training with synthetic datasets ST and 901 K, while R denotes training with the Union14M-Train dataset. ViT-S uses six multi-head attention layers, and TF3 indicates the use of three Transformer blocks.
Table 1. Ablation study of different feature extraction networks trained with synthetic datasets and Union14M-L. The Union14M-Benchmark is divided into seven subsets from left to right: Curve, Multi-Oriented, Artistic, Contextless, Salient, Multi-Words, and General. S denotes training with synthetic datasets ST and 901 K, while R denotes training with the Union14M-Train dataset. ViT-S uses six multi-head attention layers, and TF3 indicates the use of three Transformer blocks.
EncoderTrain DataIC15CurveMulti-OrientedArtisticContextlessSalientMulti-WordsGeneralAvg
ResNetS70.267.159.650.870.259.861.260.562.4
ResNet+TFsS71.068.060.551.971.060.962.562.163.5
ConvNeXtV2S70.867.860.650.971.359.962.361.763.2
ViT-SS69.567.760.150.770.158.861.060.162.3
SCRTNet (Ours)S71.668.961.252.672.661.162.863.264.3
ResNetR82.380.768.760.181.370.176.071.273.8
ResNet+TFsR83.281.572.661.282.570.977.372.075.1
ConvNeXtV2R83.081.272.261.182.270.776.971.774.9
ViT-SR83.281.071.960.281.470.476.271.574.5
SCRTNet (Ours)R85.984.176.864.985.473.781.974.678.4
ResNetS+R83.682.273.161.783.373.277.873.576.1
ResNet+TFsS+R84.983.174.862.384.774.678.575.177.3
ConvNeXtV2S+R84.382.974.162.983.673.377.474.976.7
ViT-SS+R82.781.973.161.682.972.976.672.975.8
SCRTNet (Ours)S+R86.185.777.265.186.975.183.276.379.5
Table 2. Ablation study of the contributions of different modules (VM: Vision Module, LM: Language Module, VLM: Visual–Language Module, DAM/Deformable Alignment Module).
Table 2. Ablation study of the contributions of different modules (VM: Vision Module, LM: Language Module, VLM: Visual–Language Module, DAM/Deformable Alignment Module).
VMLMVLMDAMIC15CurveMulti-OrientedArtisticContextlessSalientMulti-WordsGeneralAvg
89.287.479.976.686.577.785.777.082.5
90.989.783.479.385.379.886.179.884.3
92.189.985.579.888.581.386.583.185.8
90.989.783.479.385.379.886.179.884.3
Table 3. Recognition accuracy under different language correction thresholds τ on Union14M-Benchmark sub-datasets, reflecting the effect of dynamic vision–language fusion.
Table 3. Recognition accuracy under different language correction thresholds τ on Union14M-Benchmark sub-datasets, reflecting the effect of dynamic vision–language fusion.
τ CurveMulti-OrientedArtisticContextlessSalientMulti-WordsGeneralAvg
0.189.585.279.785.179.786.079.784.0
0.289.685.379.885.279.886.179.884.2
0.389.785.479.985.279.986.279.984.3
0.489.785.579.985.379.986.279.984.4
0.589.885.580.085.380.086.380.084.5
0.689.885.580.085.380.086.380.084.5
0.789.785.479.985.279.986.279.984.4
0.889.685.379.885.279.886.179.884.3
0.989.585.279.785.179.786.079.784.2
Table 4. Word accuracy comparison with other STR methods on 13 benchmark datasets.
Table 4. Word accuracy comparison with other STR methods on 13 benchmark datasets.
IIIT SVT  IC13(857)  IC15 (1811) SVTP CUTECurved Multi-Oriented Artistic Contextless Salient Multi-Word General
TypeMethodCommon BenchmarksAvgUnion14MAvgAvgParam
/( × 10 6 )
CTCCRNN [21]90.883.892.871.870.581.281.829.312.634.344.216.835.660.333.355.78.3
AttentionSATRN [43]97.195.298.887.391.093.993.974.964.867.276.372.274.275.872.282.267.0
MGP-STR [44]97.297.998.091.493.098.195.985.980.873.676.178.472.884.478.986.7148.0
LISTER [45]98.197.598.689.694.097.295.878.465.674.782.973.484.184.877.786.151.1
MAERec [19]98.597.898.389.594.498.696.288.883.980.085.584.987.685.985.290.335.7
LMSRN [14]95.589.395.679.283.991.589.249.720.150.761.143.951.662.748.567.351.7
VisionLAN [24]96.391.596.183.685.492.490.970.957.256.763.867.647.574.262.675.632.8
ABINet++ [23]97.295.797.987.692.294.594.275.161.565.371.272.959.279.469.280.736.7
BUSNet [46]97.697.597.989.395.497.895.982.779.171.879.277.472.983.978.186.332.1
CdisNet [28]98.097.197.988.793.697.295.481.473.973.679.178.581.482.478.686.465.5
DADCM98.397.998.792.195.398.696.889.985.579.888.581.386.587.285.590.731.8
The bolded values indicate the best performance within each dataset. The underlined values indicate the second-best performance. The italicized text in the first row of the table represents the specific sub-datasets corresponding to Common Benchmarks and Union14M, respectively.
Table 5. Average Inference Time and Computational Cost of STR Methods on the Union14M Benchmark.
Table 5. Average Inference Time and Computational Cost of STR Methods on the Union14M Benchmark.
MethodCRNNMAERecSRNVisionLANABINet++BUSNetDADCM
Time (ms)6.591.019.117.829.619.819.2
FLOPs (G)0.692.974.302.733.052.672.65
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Feng, Y.; Li, C. Towards Robust Scene Text Recognition: A Dual Correction Mechanism with Deformable Alignment. Electronics 2025, 14, 3968. https://doi.org/10.3390/electronics14193968

AMA Style

Feng Y, Li C. Towards Robust Scene Text Recognition: A Dual Correction Mechanism with Deformable Alignment. Electronics. 2025; 14(19):3968. https://doi.org/10.3390/electronics14193968

Chicago/Turabian Style

Feng, Yajiao, and Changlu Li. 2025. "Towards Robust Scene Text Recognition: A Dual Correction Mechanism with Deformable Alignment" Electronics 14, no. 19: 3968. https://doi.org/10.3390/electronics14193968

APA Style

Feng, Y., & Li, C. (2025). Towards Robust Scene Text Recognition: A Dual Correction Mechanism with Deformable Alignment. Electronics, 14(19), 3968. https://doi.org/10.3390/electronics14193968

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop