Artificially Intelligent Readers: An Adaptive Framework for Original Handwritten Numerical Digits Recognition with OCR Methods
Abstract
:1. Introduction
- We create a new handwritten custom dataset from a single contributor to ensure overall consistency of style.
- We train and apply a convolutional neural network to MNIST data for validation and then extend the method with custom data.
- Furthermore, we provide insights into the effects of training the model with variations in the proportions of custom data with and without rotations
- We provide a framework for adaptive OCR applicable to handwritten documents from a single source or belonging to a particular style of writing.
- Finally, we provide a comprehensive discussion of how the adaptive framework could be implemented and potential improvements with alternate technological solutions.
2. Related Works
3. Problem Formulation, Datasets, and Preprocessing
3.1. Problem Formulation
3.2. Datasets
- MNIST: The MNIST is a benchmarking [44] of 60,000 train handwritten digits and a test set of 10,000 (10 class labels) with each example represented as an image of 28 × 28 gray-scale pixels. Images are normalized to fit in a 20 × 20 pixel box while maintaining their aspect ratio. Images are centered in a 28 × 28 image. One column for each of the pixels because there are 28 × 28 pixels, there are 784 columns, and +1 for the labels. Pixel values are between 0 and 255, where 0 means black and 255 means white.
- Custom: Custom dataset consists of 240 training digits, 40 testing digits, and 20 validation digits, ensuring that the model is trained on a diverse set of data and can generalize well to new data points.The Custom dataset is also made available on our Githubrepository [45].
3.3. Custom Dataset Collection
- Preparing the Custom Dataset: Our Custom dataset consists of 300 handwritten digits collected using the method described in the data collection section. Two users wrote these digits, and we divided the dataset into three parts, each containing 100 digits. We then performed a stratified split on each of the three datasets to create training and testing sets. Specifically, we divided each set into 80 training images and 20 testing images, ensuring each category had an equal number of labels. This stratified split helped to prevent any bias in the model. Next, we concatenated the training images, creating a set of 240 training digits, and concatenated the testing images, creating a set of 60 digits. We then performed a stratified fold on the testing images, splitting them into sets of 40 and 20. The 20 digits were used as a validation dataset.
3.4. Preprocessing the Custom Dataset
4. Materials and Methods
4.1. Major Phases of Performing an OCR
- Source data input—this phase involves using an external optical device, such as a scanner or camera, to generate images that contain relevant alphanumeric text or symbols.
- Data preparation—covers various preprocessing operations, such as noise removal, to enhance source image quality and alignment of the image with the required standardized input format.
- Partitioning—multiple characters in the image are split into individual items in this phase so that a recognition engine can process the partitioned characters appropriately.
- Feature extraction—the partitioned characters are further “processed to extract different features.” The characters are then ‘recognized’ based on their alignment with these features.
- Classification—based on partitioning and feature extraction, the target image’s features are mapped to several categories and classified to appropriate values.
- Options—characters within source files can be classified using multiple techniques for pattern recognition using fundamental and statistical approaches.
- Validation and improvement—upon the completion of initial classification exercises, it may be observed that the results are rarely perfect. This is especially true for handwritten content with multiple authors, multifaceted fonts, and intricate languages. Quantitative, conceptual, linguistic, and probabilistic remedial approaches and fine-tuning can be performed to improve the accuracy of OCR systems.
4.2. Architecture of CNN-Based DL Model
- In our model, we stack multiple Conv2D layers to extract increasingly complex features from an input image.
- Each Conv2D layer has parameters such as the number of filters, kernel size, padding, and activation function, which are tuned to optimize the performance of the network. The Conv2D layers are used to build feature maps from the data.
- The kernel size of the filters is usually (3 × 3), and we use padding to ensure that the filters fit the input image. We also use the ReLU activation function to introduce nonlinearity in our model. Max-Pooling is used to reduce dimensionality. In Max-Pooling, the output value is just the maximum input value in each patch (for example, the maximum pixel in a span of three pixels).
- The next step is to flatten out the output from the last pooling layer because the input of the fully-connected layer is not a 2D vector but a 1D vector. To flatten means to convert a 2D matrix into a 1D matrix.
- A fully-connected hidden layer is added to perform classification. The fully-connected layers combine the features extracted from the convolutional layers to create a model. Between fully-connected layers, dropout layers are added to remove specific neurons’ contributions to the following layer while leaving the rest intact. Dropout layers are applied to reduce overfitting.
- Finally, an activation function as softmax is used to classify the outputs as digits as 0, 1, 2 … 9.
5. Experiments and Results
5.1. Experiment 1
5.2. Experiment 2
5.3. Experiment 3
5.4. Experiment 4
5.5. Experiment 5
5.6. Experiment 6
5.7. Experimental Summary
6. Discussion and Limitations
7. Conclusions and Future Work
Author Contributions
Funding
Institutional Review Board Statement
Informed Consent Statement
Data Availability Statement
Conflicts of Interest
Abbreviations
AI | Artificial Intelligence |
CNN | Convolution Neural Network |
DL | Deep Learning |
ML | Machine Learning |
OCR | Optical Character Recognition |
LSTM | Long Short-Term Memory |
NLP | Natural Language processing |
RNN | Recurrent Neural Network |
NER | Named-Entity Recognition |
ReLU | Rectified Linear Unit |
LSTM-R | Long Short-Term Memory plus Relation-aware pointer network |
HMM | Hidden Markov Models |
DNN | Deep Neural Network |
References
- Samuel, J.; Kashyap, R.; Samuel, Y.; Pelaez, A. Adaptive cognitive fit: Artificial intelligence augmented management of information facets and representations. Int. J. Inf. Manag. 2022, 65, 102505. [Google Scholar] [CrossRef]
- Thorat, C.; Bhat, A.; Sawant, P.; Bartakke, I.; Shirsath, S. A detailed review on text extraction using optical character recognition. In ICT Analysis and Applications; Springer: Singapore, 2022; pp. 719–728. [Google Scholar]
- Singh, S. Optical character recognition techniques: A survey. J. Emerg. Trends Comput. Inf. Sci. 2013, 4, 2009–2015. [Google Scholar]
- Samuel, J. A call for proactive policies for informatics and artificial intelligence technologies. Scholars Strategy Network, 19 December 2021. [Google Scholar]
- Srivastava, N.; Singh, S. Netra: Smart Hand Gloves Comprises Obstacle Detection, Object Identification & OCR Text to Speech Converter for Blinds. In Proceedings of the 2018 5th IEEE Uttar Pradesh Section International Conference on Electrical, Electronics and Computer Engineering (UPCON), Gorakhpur, India, 2–4 November 2018; pp. 1–4. [Google Scholar]
- Januszewski, A.; Kujawski, J.; Buchalska-Sugajska, N. Benefits of and obstacles to RPA implementation in accounting firms. Procedia Comput. Sci. 2021, 192, 4672–4680. [Google Scholar] [CrossRef]
- Samuel, J.; Rahman, M.M.; Ali, G.M.N.; Samuel, Y.; Pelaez, A.; Chong, P.H.J.; Yakubov, M. Feeling Positive About Reopening? New Normal Scenarios From COVID-19 US Reopen Sentiment Analytics. IEEE Access 2020, 8, 142173–142190. [Google Scholar] [CrossRef]
- Ali, G.M.N.; Rahman, M.M.; Hossain, M.A.; Rahman, M.S.; Paul, K.C.; Thill, J.C.; Samuel, J. Public perceptions of COVID-19 vaccines: Policy implications from US spatiotemporal sentiment analytics. Healthcare 2021, 9, 1110. [Google Scholar] [CrossRef]
- Manwatkar, P.M.; Yadav, S.H. Text recognition from images. In Proceedings of the 2015 International Conference on Innovations in Information, Embedded and Communication Systems (ICIIECS), Coimbatore, India, 19–20 March 2015; pp. 1–6. [Google Scholar]
- Srihari, S.N.; Shekhawat, A.; Lam, S.W. Optical character recognition (OCR). In Encyclopedia of Computer Science; Wiley: London, UK, 2003; pp. 1326–1333. [Google Scholar]
- Srihari, S.N.; Kuebert, E.J. Integration of hand-written address interpretation technology into the united states postal service remote computer reader system. In Proceedings of the Fourth International Conference on Document Analysis and Recognition, Ulm, Germany, 18–20 August 1997; Volume 2, pp. 892–896. [Google Scholar]
- Kornai, A. An experimental HMM-based postal ocr system. In Proceedings of the 1997 IEEE International Conference on Acoustics, Speech, and Signal Processing, Munich, Germany, 21–24 April 1997; Volume 4, pp. 3177–3180. [Google Scholar]
- Ifhaam, M.; Jayalal, S. Sinhala handwritten postal address recognition for postal sorting. In Proceedings of the 2019 International Research Conference on Smart Computing and Systems Engineering (SCSE), Colombo, Sri Lanka, 28 March 2019; pp. 134–141. [Google Scholar]
- Patel, M.; Reddy, S.L. An impact of grid based approach in offline handwritten Kannada word recognition. In Proceedings of the 2014 International Conference on Contemporary Computing and Informatics (IC3I), Mysore, India, 27–29 November 2014; pp. 630–633. [Google Scholar]
- Nagy, G. Disruptive developments in document recognition. Pattern Recognit. Lett. 2016, 79, 106–112. [Google Scholar] [CrossRef]
- Faizullah, S.; Ayub, M.S.; Hussain, S.; Khan, M.A. A Survey of OCR in Arabic Language: Applications, Techniques, and Challenges. Appl. Sci. 2023, 13, 4584. [Google Scholar] [CrossRef]
- Al-Hadhrami, A.A.; Allen, M.; Moffatt, C.; Jones, A.E. National characteristics and variation in Arabic handwriting. Forensic Sci. Int. 2015, 247, 89–96. [Google Scholar] [CrossRef]
- Bhagyasree, P.; James, A.; Saravanan, C. A proposed framework for recognition of handwritten cursive english characters using DAG-CNN. In Proceedings of the 2019 1st International Conference on Innovations in Information and Communication Technology (ICIICT), Chennai, India, 25–26 April 2019; pp. 1–4. [Google Scholar]
- Bhunia, A.K.; Khan, S.; Cholakkal, H.; Anwer, R.M.; Khan, F.S.; Shah, M. Handwriting transformers. In Proceedings of the IEEE/CVF International Conference on Computer Vision, Montreal, BC, Canada, 11–17 October 2021; pp. 1086–1094. [Google Scholar]
- Van Strien, D.; Beelen, K.; Ardanuy, M.C.; Hosseini, K.; McGillivray, B.; Colavizza, G. Assessing the impact of OCR quality on downstream NLP tasks. In Proceedings of the 12th International Conference on Agents and Artificial Intelligence, Valletta, Malta, 22–24 February 2020. [Google Scholar]
- Wang, J.; Tang, J.; Yang, M.; Bai, X.; Luo, J. Improving OCR-based image captioning by incorporating geometrical relationship. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Virtual, 19–25 June 2021; pp. 1306–1315. [Google Scholar]
- Nguyen, T.T.H.; Jatowt, A.; Coustaty, M.; Doucet, A. Survey of post-OCR processing approaches. ACM Comput. Surv. (CSUR) 2021, 54, 1–37. [Google Scholar] [CrossRef]
- Vijayarani, S.; Sakila, A. Performance comparison of OCR tools. Int. J. UbiComp (IJU) 2015, 6, 19–30. [Google Scholar]
- Suissa, O.; Elmalech, A.; Zhitomirsky-Geffet, M. Optimizing the Neural Network Training for OCR Error Correction of Historical Hebrew Texts; iConference 2020 Proceedings; iSchools Inc.: Grandville, MI, USA, 2020; pp. 1–10. [Google Scholar]
- Hegghammer, T. OCR with Tesseract, Amazon Textract, and Google Document AI: A benchmarking experiment. J. Comput. Soc. Sci. 2022, 5, 861–882. [Google Scholar] [CrossRef]
- Smith, R. An overview of the Tesseract OCR engine. In Proceedings of the Ninth International Conference on Document Analysis and Recognition (ICDAR 2007), Curitiba, Brazil, 23–26 September 2007; Volume 2, pp. 629–633. [Google Scholar]
- Ramiah, S.; Liong, T.Y.; Jayabalan, M. Detecting text based image with optical character recognition for English translation and speech using Android. In Proceedings of the 2015 IEEE Student Conference on Research and Development (SCOReD), Kuala Lumpur, Malaysia, 13–14 December 2015; pp. 272–277. [Google Scholar]
- Breuel, T.M. The OCRopus open source OCR system. Proc. Doc. Recognit. Retr. SPIE 2008, 6815, 120–134. [Google Scholar]
- Kittinaradorn, R. EasyOCR. 2020. Available online: https://github.com/JaidedAI/EasyOCR/tree/master (accessed on 11 April 2023).
- Awalgaonkar, N.; Bartakke, P.; Chaugule, R. Automatic license plate recognition system using ssd. In Proceedings of the 2021 International Symposium of Asian Control Association on Intelligent Robotics and Industrial Automation (IRIA), Goa, India, 20–22 September 2021; pp. 394–399. [Google Scholar]
- Troller, M. Practical OCR System Based on State of Art Neural Networks. Bachelor’s Thesis, Czech Technical University in Prague, Dejvice, Czech Republic, 2017. [Google Scholar]
- Alrasheed, N.; Prasanna, S.; Rowland, R.; Rao, P.; Grieco, V.; Wasserman, M. Evaluation of Deep Learning Techniques for Content Extraction in Spanish Colonial Notary Records. In Proceedings of the 3rd Workshop on Structuring and Understanding of Multimedia heritAge Contents, Virtual, 20 October 2021; pp. 23–30. [Google Scholar]
- Chen, Y.; Yang, J. Research on scene text recognition algorithm basedon improved CRNN. In Proceedings of the 2020 4th International Conference on Digital Signal Processing, Chengdu, China, 19–21 June 2020; pp. 107–111. [Google Scholar]
- Shi, B.; Bai, X.; Yao, C. An end-to-end trainable neural network for image-based sequence recognition and its application to scene text recognition. IEEE Trans. Pattern Anal. Mach. Intell. 2016, 39, 2298–2304. [Google Scholar] [CrossRef] [Green Version]
- Baek, Y.; Lee, B.; Han, D.; Yun, S.; Lee, H. Character region awareness for text detection. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Long Beach, CA, USA, 15–20 June 2019; pp. 9365–9374. [Google Scholar]
- Fenniak, M. PyPDF2·PyPI. Available online: https://pypi.org/project/PyPDF2/ (accessed on 4 April 2023).
- Kekare, A.; Jachak, A.; Gosavi, A.; Hanwate, P. Techniques for Detecting and Extracting Tabular Data from PDFs and Scanned Documents: A Survey. Tabula 2020, 7, 415–417. [Google Scholar]
- Mindee. docTR: Document Text Recognition. Available online: https://github.com/mindee/doctr (accessed on 4 April 2023).
- Batra, P.; Phalnikar, N.; Kurmi, D.; Tembhurne, J.; Sahare, P.; Diwan, T. OCR-MRD: Performance Analysis of Different Optical Character Recognition Engines for Medical Report Digitization. 2023. Available online: https://www.researchsquare.com/article/rs-2513255/v1 (accessed on 4 April 2023).
- Vaswani, A.; Shazeer, N.; Parmar, N.; Uszkoreit, J.; Jones, L.; Gomez, A.N.; Kaiser, L.; Polosukhin, I. Attention is All You Need. arXiv 2017, arXiv:1706.03762. [Google Scholar]
- Li, M.; Lv, T.; Cui, L.; Lu, Y.; Florencio, D.; Zhang, C.; Li, Z.; Wei, F. Trocr: Transformer-based optical character recognition with pre-trained models. arXiv 2021, arXiv::2109.10282. [Google Scholar]
- Smith, R.W. The Extraction and Recognition of Text from Multimedia Document Images. Ph.D. Thesis, University of Bristol, Bristol, UK, 1987. [Google Scholar]
- Ai, J. EasyOCR. Available online: https://github.com/JaidedAI/EasyOCR (accessed on 4 April 2023).
- Le Cun, Y.; Bottou, L.; Bengio, Y.; Haffner, P. Gradient-based learning applied to document recognition. Proc. IEEE 1998, 86, 2278–2324. [Google Scholar] [CrossRef] [Green Version]
- Lab, R. OCR-RUCILDigits-4. 2022. Available online: https://github.com/ay7n/OCR-RUCILDigits-4 (accessed on 4 April 2023).
- Eikvil, L. Optical Character Recognition. 1993. Available online: http://home.nr.no/~eikvil/OCR.pdf (accessed on 4 April 2023).
- Agrawal, A.K.; Shrivas, A.; kumar Awasthi, V. A Robust model for handwritten digit recognition using machine and deep learning technique. In Proceedings of the 2021 2nd International Conference for Emerging Technology (INCET), Belagavi, India, 21–23 May 2021; pp. 1–4. [Google Scholar]
- Kumar, V.; Recupero, D.R.; Riboni, D.; Helaoui, R. Ensembling Classical Machine Learning and Deep Learning Approaches for Morbidity Identification From Clinical Notes. IEEE Access 2021, 9, 7107–7126. [Google Scholar] [CrossRef]
- Cheng, L.; Bing, L.; He, R.; Yu, Q.; Zhang, Y.; Si, L. IAM: A Comprehensive and Large-Scale Dataset for Integrated Argument Mining Tasks. arXiv 2022, arXiv:2203.12257. [Google Scholar]
- Aico. TROCR Digit. Available online: https://huggingface.co/spaces/aico/TrOCR-digit (accessed on 4 April 2023).
- Karthick, K.; Ravindrakumar, K.; Francis, R.; Ilankannan, S. Steps involved in text recognition and recent research in OCR: A study. Int. J. Recent Technol. Eng. 2019, 8, 2277–3878. [Google Scholar]
- Yu, Y.; Si, X.; Hu, C.; Zhang, J. A review of recurrent neural networks: LSTM cells and network architectures. Neural Comput. 2019, 31, 1235–1270. [Google Scholar] [CrossRef] [PubMed]
- Williams, G.; Baxter, R.; He, H.; Hawkins, S.; Gu, L. A comparative study of RNN for outlier detection in data mining. In Proceedings of the 2002 IEEE International Conference on Data Mining, Maebashi City, Japan, 9–12 December 2002; pp. 709–712. [Google Scholar]
- Mueller-Gastell, J.; Sena, M.; Tan, C.Z. A Multi-Digit OCR System for Historical Records (Computer Vision). Available online: http://cs230.stanford.edu/projects_spring_2020/reports/38792124.pdf (accessed on 4 April 2023).
- Goodfellow, I.J.; Bulatov, Y.; Ibarz, J.; Arnoud, S.; Shet, V. Multi-digit number recognition from street view imagery using deep convolutional neural networks. arXiv 2013, arXiv:1312.6082. [Google Scholar]
- Samuel, J.; Palle, R.; Soares, E.C. Textual Data Distributions: Kullback Leibler Textual Distributions Contrasts on GPT-2 Generated Texts, with Supervised, Unsupervised Learning on Vaccine & Market Topics & Sentiment. arXiv 2022, arXiv:2107.02025. [Google Scholar]
- Rahman, M.M.; Ali, G.M.N.; Li, X.J.; Samuel, J.; Paul, K.C.; Chong, P.H.; Yakubov, M. Socioeconomic factors analysis for COVID-19 US reopening sentiment with Twitter and census data. Heliyon 2021, 7, e06200. [Google Scholar] [CrossRef]
- Samuel, J.; Ali, G.G.M.N.; Rahman, M.M.; Esawi, E.; Samuel, Y. COVID-19 public sentiment insights and machine learning for tweets classification. Information 2020, 11, 314. [Google Scholar] [CrossRef]
- Bhandari, A.; Kumar, V.; Thien Huong, P.T.; Thanh, D.N. Sentiment analysis of COVID-19 tweets: Leveraging stacked word embedding representation for identifying distinct classes within a sentiment. In Artificial Intelligence in Data and Big Data Processing, Proceedings of ICABDE 2021, Ho Chi Minh City, Vietnam, 18–19 December 2022; Springer: Berlin/Heidelberg, Germany, 2022; pp. 341–352. [Google Scholar]
- Kumar, V.; Reforgiato Recupero, D.; Helaoui, R.; Riboni, D. K-LM: Knowledge Augmenting in Language Models Within the Scholarly Domain. IEEE Access 2022, 10, 91802–91815. [Google Scholar] [CrossRef]
Sr No. | Name | Brief Description | Authors | Format |
---|---|---|---|---|
1. | Tesseract | Tesseract is performing well for high-resolution images | [42] | Machine text |
2. | TrOCR | It can produce very accurate results for machine and handwritten text and is based on Transformers | [41] | Handwritten and Machine text |
3. | KerasOCR | Keras-OCR is designed for images. It produces decent results when text is embedded in a picture and the fonts and colors are not coordinated | [31] | Machine text |
4. | DocTr | State-of-the-art performances on public document datasets, comparable with GoogleVision | [38] | Machine text |
5. | Easy OCR | EasyOCR is a lightweight model that gives accurate results with organized texts like pdf files, receipts, and bills. It supports 80+ languages | [43] | Machine text |
6. | PyPDF2 | Can be used only for PDF and is not accurate in some cases. | [36] | Machine text |
Item | Specification |
---|---|
CPU | Intel Core i3-7100 (-HT-MCP-) CPU @ 3.90 GHz |
GPU | NVIDIA T4 |
CUDA | Version 10.1 |
OS | Ubuntu 17.10 |
Python | Version 3.10 |
Tensorflow | Version 2.12.0 |
Layer (Type) | Output Shape | Number of Parameters |
---|---|---|
conv2d (Conv2D) | (None, 28, 28, 16) | 160 |
max_pooling2d (MaxPooling2D) | (None, 14, 14, 16) | 0 |
conv2d_1 (Conv2D) | (None, 14, 14, 32) | 4640 |
max_pooling2d_1 (MaxPooling2D) | (None, 7, 7, 32) | 0 |
conv2d_2 (Conv2D) | (None, 7, 7, 64) | 18,496 |
max_pooling2d_2 (MaxPooling2D) | (None, 3, 3, 64) | 0 |
conv2d_3 (Conv2D) | (None, 3, 3, 64) | 36,928 |
max_pooling2d_3 (MaxPooling2D) | (None, 1, 1, 64) | 0 |
flatten (Flatten) | (None, 64) | 0 |
dense (Dense) | (None, 64) | 4160 |
dropout (Dropout) | (None, 64) | 0 |
dense_1 (Dense) | (None, 64) | 4160 |
dropout_1 (Dropout) | (None, 64) | 0 |
dense_2 (Dense) | (None, 10) | 650 |
Model | Train Data | Validation Accuracy | Misclassified Digits | Test Accuracy |
---|---|---|---|---|
CNN | MNIST | 99.18% | 5 | 87.5% |
CNN | MNIST + 240 custom digits | 98.91% | 2 | 95% |
CNN | MNIST + 960 custom digits | 98.83% | 2 | 95% |
CNN | MNIST + 9840 custom digits | 98.96% | 2 | 95% |
TrOCR | (Pretrained model) | NA | 2 | 95% |
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content. |
© 2023 by the authors. Licensee MDPI, Basel, Switzerland. This article is an open access article distributed under the terms and conditions of the Creative Commons Attribution (CC BY) license (https://creativecommons.org/licenses/by/4.0/).
Share and Cite
Jain, P.H.; Kumar, V.; Samuel, J.; Singh, S.; Mannepalli, A.; Anderson, R. Artificially Intelligent Readers: An Adaptive Framework for Original Handwritten Numerical Digits Recognition with OCR Methods. Information 2023, 14, 305. https://doi.org/10.3390/info14060305
Jain PH, Kumar V, Samuel J, Singh S, Mannepalli A, Anderson R. Artificially Intelligent Readers: An Adaptive Framework for Original Handwritten Numerical Digits Recognition with OCR Methods. Information. 2023; 14(6):305. https://doi.org/10.3390/info14060305
Chicago/Turabian StyleJain, Parth Hasmukh, Vivek Kumar, Jim Samuel, Sushmita Singh, Abhinay Mannepalli, and Richard Anderson. 2023. "Artificially Intelligent Readers: An Adaptive Framework for Original Handwritten Numerical Digits Recognition with OCR Methods" Information 14, no. 6: 305. https://doi.org/10.3390/info14060305
APA StyleJain, P. H., Kumar, V., Samuel, J., Singh, S., Mannepalli, A., & Anderson, R. (2023). Artificially Intelligent Readers: An Adaptive Framework for Original Handwritten Numerical Digits Recognition with OCR Methods. Information, 14(6), 305. https://doi.org/10.3390/info14060305