Next Article in Journal
Label Metric for Multi-Class Multi-Target Tracking under Hierarchical Multilevel Classification
Next Article in Special Issue
Synthesis of High-Input Impedance Electronically Tunable Voltage-Mode Second-Order Low-Pass, Band-Pass, and High-Pass Filters Based on LT1228 Integrated Circuits
Previous Article in Journal
ECO6G: Energy and Cost Analysis for Network Slicing Deployment in Beyond 5G Networks
Previous Article in Special Issue
Anti-Blooming Clocking for Time-Delay Integration CCDs
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Multi-Currency Integrated Serial Number Recognition Model of Images Acquired by Banknote Counters

1
Department of Computer Science, Graduate School, Sangmyung University, Hongjimun 2-gil 20, Jongno-gu, Seoul 03016, Republic of Korea
2
Department of AI & Informatics, Graduate School, Sangmyung University, Hongjimun 2-gil 20, Jongno-gu, Seoul 03016, Republic of Korea
3
Division of Software Convergence, Sangmyung University, Hongjimun 2-gil 20, Jongno-gu, Seoul 03016, Republic of Korea
4
Department of Human-Centered Artificial Intelligence, Sangmyung University, Hongjimun 2-gil 20, Jongno-gu, Seoul 03016, Republic of Korea
*
Authors to whom correspondence should be addressed.
These authors contributed equally to this work.
Sensors 2022, 22(22), 8612; https://doi.org/10.3390/s22228612
Submission received: 4 October 2022 / Revised: 4 November 2022 / Accepted: 5 November 2022 / Published: 8 November 2022
(This article belongs to the Special Issue Feature Papers in Electronic Sensors)

Abstract

:
The objective of this study was to establish an automated system for the recognition of banknote serial numbers by developing a deep learning (DL)-based optical character recognition framework. An integrated serial number recognition model for the banknotes of four countries (South Korea (KRW), the United States (USD), India (INR), and Japan (JPY)) was developed. One-channel image data obtained from banknote counters were used in this study. The dataset used for the multi-currency integrated serial number recognition contains about 150,000 images. The class imbalance problem and model accuracy were improved through data augmentation based on geometric transforms that consider the range of errors that occur when a bill is inserted into the counter. In addition, by fine-tuning the recognition network, it was confirmed that the performance was improved when the serial numbers of the banknotes of four countries were recognized instead of the serial number of a banknote from each country from a single-currency dataset, and the generalization performance was improved by training the model to recognize the diverse serial numbers of multiple currencies. Therefore, the proposed method shows that real-time processing of less than 30 ms per image and character recognition with 99.99% accuracy are possible, even though there is a tradeoff between inference speed and serial number recognition accuracy when data augmentation based on the characteristics of banknote counters and a 1-stage object detector for banknote serial number recognition is used.

1. Introduction

Although electronic payment methods are increasing and the use of banknotes is on the decline worldwide, banknotes still play a significant role and are a major means of transactions and a secure store of wealth [1]. Although counting banknotes by hand is a common practice, to facilitate fast and accurate large-scale banknote transactions, the use of automated machines has become essential. These machines include banknote counters [2], coin counting machines [3], and automatic vending machines [4], in which money is inserted to purchase goods, as well as automated teller machines for deposits and withdrawals of banknotes. In the banking sector, devices such as banknote classifier machines, which are capable of completing large-scale transactions faster and more accurately than bank employees, have become essential for most banking transactions. Such devices perform complex functions such as banknote recognition, counterfeit status detection, and large-scale batch processing to satisfy the diverse requirements of banknote transactions. The major areas of banknote image analysis [5] include counterfeit banknote detection, serial number recognition, fitness classification, and banknote recognition. A banknote serial number, as shown in Figure 1, is engraved on each banknote in the process of production and is a unique alphanumerical identifier typically consisting of 9 to 11 digits [6]. When the unique serial number of a banknote is recognized and recorded, the source and circulation route of the banknote can be traced. This method is effective for the detection of counterfeit banknotes.
Banknote serial number recognition methods are based on optical character recognition (OCR) algorithms, which have been developed and applied in other areas, such as license plate recognition [7] and street number recognition based on Street View images [8]. In contrast to other fields adopting OCR, banknote serial number recognition requires a higher standard of accuracy because even small errors can result in large financial losses. Additionally, real-time detection is important for banknote service terminals, and the recognition process must satisfy transaction time requirements [9]. In general, the serial number recognition process for banknotes is divided into two steps. In the first step (preprocessing), the serial number region is extracted as the region of interest (ROI) containing the serial number from the banknote image captured by contact image sensors (CISs). In the extracted serial number region, segmentation is performed for each alphanumeric character. The segmentation process involves binarization and horizontal–vertical projection. In the second step (the character recognition process), feature extraction and classification are performed. In the feature extraction stage, the unique features of each alphanumeric character are extracted. These features include handcrafted features such as Gabor features and histogram of oriented gradients (HOG) features. In the classification part, each alphanumeric character is classified using the extracted features. The classifiers that can be used in this process include k-nearest neighbors, neural networks, and support vector machines (SVMs). Among them, SVMs are commonly used because they are less susceptible to overfitting and easier to use than neural networks. They also obtain a good real-time recognition rate for embedded systems installed in terminals. These algorithms can be utilized to obtain results with high accuracy for serial number recognition. In the ideal situation of banknotes with clean background patterns in the serial number region (Figure 2b), the recognition rate is excellent, but in practice, many banknotes do not have clean background patterns, as shown in Figure 2a. Banknotes with complex background patterns mainly include those with no clear boundary or outline (hindering character recognition), those with a small difference in pixel intensity between the background and characters, and those with similar styles for the background and serial number. Such notes cause difficulties in feature extraction. Additionally, the recognition accuracy is reduced by dirt, creases, and damage to the serial number region. However, with the development of artificial intelligence techniques, new methods have been introduced, and a deep learning (DL)-based approach has been developed to address the limitations of the existing algorithms and improve performance. In this study, the serial number region images of banknotes in multiple currencies (KRW, USD, INR, and JPY) are used as inputs for a convolutional neural network (CNN) to create an object detection model that localizes and classifies serial numbers for recognition. We use this approach to propose a multi-currency integrated serial number recognition model that is based on a DL 1-stage detector structure and is robust against various fonts and types of background noise. The contribution points of our proposed method are as follows.
  • The proposed method is a 1-stage method that simultaneously performs character region detection and classification instead of only character recognition.
  • The proposed method achieves state-of-the-art performance on both detection and recognition tasks because the 1-stage object detector is optimized for serial number recognition and the data are augmented according to the characteristics of banknote counters.
  • It achieves the highest serial number recognition performance compared with previous methods despite the increase in classes and the additional detection task.
  • The recognition model trained using a multi-currency dataset performed better than the recognition model trained on only single-currency datasets, indicating that the multi-currency dataset improved the generalization performance of the recognition model.
The remainder of this paper is organized as follows. Section 2 presents a literature review of serial number recognition. Section 3 describes the data collection and preprocessing steps and presents the proposed model. Section 4 presents the results obtained using the integrated serial number recognition model. In Section 5, the paper is summarized, and the conclusions are drawn.

2. Related Work

2.1. Handcrafted Feature Extraction Approach

Statistical recognition methods perform recognition using the feature vectors of each serial number character. Feature vectors are extracted from input images, the model is trained, and the probability distributions of the extracted feature vectors are obtained to separate the feature vector space. In these methods, the recognition results are affected by the definitions of the features of each serial number and the criteria for extracting feature vectors; thus, accurate definitions are crucial.
Zhao et al. [10] employed gray-level transformations from RGB images obtained using a charge-coupled device camera, binarization, and slope correction for preprocessing. Next, serial numbers were segmented using a vertical projection segmentation method. For the classifier, a genetic algorithm–backpropagation artificial neural network (GA-BP ANN) was used, and bilinear interpolation was used to address the problem of slow or no network convergence when samples of different sizes were used. The GA was used to set the node weights of the ANN. The accuracy of the results obtained using the GA-BP ANN was 95%, whereas that for a BP ANN was 82%. Ebrahimzadeh et al. [11] segmented the digits for recognition and extracted HOG features in the preprocessing step. Then, an SVM was employed using four different kernels—linear, polynomial, radial basis function (RBF), and sigmoid. Among these, the linear kernel exhibited the best performance, with an accuracy of 97.25%. This structure is efficient and useful for real-time application. Feng et al. [12] focused on RenMinBi (RMB) serial numbers. A variety of feature extraction methods use gradient direction features, Gabor features, and LeNet5 network-based features. For classifiers, an SVM with a linear discriminant function, quadratic discriminant function, and modified quadratic discriminant function was used. A multivariate normal distribution was assumed, and classification was performed using Bayes’ theorem. Using these methods, serial number recognition was performed. In another study by Feng et al. [13], points of interest were extracted using the difference of Gaussians to segment the serial number region. Then, the k-means clustering algorithm was used to remove redundant points of interest, and the key points were clustered to generate a set of local image parts for the serial number. A classification accuracy of 99.33% was achieved using SVM models with linear and RBF kernels.
In addition, other OCR algorithms, such as machine learning algorithms for scene text recognition and other methods, can be used to recognize banknote serial numbers. To recognize scene text in different languages, Tian et al. [14] extended the conventional HOG and proposed two new feature descriptors—the co-occurrence HOG (Co-HOG), which extracts spatial information by calculating the co-occurrence frequency of gradients for neighboring pixels, and the convolutional Co-HOG (ConvCo-HOG), which uses a convolutional strategy to extract Co-HOG features for all possible distributions of the co-occurrence structure. In addition, the dimensionality was reduced using principal component analysis, and scene text characters were recognized using a linear SVM. Zhou et al. proposed a method [15] for extracting characters from a serial number and removing background and noise using a hybrid binarization (HybridB) algorithm. This algorithm combines the binary algorithm based on the cumulative statistics of gray histograms, which performs projection on the horizontal and vertical character region of the serial number, and the binary algorithm based on neighborhood second expansion of gray ratios, which utilizes the four-way neighborhood information of each pixel. In addition, an adaptive character extraction algorithm was proposed to extract characters.

2.2. Deep Learning-Based Approach

DL methods employing CNNs have achieved good performance in various vision-based applications, and these methods exhibit considerable potential for serial number recognition. In DL, training is performed automatically using input data obtained using a learned feature extraction method rather than a handcrafted feature extraction method. Alwzwazy et al. [16] performed DL-based character recognition for Arabic digits, which is more difficult than the recognition of general alphanumeric patterns. The input images were the segmented Arabic digits, and because these were passed to the designed architecture, the images were preprocessed to make them 64 × 64 pixels in size. The DL architecture consisted of a convolutional layer, a max-pooling layer, and a fully connected layer. Classification was performed using the softmax activation function. This model achieved an accuracy of 95.7%. Boufenar et al. [17] performed DL-based character recognition with OAHCDB-40 [18] and AHCD, which are two datasets of segmented Arabic characters. The image size was 227 × 227 pixels, and the AlexNet architecture was used for model training. Three different training methods were applied—training from scratch, CNN fine-tuning, and the use of the CNN as a fixed feature extractor. Training from scratch achieved an accuracy of 100%. Zhao et al. [19] employed horizontal and vertical projections to determine the left- and right-side positions of each character as well as the window size of the character. Characters were extracted by a sliding window method and, using a simple CNN structure, a recognition accuracy of 99.99% and a recognition time of 5 ms were achieved using a DM642 chip. Wang et al. [20] performed preprocessing involving skew correction and the segmentation of each character on grayscale images of banknotes scanned using CISs. Image data with each extracted character were used, and the convolutional and pooling layers were replaced with dilated convolution in the CNN, which reduced its computational cost. In addition, the quantitative neural network-based method quantizes the weight parameters to an integer power of two, which can significantly reduce the learning time. In an experiment, the accuracy was increased to 99.89%, and a recognition time of less than 0.1 ms was achieved. Jang et al. [21] proposed a serial number recognition method for Indian banknotes, which have a unique font and complex background. A de-skewing process was performed on the images obtained for serial number recognition, the serial number region was cropped to extract each character, and a data shift was applied as data augmentation to prevent overfitting. The DL structure was composed of deep/shallow layers and heavy/light kernels; thus, four different models were designed: (1) deep/heavy, (2) deep/light, (3) shallow/heavy, and (4) shallow/light. With the shallow/light structure, an accuracy of 99.92% was obtained, and it was demonstrated that the DL character recognition method could recognize serial numbers in banknotes with complex backgrounds and characters with unusual fonts.
In addition, other OCR algorithms and related methods can be applied to banknote serial number recognition. Namysl et al. [22] introduced a lexicon-free OCR system. To recognize complex scene text, synthetic data were generated using a corpora and 2000 fonts. The DL model used the convolutional recurrent neural network structure, which is a hybrid method combining a CNN and long short-term memory (LSTM) and offers excellent feature extraction performance. The proposed OCR system achieved a higher accuracy than an open-source engine; it represents an application that combines a CNN and LSTM. Caldeira et al. [23] proposed an OCR system to extract the printed identification numbers of steel coils. The ROI of the coil identification number was extracted from an image obtained using a fixed camera. Then, to extract each character, the background was separated using four filters—the top hat, homomorphic, LogAbout, and Otsu filters. Each character was segmented by extracting the connected components from the background-separated binary images, and AlexNet, LeNet-5, and CIFAR models were used. Among these models, the LeNet-5 model yielded the best results, with an accuracy of 99.68%. Gang et al. [24] proposed a model for recognizing characters printed on components mounted on printed circuit boards. Data were collected automatically from inspection machines. Using the software used in the inspection process, characters were separated and labeled from the printed strings. In addition, data augmentation techniques, illumination, rotation, size, and noise were varied, and for the DL model, ResNet [25] and EfficientNet [26] were used to achieve a top-5 accuracy of 99.965%.

3. Materials and Methods

This section presents a detailed description of the development of the proposed model. Our method consists of the following four steps: data acquisition, preprocessing, DL model training, and multi-currency banknote serial number recognition. The proposed method is based on a single-shot multi-box detector (SSD), which is a 1-stage object detector [27]. The proposed method simultaneously performs detection and classification tasks, unlike existing methods that perform only character classification and do not detect character regions. The backbone of the proposed method is a CNN, which is a dense block-based network proposed by Huang et al. [28]. The proposed method is an SSD model in which the input size and backbone network are set according to the serial number recognition task, as shown in Figure 3. Experiments were conducted to determine the optimal SSD structure and hyperparameter settings for serial number recognition, and the process is described in Section 3.1 and Section 3.2.

3.1. Data Acquisition and Preprocessing

The data used in this study were acquired using a banknote counter [29] with a CIS (Canon Inc., Otaku, Japan). The CIS sensor-based banknote counter acquires approximately 16 frames per second, and the image resolution is 200 DPI. The acquired data are images of various types of banknotes from four countries (South Korea, the United States, India, and Japan). For Korean currency (KRW), denominations of 1000, 5000, 10,000, and 50,000 won were used. For US currency (USD), denominations of 1, 2, 5, 10, 20, 50, and 100 US dollars were used. For Indian currency (INR), denominations of 10, 20, 50, 100, 200, 500, and 2000 rupees were used. For Japanese currency (JPY), denominations of 1000, 2000, 5000, and 10,000 yen were used. The data acquired using the banknote counter were in a file in bag-of-features format with ROI information for serial number region. Using this file, the serial number regions were cropped and stored as 1-channel bitmap images (280 × 60). Because the same serial number can be found in two positions on each bill, the images were saved by assigning a separate ROI for each case. Although there are differences among the banknotes of different countries, the serial numbers are mainly located in the top left and bottom right corners. The serial number in the top left was marked as “first”, and the serial number in the bottom right was marked as “second”. Figure 4 shows an example of ROI selection.
An examination of the cropped images of the serial number regions of the banknotes revealed that, in general, the national currencies have the following characteristics. Korean and US banknotes have serial numbers with similar sizes. However, in Indian currency, different types of banknotes have serial numbers of different sizes. In addition, the serial numbers of US banknotes uniquely contain the character “☆”. In Japanese banknotes, the serial numbers are unclear because the pixel intensities of the background and serial numbers are similar. The images were manually labeled according to the bounding box for each character and the predefined class ID. Figure 5 shows an example of cropped images for each national currency.
When a note is inserted into the banknote counter, slight rotation and translation occur. Additionally, banknotes with frequent circulation are subjected to soiling and creases. To represent similar situations when performing data augmentation, the data were augmented 10-fold by applying mixed techniques of rotation, translation, and blur. The rotation is defined by Equation (1), and a range of –1° to +1° was applied with reference to the image center, considering that the CIS sensor acquires images aligned with the sensor’s image plane. Translation was specified in eight directions—up, down, left, right, and four diagonal directions—using Equation (2). For the padding, the pixel values at the edge were used for filling. Finally, Gaussian blur, expressed in Equation (3), was used [30]. Figure 6 shows an example of the augmentation applied to the banknote image data.
[ x y 1 ] = [ cos θ   sin θ   0 sin θ   cos θ   0       0                   0               1 ] [ x y 1 ]
[ x y 1 ] = [ 1   0   Δ x 0   1   Δ y 0   0       1 ] [ x y 1 ]
I ( x , y ; σ ) = 1 2 π σ 2 e x p ( x 2 + y 2 ) / 2 σ 2
The composition of the datasets was as follows. The first and second ROI images and labels were divided into training, validation, and test sets and then pooled to form a multi-currency dataset. For the KRW, USD, INR, and JPY banknotes, the dataset composition was as follows: 70% for the training set, 15% for the validation set, and 15% for the test set. Table 1 presents the number of banknote data for each currency.

3.2. Model Architecture

This section describes the model structure used in this study for multi-currency integrated serial number recognition and discusses the difference between the proposed and existing methods. In this study, four architectures were designed to identify the optimal architecture for multi-currency banknote serial number recognition. Figure 7 shows the structures of the four architectures. In the proposed model (Version 1), classification is performed using a high-level feature map with good feature extraction performance. In Version 2, pooling layers are added, which can reduce the training time; additionally, duplicate features are removed, and only important information is transferred. In Version 3, multiple classifier layers are added to allow the detection and recognition of small banknote serial numbers. Finally, in Version 4, DenseNet [29], which uses dense block and a transition structure, is employed. DenseNet has a structure that concatenates the feature maps of all layers. In this structure, features extracted from preceding layers can be continuously used, and the information flow is maximized. Additionally, to maximize feature extraction, the classifier layers use the extracted features from a high-level feature map.

4. Experimental Results

In this section, the process of developing the proposed model for multi-currency banknote serial number recognition is described. First, to identify the optimal model structure for this task, all four model structures illustrated in Figure 7 were tested, and for each currency, four different DL models were created to evaluate the recognition accuracy. The model structure exhibiting the highest accuracy was selected to develop a model for integrated recognition of banknote serial numbers in multiple currencies. For an accurate and fair comparison of model performance, a pre-trained model was not used, and the convolutional layer hyperparameters were set using He initialization [31]. All experiments were implemented in Python and Keras and were performed on a PC equipped with an Intel® Core (TM) i9-10900KF (3.70 GHz) CPU, an NVIDIA GeForce RTX2080 Ti GPU, and 64 GB of RAM. The output of the DL model includes the results of bounding-box detection and character classification. The errors were divided into detection errors and classification errors, and the sum of these errors was calculated to determine the total accuracy. A detection error refers to a failure in the detection of the bounding box. There are two cases of prediction failure. In the first case, the bounding box is detected, but the predicted location does not match the location of the ground truth but is elsewhere, leading to a small intersection over union [28]. In the second case, some parts of the serial number are not detected. Figure 8 illustrates examples of detection errors. The classification error refers to a failure in to predict the correct class. The accuracy values were calculated using Equations (4)–(6).
detecton   accuracy = number   of   correct   detections total   number   of   detections
recognition   accuracy = number   of   correct   recognitions total   number   of   recogntions
total   accuracy = number   of   correct   detections   and   recognitions total   number   of   data  
In the SSD model, multiple confidence values and bounding boxes are created for an object in the training stage. Among the multiple bounding boxes, only the optimal bounding box of the object should be selected, and all the other boxes should be removed. This operation is performed by the non-maximum suppression algorithm [32]. Table 2 presents the recognition accuracy (detection and classification) and model processing time of the Version 1, Version 2, Version 3, and Version 4 models. Evidently, the Version 4 model yields the highest accuracy. Experiments were conducted to evaluate the performance improvements of the model obtained by scale and aspect ratio adjustments and annotation mistake modifications. Table 3 presents the different scales and aspect ratios of the anchor boxes applied to the banknotes of multiple currencies. Table 4 and Figure 9 present the accuracy of the recognition model when these values were applied. The aspect ratio and anchor box hyperparameters were defined based on the printed characteristics of the banknote serial numbers. Setting the hyperparameters based on these characteristics considerably reduced the detection and classification errors.
Table 5 and Figure 10 compare the results of existing DL-based methods and the integrated recognition model for the banknotes of four currencies (KRW, USD, INR, and JPY). In previous studies, only classification was performed (not object detection) for the digits 0–9 and letters A–G. By contrast, in the proposed method, classification is performed over a total of 37 classes, including the digits 0–9, letters A–Z, and special character “☆” along with object detection. The comparison reveals that the proposed method is superior to the previously reported methods for banknote serial number recognition. Figure 11 presents a confusion matrix of the results of the multi-currency (KRW, USD, INR, and JPY) test set. The horizontal axis indicates the predicted label, and the vertical axis indicates the true label. Correct and incorrect predictions are shown in orange and red, respectively.

5. Discussion

In this study, an SSD-based multi-currency integrated serial number recognition model was developed as a 1-stage OCR model that considers both inference speed and OCR accuracy, unlike the 2-stage methods that detect character positions and classify characters. In contrast to these methods, the proposed method performs character detection and recognition for alphanumeric characters using an SSD-based 1-stage method. Despite the additional task of character recognition, the accuracy was confirmed to be state-of-the-art when compared with the accuracy of the existing methods. In addition, it is clear that the task assigned to the proposed method is more challenging than that of the existing methods—recognizing the banknotes of four countries using a single SSD model. When using our proposed multi-currency integrated serial number recognition model, a separate algorithm is not required to detect a character’s location, and the use of a large dataset for multiple countries improves the accuracy for the currency of a single country with an insufficient amount of data. Table 6 and Figure 12 compare the performance of models using only a single national currency dataset with the accuracy obtained when the model is trained on a large dataset containing four currencies. In the serial number recognition results for each national currency, it can be observed that using the data of multiple currencies is better than using only single-currency data. These results show that even if the font and size of the serial numbers differ among countries, the feature extraction is better when the model is trained with a large, integrated dataset. In addition, it was confirmed that the diversity and size of data were insufficient for training a model with high accuracy (approaching 100%) even when using a large dataset consisting of 40,000 or more samples for each country (see Table 1). In other words, the proposed method increased the diversity and number of data by integrating multiple currencies into one dataset, and a multi-currency integrated serial number recognition accuracy of 99.97% was obtained through model structure optimization and hyperparameters set according to the printing characteristics of the currencies. Finally, the proposed method is a serial number recognition model based on a fast SSD model with state-of-the-art character detection accuracy and can be applied to various tasks in the financial field, such as detecting counterfeit money with mismatched serial numbers or currency bill tracking.
In the future, reducing the inference time to enable high-speed processing in embedded systems in banknote counters could be a major research topic. To this end, various DL techniques and frameworks could be utilized, such as knowledge distillation, lightweight CNNs, the Open Neural Network Exchange, and TensorFlow Lite. In the case of knowledge distillation, the teacher–student structure can drive learning, such that the student model achieves teacher-level accuracy, even though the student model has fewer parameters and is faster than the teacher model. In the case of lightweight CNNs, depth-wise separable convolution or grouped convolution can be used to improve the inference speed by reducing the number of floating point operations. Furthermore, by establishing a 1-stage object detector-based integrated banknote serial number recognition system for various currencies, such as Arabic and Russian Cyrillic banknotes, a state-of-the-art multi-currency integrated serial number recognition system could be developed, even if the serial number contains characters that are not from the Latin alphabet.

6. Conclusions

In this study, a state-of-the-art multi-currency integrated serial number recognition method was developed to obtain superior accuracy and inference speed. To identify the optimal model structure for banknote serial number recognition, CNN models with four different structures were designed, and a data augmentation technique based on the insertion of notes into the banknote counter was used. A performance comparison was performed using a dataset containing multiple currencies (KRW, USD, INR, and JPY) to determine the optimal 1-stage object detector-based architecture for multi-currency integrated serial number recognition. The results confirmed that the best representation was extracted from the layers of the last block. Therefore, the best architecture was designed using a loss function, and the model output was obtained from the last block’s feature map. Using the optimal model, an integrated serial number recognition model for the banknotes of the four countries was designed. To determine the optimal hyperparameter values, experiments were performed using different scales and aspect ratios. The proposed method, using hyperparameters optimized for the printing characteristics of serial numbers, achieved state-of-the-art performance, with 99.97% accuracy and real-time serial number detection (within 30 ms) for multi-currency serial number recognition. These results indicate the importance of adjusting the parameters for the creation of anchor boxes during the training stage. It was confirmed that the proposed CNN-based banknote serial number recognition method could achieve accurate banknote serial number recognition even in the presence of complex backgrounds.

Author Contributions

Conceptualization, E.C.L. and K.L.; methodology, E.C.L., K.L., W.J. and C.L.; validation, K.L. and W.J.; investigation, D.S.J.; data curation, D.S.J.; writing—original draft preparation, K.L. and C.L.; writing—review and editing, E.C.L., K.L. and C.L.; visualization, W.J.; supervision, K.L. and E.C.L.; project administration, E.C.L.; funding acquisition, E.C.L. All authors have read and agreed to the published version of the manuscript.

Funding

This research received no external funding.

Data Availability Statement

Not applicable.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Kim, E.; Turton, T. The next generation banknote project. Reserve Bank Aust. Bull. 2014, 1–11. [Google Scholar]
  2. Wang, H.; Wang, N.; Jiang, Y.; Yang, J. Design of banknote withdrawal system of supermarket automatic change machine. In Proceedings of the 2018 8th International Conference on Applied Science, Engineering and Technology, Qingdao, China, 25–26 March 2018; Atlantis Press: Paris, France, 2018; pp. 130–134. [Google Scholar]
  3. Kavale, A.; Shukla, S.; Bramhe, P. Coin counting and sorting machine. In Proceedings of the 2019 9th International Conference on Emerging Trends in Engineering and Technology-Signal and Information Processing, Nagpur, India, 1–2 November 2019; pp. 1–4. [Google Scholar]
  4. Cardaci, R.; Burgassi, S.; Golinelli, D.; Nante, N.; Battaglia, M.A.; Bezzini, D.; Messina, G. Automatic vending-machines contamination: A pilot study. Glob. J. Health Sci. 2016, 9, 63. [Google Scholar]
  5. Lee, J.W.; Hong, H.G.; Kim, K.W.; Park, K.R. A survey on banknote recognition methods by various sensors. Sensors 2017, 17, 313. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  6. Bruna, A.; Farinella, G.M.; Guarnera, G.C.; Battiato, S. Forgery detection and value identification of Euro banknotes. Sensors 2013, 13, 2515–2529. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  7. Panahi, R.; Gholampour, I. Accurate detection and recognition of dirty vehicle plate numbers for high-speed applications. IEEE Trans. Intell. Transp. Syst. 2016, 18, 767–779. [Google Scholar] [CrossRef]
  8. Goodfellow, I.J.; Bulatov, Y.; Ibarz, J.; Arnoud, S.; Shet, V. Multi-digit number recognition from street view imagery using deep convolutional neural networks. arXiv 2013, arXiv:1312.6082. [Google Scholar]
  9. Tele, G.S.; Kathalkar, A.P.; Mahakalkar, S.; Sahoo, B.; Dhamane, V. Detection of fake Indian currency. Int. J. Adv. Res. Ideas Innov. Technol. 2019, 4, 170–176. [Google Scholar]
  10. Zhao, T.T.; Zhao, J.Y.; Zheng, R.R.; Zhang, L.L. Study on RMB number recognition based on genetic algorithm artificial neural network. In Proceedings of the 2010 3rd International Congress on Image and Signal Processing, Yantai, China, 16–18 October 2010; Volume 4, pp. 1951–1955. [Google Scholar]
  11. Ebrahimzadeh, R.; Jampour, M. Efficient handwritten digit recognition based on histogram of oriented gradients and SVM. Int. J. Comput. Appl. 2014, 104, 10–13. [Google Scholar]
  12. Feng, B.Y.; Ren, M.; Zhang, X.Y.; Suen, C.Y. Automatic recognition of serial numbers in bank notes. Pattern Recognit. 2014, 47, 2621–2634. [Google Scholar]
  13. Feng, B.Y.; Ren, M.; Zhang, X.Y.; Suen, C.Y. Part-based high accuracy recognition of serial numbers in bank notes. In IAPR Workshop on Artificial Neural Networks in Pattern Recognition; Springer: Berlin/Heidelberg, Germany, 2014; pp. 204–215. [Google Scholar]
  14. Tian, S.; Bhattacharya, U.; Lu, S.; Su, B.; Wang, Q.; Wei, X.; Lu, Y.; Tan, C.L. Multilingual scene character recognition with co-occurrence of histogram of oriented gradients. Pattern Recognit. 2016, 51, 125–134. [Google Scholar]
  15. Zhou, J.; Wang, F.; Xu, J.; Yan, Y.; Zhu, H. A novel character segmentation method for serial number on banknotes with complex background. J. Ambient Intell. Humaniz. Comput. 2019, 10, 2955–2969. [Google Scholar] [CrossRef]
  16. Alwzwazy, H.A.; Albehadili, H.M.; Alwan, Y.S.; Islam, N.E. Handwritten digit recognition using convolutional neural networks. Int. J. Innov. Res. Comput. Commun. Eng. 2016, 4, 1101–1106. [Google Scholar]
  17. Boufenar, C.; Kerboua, A.; Batouche, M. Investigation on deep learning for off-line handwritten Arabic character recognition. Cogn. Syst. Res. 2018, 50, 180–195. [Google Scholar] [CrossRef]
  18. OAHCDB-40 30000 Examples. Available online: http://www.mediafire.com/file/wks8xgfs0dmm4db/OAHCDB-40_30000_Exemples.rar/file (accessed on 23 September 2022).
  19. Zhao, N.; Zhang, Z.; Ouyang, X.; Lv, N.; Zang, Z. The recognition of RMB serial number based on CNN. In Proceedings of the 2018 Chinese Control and Decision Conference (CCDC), Shenyang, China, 9–11 June 2018; pp. 3303–3306. [Google Scholar]
  20. Wang, F.; Zhu, H.; Li, W.; Li, K. A hybrid convolution network for serial number recognition on banknotes. Inf. Sci. 2020, 512, 952–963. [Google Scholar]
  21. Jang, U.; Suh, K.H.; Lee, E.C. Low-quality banknote serial number recognition based on deep neural network. J. Inf. Process. Syst. 2020, 16, 224–237. [Google Scholar]
  22. Namysl, M.; Konya, I. Efficient, lexicon-free OCR using deep learning. In Proceedings of the 2019 International Conference on Document Analysis and Recognition (ICDAR), Sydney, Australia, 20–25 September 2019; pp. 295–301. [Google Scholar]
  23. Caldeira, T.; Ciarelli, P.M.; Neto, G.A. Industrial optical character recognition system in printing quality control of hot-rolled coils identification. J. Control Autom. Electr. Syst. 2020, 31, 108–118. [Google Scholar]
  24. Gang, S.; Fabrice, N.; Chung, D.; Lee, J. Character recognition of components mounted on printed circuit board using deep learning. Sensors 2021, 21, 2921. [Google Scholar] [CrossRef]
  25. He, K.; Zhang, X.; Ren, S.; Sun, J. Deep residual learning for image recognition. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Las Vegas, NV, USA, 27–30 June 2016; pp. 770–778. [Google Scholar]
  26. Tan, M.; Le, Q. EfficientNet: Rethinking model scaling for convolutional neural networks. In Proceedings of the International Conference on Machine Learning, PMLR, Long Beach, CA, USA, 9–15 June 2019; pp. 6105–6114. [Google Scholar]
  27. Liu, W.; Anguelov, D.; Erhan, D.; Szegedy, C.; Reed, S.; Fu, C.Y.; Berg, A.C. SSD: Single shot multibox detector. In European Conference on Computer Vision; Springer: Cham, Switzerland, 2016; pp. 21–37. [Google Scholar]
  28. Huang, G.; Liu, Z.; Van Der Maaten, L.; Weinberger, K.Q. Densely connected convolutional networks. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Honolulu, HI, USA, 21–26 July 2017; pp. 4700–4708. [Google Scholar]
  29. KISAN ELECTRONICS. Available online: https://kisane.com/ (accessed on 24 September 2022).
  30. Deng, G.; Cahill, L.W. An adaptive Gaussian filter for noise reduction and edge detection. In Proceedings of the IEEE Conference Record Nuclear Science Symposium and Medical Imaging Conference, San Francisco, CA, USA, 31 October–6 November 1993. [Google Scholar]
  31. He, K.; Zhang, X.; Ren, S.; Sun, J. Delving deep into rectifiers: Surpassing human-level performance on ImageNet classification. In Proceedings of the IEEE International Conference on Computer Vision, Santiago, Chile, 7–13 December 2015; pp. 1026–1034. [Google Scholar]
  32. Rothe, R.; Guillaumin, M.; Gool, L.V. Non-maximum suppression for object detection by passing messages between windows. In Asian Conference on Computer Vision; Springer: Berlin/Heidelberg, Germany, 2014; pp. 290–306. [Google Scholar]
Figure 1. Example of a banknote serial number (KRW 10,000 bill).
Figure 1. Example of a banknote serial number (KRW 10,000 bill).
Sensors 22 08612 g001
Figure 2. Examples of the backgrounds of serial number regions: (a) serial numbers with complex background patterns; (b) serial numbers with clean background patterns.
Figure 2. Examples of the backgrounds of serial number regions: (a) serial numbers with complex background patterns; (b) serial numbers with clean background patterns.
Sensors 22 08612 g002
Figure 3. Single shot multi-box detector (SSD) based multi-currency integrated serial number recognition model: (a) proposed model; (b) dense block comprising the backbone of the proposed model.
Figure 3. Single shot multi-box detector (SSD) based multi-currency integrated serial number recognition model: (a) proposed model; (b) dense block comprising the backbone of the proposed model.
Sensors 22 08612 g003
Figure 4. Example of ROI selections for different positions of the banknote serial number: (a) KRW; (b) USD; (c) INR; (d) JPY.
Figure 4. Example of ROI selections for different positions of the banknote serial number: (a) KRW; (b) USD; (c) INR; (d) JPY.
Sensors 22 08612 g004
Figure 5. Examples of cropped ROI images for each currency: (a) KRW; (b) USD; (c) INR; (d) JPY.
Figure 5. Examples of cropped ROI images for each currency: (a) KRW; (b) USD; (c) INR; (d) JPY.
Sensors 22 08612 g005
Figure 6. Example of image data augmentation for banknote image data. Red boxes indicate the ground truth. (a) Original image; (b) images obtained from data augmentation.
Figure 6. Example of image data augmentation for banknote image data. Red boxes indicate the ground truth. (a) Original image; (b) images obtained from data augmentation.
Sensors 22 08612 g006
Figure 7. Modified architectures: (a) Version 1; (b) Version 2; (c) Version 3; (d) Version 4.
Figure 7. Modified architectures: (a) Version 1; (b) Version 2; (c) Version 3; (d) Version 4.
Sensors 22 08612 g007
Figure 8. Examples of detection errors (yellow box: prediction; white box: ground truth): (a) first case; (b) second case.
Figure 8. Examples of detection errors (yellow box: prediction; white box: ground truth): (a) first case; (b) second case.
Sensors 22 08612 g008
Figure 9. Graphical accuracy comparison for the result of Table 4.
Figure 9. Graphical accuracy comparison for the result of Table 4.
Sensors 22 08612 g009
Figure 10. Graphical accuracy comparison for the result of Table 5 (Compared previous studies have marked “N/A” because only classification results exist without character region detection). (a) Zhao et al. [10]. (b) Ebrahimzadeh et al. [11]. (c) Feng et al. [13]. (d) Alwzwazy et al. [16]. (e) Boufenar et al. [17]. (f) Zhao et al. [19]. (g) Wang et al. [21]. (h) Jang et al. [22].
Figure 10. Graphical accuracy comparison for the result of Table 5 (Compared previous studies have marked “N/A” because only classification results exist without character region detection). (a) Zhao et al. [10]. (b) Ebrahimzadeh et al. [11]. (c) Feng et al. [13]. (d) Alwzwazy et al. [16]. (e) Boufenar et al. [17]. (f) Zhao et al. [19]. (g) Wang et al. [21]. (h) Jang et al. [22].
Sensors 22 08612 g010
Figure 11. Multi-currency (KRW, USD, INR, JPY) confusion matrix for the test dataset.
Figure 11. Multi-currency (KRW, USD, INR, JPY) confusion matrix for the test dataset.
Sensors 22 08612 g011
Figure 12. Graphical accuracy comparison for the result of Table 6.
Figure 12. Graphical accuracy comparison for the result of Table 6.
Sensors 22 08612 g012
Table 1. Size of dataset for each currency.
Table 1. Size of dataset for each currency.
CurrencyTrainingValidationTestTotal
KRW35,0007500750050,000
USD37,7248083809453,901
INR31,5006750675045,000
JPY31,2606695669644,651
Total97,08026,62421,832145,536
Table 2. Model performance comparison of different architecture versions.
Table 2. Model performance comparison of different architecture versions.
Version1234
Accuracy (%)97.8999.3399.4299.88
Number of
detection errors
5032300
Number of
classification errors
10812139
1st accuracy (%)98.9799.4999.72100
2nd accuracy (%)96.8199.1799.1299.75
Time
(ms)
44424430
Table 3. Scale and aspect ratio of anchor boxes for use in the multi-currency (KRW, USD, INR, and JPY) recognition model.
Table 3. Scale and aspect ratio of anchor boxes for use in the multi-currency (KRW, USD, INR, and JPY) recognition model.
CaseScaleAspect Ratio Set
1[0.41, 0.45, 0.61, 0.65][0.81, 0.85, 0.88]
2[0.30, 0.39, 0.47, 0.57][0.8, 0.9]
3[0.33, 0.41, 0.55, 0.62][0.5, 0.65, 0.88, 1.0, 1.1, 1.25, 1.4]
4[0.28, 0.36, 0.44, 0.52][0.5, 0.65, 0.88, 1.0, 1.1, 1.25, 1.4]
5[0.28, 0.36, 0.44, 0.52][0.5, 0.65, 0.88, 1.0, 1.1, 1.25, 1.4]
6[0.28, 0.36, 0.44, 0.52][0.5, 0.65, 0.88, 1.0, 1.1, 1.25, 1.4]
Table 4. Model performance for multi-currency (KRW, USD, INR, and JPY) recognition.
Table 4. Model performance for multi-currency (KRW, USD, INR, and JPY) recognition.
Case123456
Accuracy (%)99.7699.8199.8699.9599.9699.97
Number of
detection errors
493832860
Number of
classification errors
19166646
1st accuracy (%)99.8999.9299.9399.9599.9599.97
2nd accuracy (%)99.6399.6999.8099.9599.9599.97
Table 5. Comparison of banknote serial number recognition methods.
Table 5. Comparison of banknote serial number recognition methods.
ApproachMethodTaskNumber of ClassesDetection
Accuracy
(%)
Classification Accuracy
(%)
Handcrafted
feature
extraction
Zhao et al. [10]Classification17-95
Ebrahimzadeh et al. [11]Classification17-97.25
Feng et al. [13]Classification17-99.33
Deep-
learning
based
Alwzwazy et al. [16]Classification17-95.7
Boufenar et al. [17]Classification17-100
Zhao et al. [19]Classification17-99.99
Wang et al. [21]Classification17-99.89
Jang et al. [22]Classification17-99.92
Our methodDetection + Classification3710099.97
Table 6. Performance comparison of the single-currency dataset-based and multi-currency dataset-based models.
Table 6. Performance comparison of the single-currency dataset-based and multi-currency dataset-based models.
National
Currency
MetricsSingle-Currency ModelMulti-Currency
Integrated Model
KRWAccuracy (%)99.8899.94
Number of
detection errors
00
Number of
classification errors
94
1st accuracy (%)10099.94
2nd accuracy (%)99.7599.94
USDAccuracy (%)99.9499.97
Number of
detection errors
50
Number of
classification errors
02
1st accuracy (%)99.9599.97
2nd accuracy (%)99.9299.97
INRAccuracy (%)99.88100
Number of
detection errors
50
Number of
classification errors
30
1st accuracy (%)99.93100
2nd accuracy (%)99.83100
JPYAccuracy (%)99.80100
Number of
detection errors
00
Number of
classification errors
130
1st accuracy (%)99.98100
2nd accuracy (%)99.62100
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Share and Cite

MDPI and ACS Style

Jang, W.; Lee, C.; Jeong, D.S.; Lee, K.; Lee, E.C. Multi-Currency Integrated Serial Number Recognition Model of Images Acquired by Banknote Counters. Sensors 2022, 22, 8612. https://doi.org/10.3390/s22228612

AMA Style

Jang W, Lee C, Jeong DS, Lee K, Lee EC. Multi-Currency Integrated Serial Number Recognition Model of Images Acquired by Banknote Counters. Sensors. 2022; 22(22):8612. https://doi.org/10.3390/s22228612

Chicago/Turabian Style

Jang, Woohyuk, Chaewon Lee, Dae Sik Jeong, Kunyoung Lee, and Eui Chul Lee. 2022. "Multi-Currency Integrated Serial Number Recognition Model of Images Acquired by Banknote Counters" Sensors 22, no. 22: 8612. https://doi.org/10.3390/s22228612

APA Style

Jang, W., Lee, C., Jeong, D. S., Lee, K., & Lee, E. C. (2022). Multi-Currency Integrated Serial Number Recognition Model of Images Acquired by Banknote Counters. Sensors, 22(22), 8612. https://doi.org/10.3390/s22228612

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop