Next Article in Journal
Exergy Analysis of the Heart with a Stenosis in the Arterial Valve
Next Article in Special Issue
A New Algorithm for Medical Color Images Encryption Using Chaotic Systems
Previous Article in Journal
Learning Coefficient of Vandermonde Matrix-Type Singularities in Model Selection
Previous Article in Special Issue
Entropy in Image Analysis
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Improvement of Image Binarization Methods Using Image Preprocessing with Local Entropy Filtering for Alphanumerical Character Recognition Purposes

Faculty of Electrical Engineering, West Pomeranian University of Technology, Szczecin, 70-313 Szczecin, Poland
*
Author to whom correspondence should be addressed.
Entropy 2019, 21(6), 562; https://doi.org/10.3390/e21060562
Submission received: 26 May 2019 / Revised: 2 June 2019 / Accepted: 2 June 2019 / Published: 4 June 2019
(This article belongs to the Special Issue Entropy in Image Analysis II)

Abstract

:
Automatic text recognition from the natural images acquired in uncontrolled lighting conditions is a challenging task due to the presence of shadows hindering the shape analysis and classification of individual characters. Since the optical character recognition methods require prior image binarization, the application of classical global thresholding methods in such case makes it impossible to preserve the visibility of all characters. Nevertheless, the use of adaptive binarization does not always lead to satisfactory results for heavily unevenly illuminated document images. In this paper, the image preprocessing methodology with the use of local image entropy filtering is proposed, allowing for the improvement of various commonly used image thresholding methods, which can be useful also for text recognition purposes. The proposed approach was verified using a dataset of 140 differently illuminated document images subjected to further text recognition. Experimental results, expressed as Levenshtein distances and F-Measure values for obtained text strings, are promising and confirm the usefulness of the proposed approach.

1. Introduction

Image binarization is one of the most relevant preprocessing steps leading to significant decrease in the amount of information subjected to further analysis and allowing for an increase of its speed. Such an operation is typically applied in many systems which utilize mainly shape recognition methods and do not require the colour or texture analysis. Some good examples might be some robotic applications, including line followers and visual navigation in corridors and labyrinths, advanced driver-assistance systems (ADAS) and autonomous vehicles with lane tracking, as well as widely used optical character recognition (OCR) methods. Binary image analysis may also be applied successfully in embedded systems with limited amount of memory and low computational power.
Nevertheless, the appropriate results of binary image analysis, in particular text recognition, depend on the correct prior binarization. In some applications, where the uniform illumination of the scene can be ensured, e.g., popular flatbed scanners or some non-destructive automated book scanners, even with additional infrared cameras allowing for software straightening the scanned book pages [1], the simplest global thresholding may be sufficient. However, in many other situations the illumination may be non-uniform, especially in natural images captured by cameras, and therefore more sophisticated adaptive methods should be applied.
One of the most challenging problems related to the influence of image thresholding on further analysis is document image binarization and therefore newly developed algorithms are typically validated by using intentionally prepared document images containing various distortions. For this reason well-known document image binarization competitions (DIBCO) datasets are typically used to verify the usefulness and validate the advantages of binarization methods. These databases are prepared for yearly document image binarization competitions organized during two leading conferences in this field—the International Conference on Document Analysis and the Recognition (ICDAR) [2] and International Conference on Frontiers in Handwriting Recognition (ICFHR) [3], where the H-DIBCO datasets are used, containing only handwritten document images without machine printed samples. All DIBCO datasets contain not only the distorted document images but also “ground truth” binary images and therefore the binarization results can be compared with them at the pixel level analysing the numbers of correctly and improperly classified pixels [4,5].
Despite the fact that image binarization is not a new topic, some enhancements of algorithms are still proposed, particularly for historical document image binarization, as well as unevenly illuminated natural images. A proposal of such an improvement based on the image entropy filter, possible to apply in many commonly known binarization methods, is presented in this paper.
The rest of the paper consists of the short overview of the most widely used image binarization methods, description of the proposed approach based on the use of local entropy filter, presentation and discussion of results and final conclusions.

2. Brief Overview of Image Binarization Algorithms

Probably the most popular image thresholding method was proposed in 1979 by Nobuyuki Otsu [6], who delivered the idea of minimizing the sum of intra-class variances of two groups of pixels classified as foreground and background, assuming the bi-modal histogram of the image pixels’ intensity. Hence, this approach leads to maximization of inter-class variance and therefore a good separation of two classes of pixels, represented finally as black and white, is achieved. Due to the operations on the histograms, this method is fast, although it works properly only for uniformly illuminated images with bi-modal histograms.
A similar approach, utilizing the entropy of the histogram instead of variances was proposed by Kapur et al. [7], whereas the idea of combining the global and local Otsu and Kapur methods was presented in the paper [8]. An extended adaptive version of Otsu method, known as AdOtsu, proposed by Moghaddam and Cheriet [9], assumed some additional operations such as multi-scale background estimation and calculation of average stroke widths and line heights. Since some images with unimodal histograms cannot be properly binarized using the above mentioned histogram-based methods another interesting idea was presented by Paul Rosin [10], who proposed to determine the threshold as the corner of the histogram curve.
Since the images containing some shadows being the result of non-uniform illumination should not be binarized using a single global threshold, some adaptive algorithms, which require the analysis of each pixels’ neighbourhood, were proposed as well. The most popular approach developed by Wayne Niblack [11] assumed the determination of the local threshold as the average local intensity lowered by the local standard deviation scaled by the constant parameter k. A further modification of this approach, utilizing the additional normalization of the local standard deviation by its division by its maximum value in the image, is known as Sauvola method [12]. Its multi-scale version was further developed by Lazzara and Géraud [13].
A simple choice of the local threshold as the average of the minimum and the maximum intensity within the local window (so called midgray value) was proposed by John Bernsen [14], whereas Bradley and Roth [15] developed the method using the integral image for the calculation of the local mean intensity of the neighbourhood. The implementation of this method, also in the modified versions utilising the local median and Gaussian weighted mean, is available as MATLAB adaptthresh function.
Some other adaptive binarization methods were proposed by Wolf and Jolion [16], who used a relatively simple contrast maximization approach as a modification of Niblack’s method, as well as Feng and Tan [17], where a similar idea based on the maximization of local contrast was used, however significantly slower due to the application of additional median filtering and bilinear interpolation. Another method proposed by Gatos et al. [18] utilizes a low-pass Wiener filtering and background estimation, followed by the use of Sauvola’s thresholding with additional interpolation and post-processing using so called shrink and swell filters to remove noise and fill some foreground gaps and holes.
More recent document image binarization methods include the idea of region-based thresholding using Otsu’s method with additional use of support vector machines (SVM) presented by Chou et al. [19] as well as faster region-based approaches [20,21]. Another method utilising the SVM-based approach with local features was presented recently by Xiong et al. [22].
The algorithm proposed by Howe [23] utilizes a Laplacian operator, Canny edge detection and graph cut method to find the threshold minimizing the energy. Erol et al. [24] proposed a more general approach related to the localization of text on a document captured by mobile phone camera using morphological operations for background estimation. Another background suppression method, although working properly mainly for evenly illuminated document images, was proposed by Lu et al. [25], whereas another attempt to the application of morphological operations was presented by Okamoto et al. [26].
Lelore and Bouchara [27] proposed the extended fast algorithm for document image restoration (FAIR) algorithm based on rough text localization and likelihood estimation followed by simple thresholding of the obtained super-resolution likelihood image. A multi-scale adaptive–interpolative method was proposed by Bag and Bhowmick [28], useful for faint characters. A method proposed by Su et al. [29] exploited adaptive image contrast map combined with results of Canny edge detection, whereas an attempt to use multiple thresholding methods was presented by Yoon et al. [30].
Some faster ideas of image thresholding based on the Monte Carlo method were proposed as well [31,32,33], where the simplified histogram of the image was approximated using the limited number of randomly chosen pixels. On the other hand, Khitas et al. [34] developed recently an algorithm based on median filtering used for estimation of the background information. An application of local features with Gaussian mixtures was examined in the paper [35], whereas Chen and Wang [36] used extended non-local means method followed by adaptive thresholding with additional postprocessing.
Bataineh et al. [37] developed an algorithm inspired by Niblack’s and Sauvola’s methods with additional application of dynamic windows. Further modifications of Niblack’s method were proposed by Khurshid et al. [38], Kulyukin et al. [39] and recently by Samorodova and Samorodov [40]. A direct binarization scheme of colour document images based on multi-scale mean-shift algorithm with the use of modified Niblack’s method was recently proposed by Mysoreet al. [41]. A review of many modifications of Niblack inspired algorithms can be found in Saxena’s paper [42], whereas many other approaches are discussed in some other survey papers [43,44,45]. Some earlier methods can also be found in BinarizationShop software developed by Deng et al. [46].
Some recent trends in image binarization are related to the use of variational models [47] and deep learning methods [48]. Recently, Vo et al. [49] proposed another supervised approach based on hierarchical deep neural networks. A comprehensive overview of many document image binarization algorithms can be found in the survey paper written by Sulaiman et al. [50].
An interesting method of binarization of non-uniformly illuminated images based on Curvelet transform followed by Otsu’s thresholding was proposed by Wen et al. [51]. However, the application of this algorithms requires the additional nonlinear enhancement functions and time-consuming multi-scale processing.
Some of the binarization methods utilize the calculation of histogram entropy as well as image entropy. The most widely known approach proposed by Kapur et al. [7] may be considered as the modification of the classical Otsu’s thresholding, which is based on earlier ideas presented by Thierry Pun [52,53]. Fan et al. [54] proposed a method maximizing the 2D temporal entropy, whereas Abutaleb [55] developed a method which uses pixel’s grey level as well the average of its neighbourhood for minimization of two-dimensional entropy. Brink and Pendock [56] used the cross-entropy instead of distance or similarity between the original image and the result of binarization to optimize the threshold. Some similar multilevel methods have been further developed as well for image segmentation [57], also with the use of genetic methods [58]. A ternary entropy-based method [59], based on the classification of pixels into text, near-text, and non-text regions was proposed as well, which utilized Shannon entropy, whereas Tsallis entropy was used by Tian and Hou [60]. Nevertheless, entropy-based methods are generally less popular than simple histogram-based thresholding or some adaptive binarization methods. Apart from the typical image binarization, one can find some other applications of entropy related to classification of signals or images obtained as the results of measurements or some other experiments, e.g., in a gearbox testing system presented by Jiang et al. [61], where Shannon entropy of the vibration signal is used to detect worn and cracked gears.
Development of any new image processing algorithms usually requires their reliable validation based on the comparison of the obtained results with the other methods. Stathis et al. [62] proposed a method of evaluation of binarization algorithms based on comparison of individual pixels, using the pixel error rate (PERR), peak signal to noise ratio (PSNR) and similar metrics, whereas some other approaches were presented in the survey paper by Sezgin and Sankur [63]. A much more popular approach is the use of typical classification metrics based on precision, recall, sensitivity, specificity or F-Measure [4,5], as well as the application of misclassification penalty metric (MPM) [64] or distance reciprocal distortion (DRD) [65]. Another binarization assessment method was presented by Lins et al. [66], which utilizes a dataset of synthetic images for comparison of various thresholding algorithms. Nevertheless, considering the final results of the document image recognition as the recognized text strings, a more useful approach would be the application of metrics calculated for characters instead of individual pixels. Apart from F-Measure, some metrics dedicated for text strings, such as Levenshtein distance, defined as the number of character operations necessary to convert one string into another, may be applied as well.

3. Proposed Method and Its Experimental Verification

3.1. Description of the Method

Analysing the unevenly illuminated document images, important information can be achieved with the use of the local image entropy, which may be calculated using the MATLAB entropyfilt function. Using its default parameters the local measure of randomness of the grey levels of the neighbourhood defined by the 9 × 9 pixels mask was achieved and stored as the result for the central pixel. Such an approach may be useful for image forgery detection, switching purposes in adaptive median filtering as well as for image preprocessing followed by comparison of properties of image regions. Hence, the local entropy filter was considered in the proposed method as one of the preprocessing steps for adaptive image binarization of unevenly illuminated document images subjected to further optical text recognition.
It is worth noting that most of the OCR engines used some “built-in” thresholding procedures and therefore their results are dependent also on the quality of the input data. For example, widely used freeware Tesseract OCR developed by Google utilized global Otsu’s thresholding, whereas the commercial ABBYY FineReader software employed the adaptive Bradley’s method. Therefore, the application of some other image binarization methods may improve or decrease the recognition accuracy, since the OCR “internal” thresholding does not change the input binary image. Hence, prior image thresholding may be considered as a replacement of the default methods used in the OCR engines.
The proposed method caused the equalization of illumination of an image, increasing also its contrast, making it easier to conduct the proper binarization and further recognition of alphanumerical characters. It is based on the analysis of the local entropy, assuming its noticeably higher values in the neighbourhood of the characters. Hence, only the relatively high entropy regions should be further analyzed as potentially containing some characters, whereas low entropy regions may be considered as the background. The proposed algorithm consists of the following steps:
  • entropy filter—calculation of the local entropy using the predefined mask (in our experiments the most appropriate size is 19 × 19 pixels) leading to the local entropy map;
  • negative—simple negation leads to more readable dark characters on a bright background; assuming the maximum entropy value equal to eight (considering eight bits necessary to store 256 grey levels), the additional normalization can be applied with the formula Y = 1 X 8 , where X is the local entropy map and the final range of the output image Y is 0 ; 1 ;
  • thresholding—one of the global binarization methods may be used for this purpose, in our experiments the classical Otsu’s thresholding was used, leading to the image M with segmented regions containing text and representing the background;
  • masking—the obtained binary image M was used as the mask for the original input image, leading to the background image B with removed text regions;
  • morphological dilation—the purpose of this operation was to fill the gaps containing the characters making it possible to obtain a full estimate of the background; a critical element of this step is an appropriate choice of the size of the structuring element (in our experiments the square 20 × 20 pixels one was sufficient and larger structuring elements caused an increase of the computation time);
  • background subtraction—the expected result of the subtraction of the background estimate from the original input image should contain a bright text and the dark background with equalized illumination;
  • negation with increase of contrast—a simple operation leading to the dark text and the bright background with improved readability;
  • final binarization—the last step conducted after pre-processing, which can utilize any of commonly used binarization methods (in our experiments good results were obtained using adaptive Bradley’s and Niblack’s thresholding).
The simplified flowchart of the method is shown in Figure 1, whereas the illustration of results obtained after consecutive steps of the algorithm is presented in Figure 2.

3.2. Practical Verification

The verification of the proposed method was conducted using the database of document images, prepared applying various illuminations (uniform lighting and six types of non-uniform or directional shadows). The well-known quasi-Latin text Lorem ipsum, used as the basis for the generated sample pages containing 536 words, was printed using five various font shapes (Arial, Times New Roman, Calibri, Verdana and Courier) and their style modifications (normal, bold, italics and bold+italics). Such printed 20 sheets of paper were photographed applying 7 types of illuminations mentioned above (six unevenly illuminated examples are shown in Figure 3). These 140 captured images were binarized in two scenarios: with and without the proposed preprocessing. In both cases several binarization algorithms were applied to verify the proposed approach in practice. All the obtained binary images were used as the input data for the Google Tesseract OCR engine. For each of the images, the number of correctly and incorrectly recognized characters were determined, allowing for the calculation of some typical classification metrics, such as F-Measure defined as:
F M = 2 · P R · R C P R + R C ,
where P R and R C stand for the precision (true positives to sum of all positives ratio) and recall (ratio of true positives to sum of true positives and false negatives). Hence, they can be expressed as:
P R = T P T P + F P and R C = T P T P + F N ,
where T P are true positives and F N false negatives, respectively. All positive and negative values are considered as the numbers of correctly and incorrectly recognized characters.
The additional metric, which may be applied for the evaluation of text similarity, is known as Levenshtein distance, representing the minimum number of text changes (insertions, deletions or substitutions of individual characters) necessary to change the analyzed text into another. This metric was also applied for evaluation purposes, assuming the knowledge of the original text string (Lorem ipsum-based in these experiments).

4. Results and Discussion

The development of the final preprocessing algorithm allowing for the increase of the final OCR accuracy required an appropriate choice of some parameters mentioned earlier. The first of them is the size of the block used for the entropy filter which influences significantly the obtained results. Too small size of the filter would not be efficient due to its sensitivity to small details and noise whereas too big windows would be vulnerable to averaging effects. Since the default size of the filter in MATLAB entropyfilt function is 9 × 9 pixels, the first experiments were conducted using various windows to verify the influence of their size on the OCR results. The obtained results are presented in Figure 4, where the best values can be observed for 19 × 19 pixels filter. Therefore, the application of the default values would be inappropriate, particularly for the series #5 containing the non-uniformly illuminated images with sharp shadow edges as shown in Figure 3d.
A similar difference may be observed during the choice of the most appropriate size of the structuring element applied during the morphological dilation, since the results obtained for the series #5 differ significantly from the others. Nevertheless, in all cases the choice of a similar size of the structuring element to the size of the block in the entropy filter leads to the best results as illustrated in Figure 5 (in our experiments 20 × 20 pixels structuring element was chosen).
The additional reason of the choice of such structuring element was the processing time, which increased noticeably for bigger structuring elements as shown in Figure 6, where its values normalized according to the computation time obtained using the selected 20 × 20 pixels structuring element are presented. Unfortunately, relatively shorter processing did not guarantee good enough OCR accuracy, whereas increase of the structuring element’s size and computation time did not enhance the obtained results significantly. Since the experiments were conducted using a personal computer, some processes running in background (including the Tesseract OCR engine) might have influenced the obtained results. Nonetheless, the relation between the size of structuring element and the processing time can be considered as nearly linear. Hence, the most reasonable choice was the smallest possible structuring element not affecting the acceptable OCR accuracy level.
Having chosen the most appropriate parameters of the proposed preprocessing method, the obtained F-Measure values and Levenshtein distances for the whole dataset and each of the illumination types, as well as individual font faces and style modifications, were compared with some other methods applied without the proposed preprocessing. The comparison of the influence of the proposed preprocessing method on the F-Measure values is presented in Table 1, whereas respective Levenshtein distances are shown in Table 2. Analysing the results, a significant decrease of the Levenshtein distance, as well as the increase of the F-Measure values, may be observed for all methods, proving the usefulness of the proposed approach. The best results were achieved for Niblack, Sauvola and Wolf thresholding, as well as the simple Meanthresh method, which was significantly improved by the use of the entropy filtering-based preprocessing.
Some exemplary results obtained using the proposed preprocessing as well as its application for Bradley binarization with Gaussian kernel are illustrated in Figure 7. The additional illustration of its advantages for three exemplary images with the use of Niblack and Sauvola methods is shown in Figure 8, whereas another such comparison for Bernsen and Meanthresh methods is presented in Figure 9.
Since the properties of the proposed method may differ for various font shapes and styles, particularly for some of the thresholding algorithms, more detailed results are presented for them in Table 3 and Table 4, where F-Measure values can be compared for the same methods with and without the proposed entropy-based preprocessing method.
Comparing the influence of the proposed approach on the obtained OCR accuracy expressed as the F-Measure values calculated for individual text characters, relatively smaller enhancement may be observed for adaptive binarization methods, which achieve good results even without the proposed preprocessing method, such as Niblack or Sauvola. Nevertheless, in all cases the improvements may be noticed, also for the binarization method proposed by Wolf, which achieved much worse results for Courier fonts without the presented preprocessing method. A great improvement may also be observed for the simple mean thresholding as well as the direct usage of OCR engine’s built-in binarization, whereas the proposed method caused a small decrease of recognition accuracy after Bernsen thresholding for some font shapes (Courier and Times New Roman). It is worth to note that the proposed entropy-based preprocessing method always leads to better text recognition of bold fonts.

5. Conclusions

Binarization of unevenly illuminated and degraded document images is still an open and challenging field of research. Considering the necessity of fast image processing, many sophisticated methods, which cannot be effectively applied in many applications, may be replaced by simpler thresholding supported by less complicated preprocessing methods without the necessity of shape analysis or training procedures.
The approach proposed in the paper may be efficiently applied as the preprocessing step for many binarization methods in the presence of non-uniform illumination of document images, increasing significantly the accuracy of further text recognition, as shown in experimental results. Since its potential applicability is not limited to binarization of document images for OCR purposes, our further research may concentrate on the development of similar approaches for some other applications related to binarization of natural images and machine vision in robotics, particularly in unknown lighting conditions.

Author Contributions

H.M. worked under the supervision of K.O., H.M. prepared the data and sample document images, H.M. and K.O. designed the concept and methodology and proposed the algorithm, H.M. implemented the algorithm, performed the calculations and made the data visualization, K.O. validated the results and wrote the final version of the paper.

Funding

This research received no external funding.

Conflicts of Interest

The authors declare no conflict of interest.

Abbreviations

The following abbreviations are used in this manuscript:
ADASadvanced driver-assistance system
DIBCOdocument image binarization competition
DRDdistance reciprocal distortion
FAIRfast algorithm for document image restoration
ICDARInternational Conference on Document Analysis and Recognition
ICFHRInternational Conference on Frontiers in Handwriting Recognition
MPMmisclassification penalty metric
OCRoptical character recognition
PERRpixel error rate
PSNRpeak signal to noise ratio
SVMsupport vector machines

References

  1. Guizzo, E. Superfast Scanner Lets You Digitize Book by Flipping Pages. Available online: https://spectrum.ieee.org/automaton/robotics/robotics-software/book-flipping-scanning (accessed on 3 June 2019).
  2. Pratikakis, I.; Zagoris, K.; Barlas, G.; Gatos, B. ICDAR2017 Competition on Document Image Binarization (DIBCO 2017). In Proceedings of the 2017 14th IAPR International Conference on Document Analysis and Recognition (ICDAR), Kyoto, Japan, 9–15 November 2017; Volume 1, pp. 1395–1403. [Google Scholar] [CrossRef]
  3. Pratikakis, I.; Zagori, K.; Kaddas, P.; Gatos, B. ICFHR 2018 Competition on Handwritten Document Image Binarization (H-DIBCO 2018). In Proceedings of the 2018 16th International Conference on Frontiers in Handwriting Recognition (ICFHR), Niagara Falls, NY, USA, 5–8 August 2018; pp. 489–493. [Google Scholar] [CrossRef]
  4. Ntirogiannis, K.; Gatos, B.; Pratikakis, I. Performance evaluation methodology for historical document image binarization. IEEE Trans. Image Process. 2013, 22, 595–609. [Google Scholar] [CrossRef] [PubMed]
  5. Sokolova, M.; Lapalme, G. A systematic analysis of performance measures for classification tasks. Inf. Process. Manag. 2009, 45, 427–437. [Google Scholar] [CrossRef]
  6. Otsu, N. A threshold selection method from gray-level histograms. IEEE Trans. Syst. Man Cybern. 1979, 9, 62–66. [Google Scholar] [CrossRef]
  7. Kapur, J.; Sahoo, P.; Wong, A. A new method for gray-level picture thresholding using the entropy of the histogram. Comput. Vis. Graph. Image Process. 1985, 29, 273–285. [Google Scholar] [CrossRef]
  8. Lech, P.; Okarma, K.; Wojnar, D. Binarization of document images using the modified local-global Otsu and Kapur algorithms. Przegląd Elektrotechniczny 2015, 91, 71–74. [Google Scholar] [CrossRef]
  9. Moghaddam, R.F.; Cheriet, M. AdOtsu: An adaptive and parameterless generalization of Otsu’s method for document image binarization. Pattern Recognit. 2012, 45, 2419–2431. [Google Scholar] [CrossRef]
  10. Rosin, P.L. Unimodal thresholding. Pattern Recognit. 2001, 34, 2083–2096. [Google Scholar] [CrossRef]
  11. Niblack, W. An Introduction to Digital Image Processing; Prentice Hall: Englewood Cliffs, NJ, USA, 1986. [Google Scholar]
  12. Sauvola, J.; Pietikäinen, M. Adaptive document image binarization. Pattern Recognit. 2000, 33, 225–236. [Google Scholar] [CrossRef] [Green Version]
  13. Lazzara, G.; Géraud, T. Efficient multiscale Sauvola’s binarization. Int. J. Doc. Anal. Recognit. 2014, 17, 105–123. [Google Scholar] [CrossRef]
  14. Bernsen, J. Dynamic thresholding of grey-level images. In Proceedings of the 8th International Conference on Pattern Recognition (ICPR), Paris, France, 27–31 October 1986; pp. 1251–1255. [Google Scholar]
  15. Bradley, D.; Roth, G. Adaptive thresholding using the integral image. J. Graph. Tools 2007, 12, 13–21. [Google Scholar] [CrossRef]
  16. Wolf, C.; Jolion, J.M. Extraction and recognition of artificial text in multimedia documents. Form. Pattern Anal. Appl. 2004, 6, 309–326. [Google Scholar] [CrossRef] [Green Version]
  17. Feng, M.L.; Tan, Y.P. Adaptive binarization method for document image analysis. In Proceedings of the 2004 IEEE International Conference on Multimedia and Expo (ICME), Taipei, Taiwan, 27–30 June 2004; Volume 1, pp. 339–342. [Google Scholar] [CrossRef]
  18. Gatos, B.; Pratikakis, I.; Perantonis, S. Adaptive degraded document image binarization. Pattern Recognit. 2006, 39, 317–327. [Google Scholar] [CrossRef]
  19. Chou, C.H.; Lin, W.H.; Chang, F. A binarization method with learning-built rules for document images produced by cameras. Pattern Recognit. 2010, 43, 1518–1530. [Google Scholar] [CrossRef] [Green Version]
  20. Michalak, H.; Okarma, K. Region based adaptive binarization for optical character recognition purposes. In Proceedings of the International Interdisciplinary PhD Workshop (IIPhDW), Swinoujscie, Poland, 9–12 May 2018; pp. 361–366. [Google Scholar] [CrossRef]
  21. Michalak, H.; Okarma, K. Fast adaptive image binarization using the region based approach. In Artificial Intelligence and Algorithms in Intelligent Systems; Silhavy, R., Ed.; Springer International Publishing: Cham, Switzerland, 2019; Volume 764, AISC; pp. 79–90. [Google Scholar] [CrossRef]
  22. Xiong, W.; Xu, J.; Xiong, Z.; Wang, J.; Liu, M. Degraded historical document image binarization using local features and support vector machine (SVM). Optik 2018, 164, 218–223. [Google Scholar] [CrossRef]
  23. Howe, N.R. A Laplacian energy for document binarization. In Proceedings of the 2011 International Conference on Document Analysis and Recognition (ICDAR), Beijing, China, 18–21 September 2011; pp. 6–10. [Google Scholar] [CrossRef]
  24. Erol, B.; Antúnez, E.R.; Hull, J.J. HOTPAPER: multimedia interaction with paper using mobile phones. In Proceedings of the 16th International Conference on Multimedia 2008, Vancouver, BC, Canada, 26–31 October 2008; pp. 399–408. [Google Scholar] [CrossRef]
  25. Lu, S.; Su, B.; Tan, C.L. Document image binarization using background estimation and stroke edges. Int. J. Doc. Anal. Recognit. (IJDAR) 2010, 13, 303–314. [Google Scholar] [CrossRef]
  26. Okamoto, A.; Yoshida, H.; Tanaka, N. A binarization method for degraded document images with morphological operations. In Proceedings of the 2013 IAPR International Conference on Machine Vision Applications (MVA 2013), Kyoto, Japan, 20–23 May 2013; pp. 294–297. [Google Scholar]
  27. Lelore, T.; Bouchara, F. Super-resolved binarization of text based on the FAIR algorithm. In Proceedings of the 2011 International Conference on Document Analysis and Recognition (ICDAR), Beijing, China, 18–21 September 2011; pp. 839–843. [Google Scholar] [CrossRef]
  28. Bag, S.; Bhowmick, P. Adaptive-interpolative binarization with stroke preservation for restoration of faint characters in degraded documents. J. Vis. Commun. Image Represent. 2015, 31, 266–281. [Google Scholar] [CrossRef]
  29. Su, B.; Lu, S.; Tan, C.L. Robust document image binarization technique for degraded document images. IEEE Trans. Image Process. 2013, 22, 1408–1417. [Google Scholar] [CrossRef] [PubMed]
  30. Yoon, Y.; Ban, K.D.; Yoon, H.; Lee, J.; Kim, J. Best combination of binarization methods for license plate character segmentation. ETRI J. 2013, 35, 491–500. [Google Scholar] [CrossRef]
  31. Lech, P.; Okarma, K. Optimization of the fast image binarization method based on the Monte Carlo approach. Elektronika Ir Elektrotechnika 2014, 20, 63–66. [Google Scholar] [CrossRef]
  32. Lech, P.; Okarma, K. Fast histogram based image binarization using the Monte Carlo threshold estimation. In Computer Vision and Graphics; Chmielewski, L.J., Kozera, R., Shin, B.S., Wojciechowski, K., Eds.; Springer International Publishing: Cham, Switzerland, 2014; Volume 8671, pp. 382–390. [Google Scholar] [CrossRef]
  33. Lech, P.; Okarma, K. Prediction of the optical character recognition accuracy based on the combined assessment of image binarization results. Elektronika Ir Elektrotechnika 2015, 21, 62–65. [Google Scholar] [CrossRef]
  34. Khitas, M.; Ziet, L.; Bouguezel, S. Improved degraded document image binarization using median filter for background estimation. Elektronika ir Elektrotechnika 2018, 24, 82–87. [Google Scholar] [CrossRef]
  35. Mitianoudis, N.; Papamarkos, N. Document image binarization using local features and Gaussian mixture modeling. Image Vis. Comput. 2015, 38, 33–51. [Google Scholar] [CrossRef]
  36. Chen, Y.; Wang, L. Broken and degraded document images binarization. Neurocomputing 2017, 237, 272–280. [Google Scholar] [CrossRef]
  37. Bataineh, B.; Abdullah, S.N.H.S.; Omar, K. An adaptive local binarization method for document images based on a novel thresholding method and dynamic windows. Pattern Recognit. Lett. 2011, 32, 1805–1813. [Google Scholar] [CrossRef]
  38. Khurshid, K.; Siddiqi, I.; Faure, C.; Vincent, N. Comparison of Niblack inspired binarization methods for ancient documents. In Proceedings of the Document Recognition and Retrieval XVI, San Jose, CA, USA, 18–22 January 2009; Volume 7247, pp. 7247:1–7247:9. [Google Scholar] [CrossRef]
  39. Kulyukin, V.; Kutiyanawala, A.; Zaman, T. Eyes-free barcode detection on smartphones with Niblack’s binarization and Support Vector Machines. In Proceedings of the 16th International Conference on Image Processing, Computer Vision, and Pattern Recognition (IPCV’2012), Las Vegas, NV, USA, 16–19 July 2012; Volume 1, pp. 284–290. [Google Scholar]
  40. Samorodova, O.A.; Samorodov, A.V. Fast implementation of the Niblack binarization algorithm for microscope image segmentation. Pattern Recognit. Image Anal. 2016, 26, 548–551. [Google Scholar] [CrossRef]
  41. Mysore, S.; Gupta, M.K.; Belhe, S. Complex and degraded color document image binarization. In Proceedings of the 2016 3rd International Conference on Signal Processing and Integrated Networks (SPIN), Noida, India, 11–12 February 2016; pp. 157–162. [Google Scholar] [CrossRef]
  42. Saxena, L.P. Niblack’s binarization method and its modifications to real-time applications: A review. Artif. Intell. Rev. 2017, 1–33. [Google Scholar] [CrossRef]
  43. Leedham, G.; Yan, C.; Takru, K.; Tan, J.H.N.; Mian, L. Comparison of some thresholding algorithms for text/background segmentation in difficult document images. In Proceedings of the 7th International Conference on Document Analysis and Recognition (ICDAR 2003), Edinburgh, UK, 3–6 August 2003; pp. 859–864. [Google Scholar] [CrossRef]
  44. Shrivastava, A.; Srivastava, D.K. A review on pixel-based binarization of gray images. In ICICT 2015; Springer: Singapore, 2016; Volume 439, pp. 357–364. [Google Scholar] [CrossRef]
  45. Mustafa, W.A.; Kader, M.M.M.A. Binarization of document images: A comprehensive review. J. Phys. Conf. Ser. 2018, 1019, 012023. [Google Scholar] [CrossRef]
  46. Deng, F.; Wu, Z.; Lu, Z.; Brown, M.S. Binarizationshop: A user assisted software suite for converting old documents to black-and-white. In Proceedings of the Annual Joint Conference on Digital Libraries, Gold Coast, Queensland, Australia, 21–25 June 2010; pp. 255–258. [Google Scholar] [CrossRef]
  47. Feng, S. A novel variational model for noise robust document image binarization. Neurocomputing 2019, 325, 288–302. [Google Scholar] [CrossRef]
  48. Tensmeyer, C.; Martinez, T. Document image binarization with fully convolutional neural networks. In Proceedings of the 14th IAPR International Conference on Document Analysis and Recognition (ICDAR 2017), Kyoto, Japan, 9–15 November 2017; pp. 99–104. [Google Scholar] [CrossRef]
  49. Vo, Q.N.; Kim, S.H.; Yang, H.J.; Lee, G. Binarization of degraded document images based on hierarchical deep supervised network. Pattern Recognit. 2018, 74, 568–586. [Google Scholar] [CrossRef]
  50. Sulaiman, A.; Omar, K.; Nasrudin, M.F. Degraded historical document binarization: A review on issues, challenges, techniques, and future directions. J. Imaging 2019, 5, 48. [Google Scholar] [CrossRef]
  51. Wen, J.; Li, S.; Sun, J. A new binarization method for non-uniform illuminated document images. Pattern Recognit. 2013, 46, 1670–1690. [Google Scholar] [CrossRef]
  52. Pun, T. A new method for grey-level picture thresholding using the entropy of the histogram. Signal Process. 1980, 2, 223–237. [Google Scholar] [CrossRef]
  53. Pun, T. Entropic thresholding, a new approach. Comput. Graph. Image Process. 1981, 16, 210–239. [Google Scholar] [CrossRef] [Green Version]
  54. Fan, J.; Wang, R.; Zhang, L.; Xing, D.; Gan, F. Image sequence segmentation based on 2D temporal entropic thresholding. Pattern Recognit. Lett. 1996, 17, 1101–1107. [Google Scholar] [CrossRef]
  55. Abutaleb, A.S. Automatic thresholding of gray-level pictures using two-dimensional entropy. Comput. Vis. Graph. Image Process. 1989, 47, 22–32. [Google Scholar] [CrossRef]
  56. Brink, A.; Pendock, N. Minimum cross-entropy threshold selection. Pattern Recognit. 1996, 29, 179–188. [Google Scholar] [CrossRef]
  57. Li, J.; Tang, W.; Wang, J.; Zhang, X. A multilevel color image thresholding scheme based on minimum cross entropy and alternating direction method of multipliers. Optik 2019, 183, 30–37. [Google Scholar] [CrossRef]
  58. Tang, K.; Yuan, X.; Sun, T.; Yang, J.; Gao, S. An improved scheme for minimum cross entropy threshold selection based on genetic algorithm. Knowl.-Based Syst. 2011, 24, 1131–1138. [Google Scholar] [CrossRef]
  59. Le, T.H.N.; Bui, T.D.; Suen, C.Y. Ternary entropy-based binarization of degraded document images using morphological operators. In Proceedings of the 2011 International Conference on Document Analysis and Recognition, Beijing, China, 18–21 September 2011; pp. 114–118. [Google Scholar] [CrossRef]
  60. Tian, X.; Hou, X. A Tsallis-entropy image thresholding method based on two-dimensional histogram obique segmentation. In Proceedings of the 2009 WASE International Conference on Information Engineering, Taiyuan, China, 10–11 July 2009; Volume 1, pp. 164–168. [Google Scholar] [CrossRef]
  61. Jiang, Y.; Zhu, H.; Malekian, R.; Ding, C. An improved quantitative recurrence analysis using artificial intelligence based image processing applied to sensor measurements. Concurr. Comput. Pract. Exp. 2019, 31, e4858. [Google Scholar] [CrossRef]
  62. Stathis, P.; Kavallieratou, E.; Papamarkos, N. An evaluation technique for binarization algorithms. J. UCS 2008, 14, 3011–3030. [Google Scholar] [CrossRef]
  63. Sezgin, M.; Sankur, B. Survey over image thresholding techniques and quantitative performance evaluation. J. Electron. Imaging 2004, 13, 146–165. [Google Scholar] [CrossRef]
  64. Young, D.P.; Ferryman, J.M. PETS metrics: On-line performance evaluation service. In Proceedings of the 2005 IEEE International Workshop on Visual Surveillance and Performance Evaluation of Tracking and Surveillance, Beijing, China, 15–16 October 2005; pp. 317–324. [Google Scholar] [CrossRef]
  65. Lu, H.; Kot, A.; Shi, Y. Distance-reciprocal distortion measure for binary document images. IEEE Signal Process. Lett. 2004, 11, 228–231. [Google Scholar] [CrossRef]
  66. Lins, R.D.; de Almeida, M.M.; Bernardino, R.B.; Jesus, D.; Oliveira, J.M. Assessing binarization techniques for document images. In Proceedings of the 2017 ACM Symposium on Document Engineering (DocEng), Valletta, Malta, 4–7 September 2017; pp. 183–192. [Google Scholar] [CrossRef]
Figure 1. The simplified flowchart of the proposed method.
Figure 1. The simplified flowchart of the proposed method.
Entropy 21 00562 g001
Figure 2. Results of the consecutive steps of the proposed algorithm obtained for an exemplary document image: (a) original input image, (b) local entropy map, (c) normalized negative entropy image, (d) binarized entropy image, (e) result of masking, (f) dilated masked image being the full background estimate, (g) result of background subtraction, (h) negative with eliminated background, and (i) final result of adaptive Niblack’s thresholding after preprocessing.
Figure 2. Results of the consecutive steps of the proposed algorithm obtained for an exemplary document image: (a) original input image, (b) local entropy map, (c) normalized negative entropy image, (d) binarized entropy image, (e) result of masking, (f) dilated masked image being the full background estimate, (g) result of background subtraction, (h) negative with eliminated background, and (i) final result of adaptive Niblack’s thresholding after preprocessing.
Entropy 21 00562 g002
Figure 3. Exemplary unevenly illuminated images used in experiments: (a) side shading—series #2, (b) shading from the bottom—series #3, (c) diagonal shading—series #4, (d) irregular sharp shadow edges—series #5, (e) arc type shadows—series #6, (f) overexposure in the central part with underexposed boundaries—series #7.
Figure 3. Exemplary unevenly illuminated images used in experiments: (a) side shading—series #2, (b) shading from the bottom—series #3, (c) diagonal shading—series #4, (d) irregular sharp shadow edges—series #5, (e) arc type shadows—series #6, (f) overexposure in the central part with underexposed boundaries—series #7.
Entropy 21 00562 g003
Figure 4. Experimental OCR results obtained for various size of blocks applied in the entropy filter: (a) F-Measure values, (b) Levenshtein distance.
Figure 4. Experimental OCR results obtained for various size of blocks applied in the entropy filter: (a) F-Measure values, (b) Levenshtein distance.
Entropy 21 00562 g004
Figure 5. Experimental optical character recognition (OCR) results obtained for various size of structuring element applied for morphological dilation: (a) F-Measure values, (b) Levenshtein distance.
Figure 5. Experimental optical character recognition (OCR) results obtained for various size of structuring element applied for morphological dilation: (a) F-Measure values, (b) Levenshtein distance.
Entropy 21 00562 g005
Figure 6. Normalized processing time for various size of structuring elements used in morphological dilation relatively to the time obtained applying the 20 × 20 pixels structuring element.
Figure 6. Normalized processing time for various size of structuring elements used in morphological dilation relatively to the time obtained applying the 20 × 20 pixels structuring element.
Entropy 21 00562 g006
Figure 7. Comparison of binarization results obtained for exemplary unevenly illuminated images before the binarization: (a) without preprocessing, (b) with the proposed preprocessing, as well as using the Bradley method with a Gaussian kernel: (c) without preprocessing, (d) with the proposed preprocessing.
Figure 7. Comparison of binarization results obtained for exemplary unevenly illuminated images before the binarization: (a) without preprocessing, (b) with the proposed preprocessing, as well as using the Bradley method with a Gaussian kernel: (c) without preprocessing, (d) with the proposed preprocessing.
Entropy 21 00562 g007
Figure 8. Comparison of binarization results obtained for exemplary unevenly illuminated images using the Niblack method: (a) without preprocessing, (b) with the proposed preprocessing, as well as Sauvola thresholding: (c) without preprocessing, (d) with the proposed preprocessing.
Figure 8. Comparison of binarization results obtained for exemplary unevenly illuminated images using the Niblack method: (a) without preprocessing, (b) with the proposed preprocessing, as well as Sauvola thresholding: (c) without preprocessing, (d) with the proposed preprocessing.
Entropy 21 00562 g008
Figure 9. Comparison of binarization results obtained for exemplary unevenly illuminated images using the Bernsen method: (a) without preprocessing, (b) with the proposed preprocessing, as well as using the Meanthresh: (c) without preprocessing, (d) with the proposed preprocessing.
Figure 9. Comparison of binarization results obtained for exemplary unevenly illuminated images using the Bernsen method: (a) without preprocessing, (b) with the proposed preprocessing, as well as using the Meanthresh: (c) without preprocessing, (d) with the proposed preprocessing.
Entropy 21 00562 g009
Table 1. Comparison of F-Measure values obtained for various binarization methods with and without the proposed preprocessing.
Table 1. Comparison of F-Measure values obtained for various binarization methods with and without the proposed preprocessing.
Binarization MethodSeries
#1#2#3#4#5#6#7All
None0.96380.62010.81390.66500.66930.74600.62600.7291
+ preprocessing0.97280.64750.87290.72720.81670.80270.95840.8283
Otsu (global) [6]0.96140.62810.79080.66620.68410.75980.65830.7355
+ preprocessing0.97370.64000.85730.73120.80490.79470.95610.8226
Region-based [20]0.96160.75790.86610.84070.77370.83180.95280.8550
+ preprocessing0.95250.83770.88610.82540.74380.84680.91040.8575
Niblack [11]0.96140.79200.86680.84440.85100.85670.95890.8759
+ preprocessing0.95960.94390.94510.95160.88780.94360.96740.9427
Sauvola [12]0.97090.95810.96460.97220.76600.96550.97210.9385
+ preprocessing0.96740.96350.96650.96680.84010.96710.96940.9487
Wolf [16]0.96610.94820.95130.95140.76140.95940.97030.9297
+ preprocessing0.96910.96610.96430.96620.85610.96210.96570.9499
Bradley (mean) [15]0.96650.91910.90930.84840.73690.89760.96990.8925
+ preprocessing0.96660.88960.91690.92620.80400.91030.96420.9111
Bradley (Gaussian) [15]0.96630.85210.82950.75280.72670.79070.94890.8381
+ preprocessing0.96780.88630.89910.87410.75210.87860.91240.8815
Feng [17]0.91100.37820.79240.63120.72920.79380.84610.7285
+ preprocessing0.92610.44180.79900.64890.71030.80760.86880.7432
Bernsen [14]0.69480.64140.68440.64670.62860.71220.72450.6764
+ preprocessing0.69710.66880.69380.67520.63120.70470.71410.6836
Meanthresh0.95970.73480.83140.79210.83170.79470.93080.8393
+ preprocessing0.96510.95700.95960.96020.89700.96060.96840.9525
Table 2. Comparison of Levenshein distances obtained for various binarization methods with and without the proposed preprocessing.
Table 2. Comparison of Levenshein distances obtained for various binarization methods with and without the proposed preprocessing.
Binarization MethodSeries
#1#2#3#4#5#6#7All
None56.401897.201031.801362.401548.301387.901815.501299.93
+ preprocessing10.901665.10718.401045.20512.551063.8568.15726.31
Otsu (global) [6]62.751878.201039.801393.401514.551358.551715.801280.44
+ preprocessing12.601671.85720.051047.05514.201066.7576.10729.80
Region-based [20]27.30537.40388.35217.50294.35423.6044.75276.18
+ preprocessing27.40133.5578.90141.60378.55166.3048.15139.21
Niblack [11]30.50560.55359.95388.00222.10398.0531.15284.33
+ preprocessing26.0042.9035.4025.5579.4532.2016.5536.86
Sauvola [12]20.3022.8517.3514.80651.6017.7512.40108.15
+ preprocessing22.4030.2523.0517.35197.5519.7515.9546.61
Wolf [16]21.3554.9069.9074.05923.6558.5517.60174.29
+ preprocessing21.4527.7519.7523.50202.7517.8016.6547.10
Bradley (mean) [15]26.4563.15157.15389.451231.95188.2517.35296.25
+ preprocessing26.3075.6052.0544.10312.1554.0519.0583.33
Bradley (Gaussian) [15]27.10355.80731.00950.751282.401136.7032.25645.14
+ preprocessing25.7591.05193.00219.95700.15149.5019.95199.91
Feng [17]66.202518.001069.501507.501030.501037.10174.201057.57
+ preprocessing59.152385.251015.751435.10887.75945.70142.30981.57
Bernsen [14]467.751471.251071.001273.651634.151167.40623.101101.19
+ preprocessing490.401178.101046.751011.851402.251093.35687.35987.15
Meanthresh20.85776.30529.10519.85250.70763.3572.40418.94
+ preprocessing21.9526.1021.5517.2581.6520.0014.2028.96
Table 3. Comparison of F-Measure values obtained for various binarization methods with and without the proposed preprocessing for various font faces.
Table 3. Comparison of F-Measure values obtained for various binarization methods with and without the proposed preprocessing for various font faces.
Binarization MethodFont Face
ArialTimes New RomanCalibriCourierVerdana
None0.75560.74320.73740.64830.7612
+ preprocessing0.85410.82770.78860.81730.8539
Otsu (global) [6]0.74890.75280.75250.65980.7637
+ preprocessing0.85060.82140.78020.80580.8548
Region-based [20]0.87260.87990.87380.79700.8514
+ preprocessing0.85130.86890.85900.84250.8659
Niblack [11]0.87760.90120.87290.84990.8777
+ preprocessing0.94630.95500.94750.91950.9452
Sauvola [12]0.93950.94760.94120.92390.9402
+ preprocessing0.95550.95400.94500.93950.9495
Wolf [16]0.93990.95070.93550.88260.9400
+ preprocessing0.95670.95580.95510.93100.9511
Bradley (mean) [15]0.90360.90040.89460.86760.8963
+ preprocessing0.91580.91860.91580.89060.9147
Bradley (Gaussian) [15]0.84750.84480.84340.80870.8463
+ preprocessing0.90040.89920.88730.86090.8599
Feng [17]0.71370.74300.71130.75280.7210
+ preprocessing0.73680.74620.73040.75400.7487
Bernsen [14]0.67350.69700.69380.62130.6971
+ preprocessing0.70620.69170.69560.60410.7202
Meanthresh0.82510.86980.84830.81970.8337
+ preprocessing0.95110.96230.95160.94290.9548
Table 4. Comparison of F-Measure values obtained for various binarization methods with and without the proposed preprocessing for various font styles.
Table 4. Comparison of F-Measure values obtained for various binarization methods with and without the proposed preprocessing for various font styles.
Binarization MethodFont Style
NormalBoldItalicBold + Italic
None0.69450.74970.72210.7291
+ preprocessing0.80490.84550.80760.8283
Otsu (global) [6]0.70950.75440.72720.7355
+ preprocessing0.79800.84260.80380.8226
Region-based [20]0.86310.84440.87000.8550
+ preprocessing0.86210.85900.85930.8575
Niblack [11]0.87810.88980.86690.8759
+ preprocessing0.93960.94440.94240.9427
Sauvola [12]0.93660.93770.93400.9385
+ preprocessing0.94640.95450.94630.9487
Wolf [16]0.91650.94300.92230.9297
+ preprocessing0.94280.95670.94670.9499
Bradley (mean) [15]0.88880.89420.89160.8925
+ preprocessing0.90310.92300.90990.9111
Bradley (Gaussian) [15]0.83420.84180.83700.8381
+ preprocessing0.88010.87380.87540.8815
Feng [17]0.73330.73420.73910.7285
+ preprocessing0.73680.74580.75180.7432
Bernsen [14]0.67220.67860.67180.6764
+ preprocessing0.66560.70600.65730.6836
Meanthresh0.83790.84540.83810.8393
+ preprocessing0.95470.95190.95410.9525

Share and Cite

MDPI and ACS Style

Michalak, H.; Okarma, K. Improvement of Image Binarization Methods Using Image Preprocessing with Local Entropy Filtering for Alphanumerical Character Recognition Purposes. Entropy 2019, 21, 562. https://doi.org/10.3390/e21060562

AMA Style

Michalak H, Okarma K. Improvement of Image Binarization Methods Using Image Preprocessing with Local Entropy Filtering for Alphanumerical Character Recognition Purposes. Entropy. 2019; 21(6):562. https://doi.org/10.3390/e21060562

Chicago/Turabian Style

Michalak, Hubert, and Krzysztof Okarma. 2019. "Improvement of Image Binarization Methods Using Image Preprocessing with Local Entropy Filtering for Alphanumerical Character Recognition Purposes" Entropy 21, no. 6: 562. https://doi.org/10.3390/e21060562

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop