Next Article in Journal
Evaluation of Online Teaching Quality Based on Facial Expression Recognition
Previous Article in Journal
Overview of Blockchain Oracle Research
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Toward Semi-Supervised Graphical Object Detection in Document Images

by
Goutham Kallempudi
1,
Khurram Azeem Hashmi
1,2,3,*,
Alain Pagani
3,
Marcus Liwicki
4,
Didier Stricker
1,3 and
Muhammad Zeshan Afzal
1,2,3
1
Department of Computer Science, Technical University of Kaiserslautern, 67663 Kaiserslautern, Germany
2
Mindgarage, Technical University of Kaiserslautern, 67663 Kaiserslautern, Germany
3
German Research Institute for Artificial Intelligence (DFKI), 67663 Kaiserslautern, Germany
4
Department of Computer Science, Luleå University of Technology, 97187 Lulea, Sweden
*
Author to whom correspondence should be addressed.
Future Internet 2022, 14(6), 176; https://doi.org/10.3390/fi14060176
Submission received: 29 April 2022 / Revised: 28 May 2022 / Accepted: 5 June 2022 / Published: 8 June 2022
(This article belongs to the Section Big Data and Augmented Intelligence)

Abstract

:
The graphical page object detection classifies and localizes objects such as Tables and Figures in a document. As deep learning techniques for object detection become increasingly successful, many supervised deep neural network-based methods have been introduced to recognize graphical objects in documents. However, these models necessitate a substantial amount of labeled data for the training process. This paper presents an end-to-end semi-supervised framework for graphical object detection in scanned document images to address this limitation. Our method is based on a recently proposed Soft Teacher mechanism that examines the effects of small percentage-labeled data on the classification and localization of graphical objects. On both the PubLayNet and the IIIT-AR-13K datasets, the proposed approach outperforms the supervised models by a significant margin in all labeling ratios ( 1 % ,   5 % , and 10 % ) . Furthermore, the 10 % PubLayNet Soft Teacher model improves the average precision of Table, Figure, and List by + 5.4 , + 1.2 , and + 3.2 points, respectively, with a similar total mAP as the Faster-RCNN baseline. Moreover, our model trained on 10 % of IIIT-AR-13K labeled data beats the previous fully supervised method + 4.5 points.

1. Introduction

Visual summary is an important aspect in a variety of applications, including summarizing the contents of a document and detecting graphical elements in the visualization pipeline. As a result, identifying and localizing graphical items will be an important step in document summary and analysis. With the increase in documents, it has become impractical to manually extract visual objects. Automated methods provide reliable, effective and efficient solutions for manual tasks. For instance, ref. [1] verified that the machine learning method performs better than humans in domain knowledge and attention-demanding tasks. Similarly, several automated methods [2,3,4] have been proposed to identify graphical objects, but these automated methods are typically rule-based, because the documents lack established dimension or structure [5].
The graphical page object detection aims at localizing and classifying multiple objects such as tables, images, and figures in a document. For instance, Figure 1 and Figure 2 detect and localize objects in PubLayNet and IIIT-AR-13K datasets, respectively. As opposed to natural images and scenes, graphical objects have very little difference in their appearance; for example, Figure 1 (right) consists of the text block and a list of items block which appear to be similar, but it is necessary to classify them separately, making graphical page object detection more challenging.
Although defining rules for all graphical objects and running an optical character recognition tool helps localize objects, they cannot generalize to the new objects. This problem has resulted in the usage of deep learning models to detect graphical objects [6,7,8,9]. The deep learning models are rule independent and overcome the generalization problem with high precision. To train a deep learning model, a large amount of annotated data must be available, and as a result, either a manual labeling process or another pre-processing technique must be involved in generating high-quality data, which is both time-consuming and error-prone [10]. Owing to the concerns with training data, the problem changes its state from supervised to semi-supervised. This paper’s primary goal is to leverage the unlabeled data in identifying graphical objects without compromising the mean average precision and performance. We leverage the recently proposed Soft Teacher mechanism [10] to design a semi-supervised pipeline for graphical object detection on scanned document images.
The semi-supervised deep learning techniques in object detection [11,12,13,14] follow a predefined pipeline where an initial detector is trained using a small amount of available annotated data, and this detector is used to generate pseudo labels for unlabeled data. Finally, the labeled data along with unlabeled data with pseudo labels are used to retrain the model. The quality of these pseudo-labeling approaches depends on the initial model, and a weak initial model may degrade the pseudo-labeling process.
To alleviate the weak initial model problem, the end-to-end semi-supervised framework simultaneously performs pseudo labeling for unlabeled data and trains a detector using these pseudo labels with few annotated data in each iteration. The approach trains two models, one for detection known as Student and the other for pseudo labeling known as Teacher. In addition to this, the teacher model is simply an Exponential Moving Average (EMA) of the student model; this ensures that the pseudo-labeling process is constantly updated by the detection process and vice versa. Therefore, both the models reinforce each other.
The problem of GOD can be formulated as a generic object detection problem where we treat document images and graphical page objects, such as figures and tables, as natural objects. Therefore, inspired by the idea of [10], we leverage the use of the end-to-end semi-supervised framework on the graphical page object detection (GOD). Unlike the traditional pseudo-labeling approaches which are multi-staged but not end-to-end, the Soft Teacher ensures that a single round of iteration is complete from pseudo labeling to generating the loss by combining both labeled and pseudo-labeled data. Additionally, the framework provides a reinforcement effect, ensuring that the student model is always monitored by the teacher. Furthermore, this paper empirically shows that semi-supervised GOD methods can produce comparable results to fully labeled supervised GOD approaches.
The rest of the paper is structured as follows. Section 2 introduces multiple ways for object detection and categorizes them into rule-based, learning-based, deep learning and semi-supervised methods. The novel Soft Teacher framework is described in Section 3, which is followed by a dataset introduction in Section 4.1, evaluation protocol in Section 4.2, implementation details in Section 4.3 and results in Section 4.4. Finally, Section 5 concludes the paper with some future directions.

2. Related Work

There are numerous techniques for localizing objects in an image, ranging from conventional OCR rule-based systems to more current and accurate deep learning methods. Although each method aims to tackle the same problem, they confront a few issues with data, procedure, and performance. This section discusses the object detection methods, a few disadvantages, and a semi-supervised deep learning-based solution to overcome the problems.

2.1. Rule-Based Methods

Before the success of deep learning techniques, rule-based solutions for graphical object recognition began with the requirement to recognize tables and figures from the scanned text. There were various rule-based algorithms developed to localize the tables. They identify graphical objects in a document by using predefined rules; Ref. [2] uses the spectator mining technique to detect potential objects in a PDF document. This method assumes that white spaces or lines can differentiate a table cell.
To tackle the problems with varying table layouts, ref. [15] proposes a practical algorithm that can detect tables from various sources, including newspapers and company articles. Unlike most of the rule-based algorithms, which focus on detecting objects from structured documents, ref. [3] employs a correlation-based approach along with dynamic programming to identify tables in noisy handwritten documents. The fundamental problem of rule-based techniques is that they rely on predefined structure and content arrangement in documents.

2.2. Learning-Based Methods

The supervised learning models are the initial and most successful approaches to solving the graphical page object detection. With the popularity of machine learning in object detection, Ref. [16] used an SVM model to identify table cells in a document by creating 26 low-level features for each group of intersecting horizontal and vertical lines. The classifier then determines whether a cell belongs to a table or not.
In [17], the authors verify that the documents can be described by the MXY tree, which is a hierarchical representation. The method identifies tables by locating perpendicular lines and white spaces. Ref. [18] describes a probabilistic graphical model for document analysis. This model incorporates different document structures into Hidden Markov Models. Unsupervised learning algorithms are also proposed to cluster tables and text separately. The domain-independent technique proposed by [19] applies bottom–up clustering to word segments in recognizing tables. Although the learning-based models are accurate and independent compared to rule-based models, they do not generalize to new structures and need a large amount of training data.

2.3. Deep Learning Methods

With the success of Convolution Neural Networks (CNN) in the field of object detection and computer vision, the CNNs are also applied to predict and localize the documents; Ref. [20] proposes a method that first identifies table-like areas using a rule-based approach and applies a CNN on the output to detect potential tables. This method also considers non-visual features such as characters for better localization. To eliminate the dependency of rules and other metadata in training, Ref. [21] introduced DeepDeSRT, a two-stage approach for detection and structure detection, and later proved that this method generalizes to new structures. A Deep Convolution Neural Network (DCNN), which adds fully connected conditional random fields to convolution layers, performs multi-scale reasoning on visual cues for localizing objects. Hashmi et al. [22] made a few modifications to Cascade Mask R-CNN to identify mathematical formulas in document images.
The state-of-the-art object detection networks that depend on the regional proposals are introduced to the field of graphical object detection by [7]. They applied a Region Proposal Network (RPN), which shares full-image convolution features, enabling cost-free region proposals. There are many [23,24,25] region, pixel and connected component-based models proposed to identify textual and non-textual components in a document. The Faster-RCNN and Mask-RCNN detectors are used by [26] to localize and segment objects.
Based on a dynamic programming approach, [27] identifies region proposals in page object detection and [28] outlines and summarizes multiple deep learning approaches for graphical page object detection. The authors from [29] conduct performance analysis on neural networks which recognize tables in document images. The deep learning-based approaches are highly effective with some detectors trained on a particular dataset capable of detecting and localizing graphical structures on unseen new datasets [30]. The major disadvantage of the deep learning-based methods is the hunger for annotated data. These approaches require a lot of characterized data that is hard to obtain, and the annotation labeling process is manual, error-prone, and time-consuming.

2.4. Semi-Supervised Approaches

Due to the bottleneck in the labeling annotation process, the object detection techniques are leveraging unlabeled data via semi-supervised methods. Every day, numerous new documents are created that may or may not be related to current datasets and may represent new datasets. One of the early studies [31] employs semi-supervised learning on fully convolution networks to accomplish MS Lesion Segmentation. Attention-based semi-supervised deep networks proposed by [32] use region-attention to leverage unlabeled training data to segment medical images in an end-to-end fashion. Similarly, Ref. [33] exploits unlabeled endoscopic videos to learn representations of the target domain. The role of semi-supervised deep learning is employed in various fields such as label propagation [34,35], anomaly detection [36], and segmentation [37,38].
In image classification, consistency-based semi-supervised methods improve classification performance by using all the unlabeled input. In these methods, a consistency requirement is used, with multiple modifications of the same image producing comparable results. Ref. [39] modifies the model, Ref. [40] proposes a new regularization method based on virtual adversarial loss, and [41] uses randomized image augmentations. Self-training approaches also known as pseudo-labeling approaches such as [42,43] train an initial classification model on unlabeled data to generate pseudo labels that help refine the initial classifier.
Semi-supervised object detection algorithms, such as image classification, can be consistency [39,44] or pseudo label based [11,45,46]. Some of the semi-supervised approaches are proved to perform better object detection than supervised approaches on the MS COCO image dataset [11,12,14,47,48]. These approaches are multi-staged, where an initial detector is applied to pre-labeled data before generating pseudo labels for the unlabeled data. Finally, the combination of annotated and non-annotated data is trained for accuracy. To improve the performance of these multi-stage detectors, an end-to-end pseudo label approach is proposed by [10], which simultaneously proposes pseudo boxes and performs detection training. The application of this model on the MS COCO dataset to detect and localize images showed significant improvement from the previous supervised and semi-supervised approaches. Hence, the primary objective of this paper is to leverage this novel end-to-end semi-supervised framework on graphical page object detection.

3. Proposed Framework

Semi-supervised object detection algorithms are broadly divided into consistency and pseudo labeled. The pseudo-labeled algorithms are multi-staged and run in two stages where an initial classifier is generated using the labeled data in the first stage, In the next stage, the initial classifier is used to create pseudo labels before updating the initial classifier with the combination of labeled and pseudo-labeled data. The Soft Teacher [10] framework, which is proven to produce supervised-like results in object detection is being used to prove the same for graphical page object detection in a pseudo-labeling framework with few modifications to the training process. Firstly, the framework is made end-to-end, without any initial classifier, and then, weak augmentation is used to generate pseudo labels. To address the issue of low training data, the strongly augmented data along with labeled data are used for detection training.
The framework comprises two models: the “student” model, which is in charge of detection training, and the “teacher” model creates pseudo boxes for unlabeled input. The teacher model is the student model’s exponential moving average (EMA) and learns to generate pseudo labels on weakly augmented unlabeled data. In contrast, the student model is trained on both labeled and strongly augmented unlabeled data to minimize the loss. The teacher model generates two sets of pseudo labels: one for the classification task and the other for detecting bounding boxes. We employ the FixMatch technique [49] of using data augmentation for training two multiple branches.
During the training process shown in Figure 3, the training data are split into different batches, each of which contains a random selection of labeled and unlabeled data. The teacher model creates pseudo labels from weakly augmented unlabeled data, while the student model uses both labeled and strongly augmented data. The pseudo labels generated by the teacher model are exploited as ground truth for strongly augmented unlabeled data to accomplish detection training. In contrast to the traditional classification problem, which determines whether a specific image corresponds to a specific label, the localization problem should identify and localize several entities in the same image. Identifying a particular object can be treated as a classification task, and the localization branch can be handled as a regression task. The framework’s total loss is the weighted sum of supervised and unsupervised losses
L t o t a l = L s u p + α L u n s u p
Here, L s u p and L u n s u p represent the supervised and unsupervised loss, respectively, and if N l a and N u n represent the number of labeled and unlabeled images in a batch with I l a i and I u n i representing a particular labeled and unlabeled images, respectively, the supervised and unsupervised loss can be defined as below
L s u p = 1 N l a i = 1 N l a ( L c l s ( I l a i ) + L r e g ( I l a i ) )
L u n s u p = 1 N u n i = 1 N u n ( L u n s u p c l s ( I u n i ) + L u n s u p r e g ( I u n i ) )
Both the teacher and student model are randomly initialized initially, and during the training, the teacher model, in addition to providing pseudo labels, guides the student model. This effect occurs because the teacher model is a simple EMA of the student. During the training of the teacher model, multiple pseudo boxes are generated, and the Non-Maxima Suppression technique is leveraged to discard a few pseudo boxes. However, there can still be multiple pseudo boxes after the suppression technique. To segregate the available boxes into foreground and background, a threshold is employed, and the predictions with box scores higher than the threshold are used as foreground boxes in training.

3.1. Limitations

Although the above base framework is capable of handling unlabeled data for training, there are two different problems. The first problem is due to the high threshold value for identifying the foreground boxes. Setting a high threshold value will result in higher precision, but the recall drops. Hence, if an IoU is used between the student and teacher-generated boxes for label assignment, most of the foreground boxes will be assigned as background. Another problem is that the localization accuracy and foreground scores are not strongly co-related. Therefore, the localization is inaccurate. These problems are solved in [10] by introducing two techniques, namely the Soft Teacher and Box Jittering, which handle the classification and regression losses of the unsupervised branch.

3.2. Soft Teacher

The Teacher model and pseudo boxes are leveraged to reduce the wrong assignment of foreground boxes to background boxes. The main idea of the Soft Teacher is to assess the reliability of each Student-generated box to be a real background, which is then used to calculate the background loss. Let b i f g and b i b g denote the set of foreground and background boxes; the unsupervised classification loss can be calculated as the following
L u n s u p c l s = 1 N f g i = 1 N f g L c l s ( b i f g , G c l s ) + j = 1 N b g w j L c l s ( b j b g , G c l s )
w j = r j k = 1 N b g r k
The N f g and N b g denote the number of images in b i f g and b i b g , G c l s denotes all the pseudo labels generated by the teacher model. The L c l s is the classification loss and r j denotes the reliability for j-th background box. Ref. [10] proved that the simple background score produced by the teacher model serves as the reliability score, and Figure 4 shows the classification bounding boxes that are generated during the training process.

3.3. Box Jittering

Due to a high threshold in identifying foreground boxes, the boxes generated are not accurate. In order to alleviate the issue, ref. [10] proposes a box-jittering approach in which a random box b i is identified across the real foreground box and is refined with the teacher model multiple times N j i t e r to generate more accurate box coordinates b i *
b i * = r e f i n e j i t e r ( b i )
Hence, N j i t e r number of new box coordinate sets { b i , j * } are generated. For each of the boxes generated, a regression variance is calculated to identify the regression box with high localization accuracy. The box regression variance is defined by
σ i ¯ = 1 4 k = 1 k σ k *
σ k * = σ k 0.5 * ( h ( b t ) + w ( b t ) )
Here, σ k denotes the standard deviation of the k -th coordinate of the refined box coordinate set { b i , j * } , σ k * is the normalized standard deviation and h and w are the height and width of the t -th refined box b t . A smaller box regression variance σ i ¯ indicates better localization, but calculating the box regression variance for all the foreground boxes is highly time and resource-consuming. Hence, it is proposed to calculate it for foreground boxes with a threshold greater than 0.5. Finally, if G r e g are the coordinates of all the regression pseudo labels generated by the teacher model, the unsupervised regression loss can be calculated using
L u n s u p r e g = 1 N f g i = 1 N f g L r e g ( b i f g , G r e g )
Figure 5 shows the regression bounding boxes that are generated after the box jittering approach in the training process.

4. Experiments

4.1. Datasets

The experiments are performed on three datasets: namely, PubLayNet [26], IIIT-AR-13K [50] and DocBank [51], and the following subsections describe the datasets in detail.

4.1.1. Publaynet

The PubLayNet dataset is created by matching the XML representations and content of over 1 million publicly available documents and is primarily used for training the model. This dataset consists of 335,703 training images and 11,245 validation images for training and evaluation. Table 1 shows the distribution of labels in the training dataset. In addition to this, a new subset dataset termed sub-PubLayNet is created, which consists of two labels: Figure and Table. These data are used to cross-validate a PubLayNet trained model on the sub-DocBank and sub-IIIT-AR-13K datasets.

4.1.2. IIIT-AR-13K

The IIIT-AR-13K consists of a set of business documents, mostly annual reports. The dataset contains 9333 training images and 1955 validation images. Table 2 represents the class-wise training dataset distribution. Similar to the PubLayNet, a subset (sub-IIIT-AR-13K) dataset is created with two labels: Figure and Table. This subset dataset is used to train a Soft Teacher model and to validate the model on sub-PubLayNet and sub-DocBank datasets.

4.1.3. Docbank

The DocBank dataset consists of 16 classes. To understand the cross-validated performance of the DocBank trained model on other datasets, a new subset (sub-DocBank) is created, which contains the labels Table and Figure. Table 3 and Table 4 show the total image and annotation count in the subset datasets sub-PubLayNet, sub-IIIT-AR-13K, and sub-DocBank, respectively.

4.2. Evaluation Protocol

The partially labeled data creation follows the STAC [52] setting, where the training data are divided into 1%, 5% and 10%, respectively, and these samples are considered annotated data in these three training settings, respectively. The rest of the data in each of the setting are used as unlabeled data in the training process. For each protocol, STAC provides five different folds, and the final performance is the average of all the five folds. To compare the supervised training models with this semi-supervised framework, the unlabeled data are treated as not useful for the training of supervised models. The mean Average Precision (mAP) on the validation data is evaluated for the comparison of various semi-supervised models.

4.2.1. Precision

The Precision [53] is the fraction of relevant instances (True Positives) among the retrieved instances (True Positives + False Positives).
P r e c i s i o n = T r u e P o s i t i v e s T r u e P o s i t i v e s + F a l s e P o s i t i v e s

4.2.2. Recall

The Recall [53] is the fraction of relevant instances (True Positives) that were retrieved (True Positives + False Negatives).
R e c a l l = T r u e P o s i t i v e s T r u e P o s i t i v e s + F a l s e N e g a t i v e s

4.2.3. F1-Score

The F1-score [53] is the harmonic mean between the Precision and Recall. It is mathematically defined as follows
F 1 - S c o r e = 2 1 P r e c i s i o n + 1 R e c a l l

4.2.4. Intersection over Union (IoU)

The Intersection over Union (IoU) [22], which is also known as the Jaccard index, measures the similarity between finite sample sets. It is defined as the ratio between the size of the intersection and the size of the union of the two sample sets. In the machine learning setting, it predicts the region between the predicted and ground truth region.
IoU ( A , B ) = A r e a o f O v e r l a p r e g i o n A r e a o f U n i o n r e g i o n = A B A B

4.2.5. Average Precision (Ap)

Average Precision is defined as the area under the Precision–Recall curve. In the context of object localization, the average precision is defined for multiple IoUs. For instance, the AP@[0.5] indicates the area under the Precision–Recall curve when the IoU threshold is set to 0.5. In this case, an element is correctly classified if it overlaps with 50 % of the ground truth. In the context of MS COCO [48] evaluation, the Average Precision of the classes is averaged to get the final Average Precision.

4.2.6. Mean Average Precision (Map)

The mean Average Precision (mAP) is calculated by taking the mean of all classes average precision over all the IoU thresholds defined. For MS COCO [48] evaluation, the mAP is averaged over 10 IoU thresholds from 0.50 to 0.95 with a step size of 0.5.

4.3. Implementation Details

For the PubLayNet Partial data (i.e., based on the percentage of annotated data) and the PubLayNet intersection data (i.e., based on the percentage of annotated data for Table and Figure), the implementation considers Faster R-CNN [7] equipped with FPN [54] as the default detection framework for the evaluation. Along with the ImageNet pre-trained ResNet-50 [55] and ResNet-101 [55] backbones on 1%, 5% and 10%, a Swin-T [56] model on 5% and 10% of annotated data is also trained for the comparison. Anchors with five scales and three aspect ratios are used, and 2000 and 1000 region proposals are generated with a non-maximum suppression threshold of 0.7 on both training and inference. Finally, in each training step, 512 proposals are samples from 2000 proposals as box candidates to train RCNN.
All the models are trained for 150,000 iterations on two GPUs with eight images per GPU batch size. The foreground threshold is set to 0.9, and the data sampling ratio, which is the ratio of annotated to non-annotated images in each batch, is set to 0.2 and gradually decreases to 0 in the last 10,000 iterations. For the stochastic gradient descent, the learning rate is set to 0.01 and is divided by ten at 110,000 iterations. To identify the pseudo labels for bounding boxes with high localization reliability, the N j i t e r is set to 10 with a threshold of 0.02. In addition, the same augmentation techniques as in [52] are considered for training.
The hyper-parameters foreground threshold, suppression threshold, and N j i t e r are proved to be optimal by [10] while training the Soft Teacher model on the MS-COCO object detection dataset [48]. Finally, the IIIT-AR-13K is trained only for 50,000 iterations with Faster-RCNN equipped with FPN using ImageNet pre-trained ResNet-50 as the backbone. The learning rate is set to 0.01, and other parameters are similar to the PublayNet trained models, since the IIIT-AR-13K data in all the splits are small (less than 1100 images).

4.4. Results and Discussion

4.4.1. Publaynet

In this section, the results of multiple models trained on the PubLayNet dataset with the proposed framework are compared with the models trained using supervised techniques. The partially annotated PubLayNet models are first compared with the corresponding supervised models. Table 5 shows the comparison of the Faster-RCNN supervised models on the ResNet-50 backbone with different versions of Soft Teacher that are trained using Faster-RCNN and Swin-T transformers. For the detailed comparison, the Faster-RCNN is further trained using two backbones, i.e., ResNet50 and ResNet101. The semi-supervised models showed around 4.9 points, 5.8 points and 5.5 points improvement in the bounding box mean Average Precision when compared under 1 % ,   5 % and 10 % of annotated data, respectively.
Further, a Soft Teacher (Faster-RCNN + ResNet101) model which is trained on 10 % of labeled data for 180,000 iterations with a batch size of 8 is compared with the existing baseline model by [26]. The baseline models are trained using fully labeled PubLayNet data on Faster-RCNN for 180,000 iterations. The results in Table 6 show that the Soft Teacher model outperforms the supervised model’s average precision at (IoU = 0.5) by 2.7% and exhibits a similar mean Average Precision of around 90.0%. Finally, the inference times in terms of FPS of our models are shown in Table 5. Although there is no reference FPS from the earlier methods, we tried to compare inference time with our models and found that our model equipped with Faster R-CNN and ResNet50 outperformed ResNet101 and Swin-T.
In addition to the total mAP, the label-wise AP comparison is shown in Table 7. The graphical objects Table, Figure and List outperform the baseline model by 5.4%, 1.2% and 3.2%, respectively. Moreover, Table 8 as well as Figure 6 and Figure 7 explain and show the detailed metrics of Precision, Recall and F-1 score at various IoU thresholds. The results for the Soft Teacher model trained on 10% labeled PubLayNet data are shown in Figure 1. Additionally, Figure 8 portrays the qualitative results, and Figure 8 (right) shows the model figure and table resulting in False Positives.

4.4.2. IIIT-AR-13K

The Supervised Faster RCNN and Soft Teacher models trained on IIIT-AR-13K with ResNet50 backbone are compared under three different scenarios of labeled data: 1 % ,   5 % , and 10 % . Table 9 shows that the proposed framework performs better than the Faster RCNN model by 6.4%, 1% and 5.1%, respectively. Additionally, a YOLO F [57] model trained on the complete data is compared with the Soft Teacher model that is trained on 10% labeled data. The Soft Teacher model showed a 3.7% mAP improvement, and it outperforms the YOLO F model in detecting natural images, logos, and signatures. Table 10 and Table 11 shows the detailed comparison of YOLO F and Soft Teacher models along with the inference times. Figure 2 show the qualitative results of two IIIT-AR-13K, and Figure 9 displays a few samples of True Positives and False Positives documents trained using the Soft Teacher model. Finally, Table 12 and Figure 10 and Figure 11 visually explain the Precision, Recall and F1 scores of a soft teacher model trained on 10 % of IIIT-AR-13K labeled dataset.

4.4.3. Cross-Validation

The datasets PubLayNet, IIIT-AR-13K, and DocBank are the most common datasets used for graphical objected detection, and the three datasets possess objects which are different from each other. To validate a model trained on one dataset with another dataset, there is a need to train models with the objects common to all the datasets. Table and Figure are the common classes among datasets, and they are different from each other; for instance, the IIIT-AR-13K dataset is based on business documents, the PubLayNet is created from publicly available docs, and DocBank is created in a way to ease layout analysis. The primary idea is to identify how similar the datasets are in terms of the graphical objects. The new datasets sub-PubLayNet, sub-IIIT-AR-13K and sub-DocBank that contain Table and Figure annotations are created, and multiple Soft Teacher models are trained with 5 and 10% of sub-PubLayNet, sub-IIIT-AR-13K, and sub-DocBank.
The results of models trained on sub-PubLayNet and cross-validated on sub-IIIT-AR-13K and sub-DocBank are shown in Table 13. The Soft Teacher model trained on 10% of labeled trained data on Faster-RCNN with ResNet101 showed the highest mAP of 95.1%. The validation performed on other datasets proves that sub-DocBank is more similar to sub-PubLayNet than the sub-IIIT-AR-13K. Table 14 also proves that sub-DocBank is similar to sub-PubLayNet, and interestingly, the mAP on sub-PubLayNet is higher than that of sub-DocBank, even though the model is trained completely trained on sub-Docbank. From Table 15, it is shown the models trained on IIIT-AR-13K perform badly on the other two datasets. This is because the dataset has mostly business reports, unlike the other two datasets. Figure 12 depicts the qualitative evaluation of the table and figure on the sub-PubLayNet dataset by displaying samples of True Positives and False Positives.

5. Conclusions and Future Work

This paper presents the semi-supervised framework for graphical page object detection by employing a multi-stage semi-supervised technique known as Soft Teacher for the detection of graphical objects. In addition to operating on minimal data, this approach combines the complex process of pseudo labeling into a single pipeline. As the framework generates pseudo labels simultaneously with detection training, the flywheel effect occurs, which means one model constantly refines the pseudo boxes generated by the other model during the training process. This framework refines the classification and regression pseudo boxes using two different techniques, Soft Teacher and Box Jittering. These two processes work independently to render accurate classification and bounding box predictions.
This approach outperforms the supervised models in the labeling ratios (1%, 5%, and 10%) of PubLayNet and IIIT-AR-13K training data. Additionally, the Soft Teacher models trained on 10% PubLayNet labeled data performed similarly to the existing supervised baseline, whereas the Soft Teacher model on IIIT-AR-13K outperformed the YOLO F model. Finally, new datasets are created to perform inter-dataset cross-validation and prove that PubLayNet and DocBank datasets are more related compared to IIIT-AR-13K.
In future work, we plan to use powerful two-stage detectors such as Cascade-RCNN for better detection results compared to the current Faster-RCNN. Furthermore, we will examine the effect of the percentage of labeled data on the final performance and try to design robust models relying on minimal annotated data. Finally, the aforementioned results show that the proposed framework with Faster R-CNN performs similarly to Faster R-CNN; nevertheless, we must show that this statement holds for multiple backbones. Finally, to establish a baseline for semi-supervised graphical page object detection, we want to assess various semi-supervised frameworks, including consistency and pseudo labeling.

Author Contributions

Writing—original draft preparation, G.K., K.A.H., M.Z.A.; writing—review and editing, K.A.H., M.Z.A., M.L.; supervision and project administration, A.P., D.S. All authors have read and agreed to the submitted version of the manuscript.

Funding

The work leading to this publication has been partially funded by the European project INFINITY under Grant Agreement ID 883293.

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

Not applicable.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Orosz, T.; Vági, R.; Csányi, G.M.; Nagy, D.; Üveges, I.; Vadász, J.P.; Megyeri, A. Evaluating Human versus Machine Learning Performance in a LegalTech Problem. Appl. Sci. 2021, 12, 297. [Google Scholar] [CrossRef]
  2. Fang, J.; Gao, L.; Bai, K.; Qiu, R.; Tao, X.; Tang, Z. A Table Detection Method for Multipage PDF Documents via Visual Seperators and Tabular Structures. In Proceedings of the 2011 International Conference on Document Analysis and Recognition, ICDAR 2011, Beijing, China, 18–21 September 2011; IEEE Computer Society: Piscataway, NJ, USA, 2011; pp. 779–783. [Google Scholar] [CrossRef]
  3. Chen, J.; Lopresti, D.P. Table Detection in Noisy Off-line Handwritten Documents. In Proceedings of the 2011 International Conference on Document Analysis and Recognition, ICDAR 2011, Beijing, China, 18–21 September 2011; IEEE Computer Society: Piscataway, NJ, USA, 2011; pp. 399–403. [Google Scholar] [CrossRef]
  4. Hashmi, K.A.; Ponnappa, R.B.; Bukhari, S.S.; Jenckel, M.; Dengel, A. Feedback learning: Automating the process of correcting and completing the extracted information. In Proceedings of the 2019 International Conference on Document Analysis and Recognition Workshops (ICDARW), Sydney, Australia, 22–25 September 2019; IEEE: Piscataway, NJ, USA, 2019; Volume 5, pp. 116–121. [Google Scholar]
  5. Saha, R.; Mondal, A.; Jawahar, C.V. Graphical Object Detection in Document Images. In Proceedings of the 2019 International Conference on Document Analysis and Recognition, ICDAR 2019, Sydney, Australia, 20–25 September 2019; IEEE: Piscataway, NJ, USA, 2019; pp. 51–58. [Google Scholar] [CrossRef]
  6. Girshick, R.B. Fast R-CNN. In Proceedings of the 2015 IEEE International Conference on Computer Vision, ICCV 2015, Santiago, Chile, 7–13 December 2015; IEEE Computer Society: Piscataway, NJ, USA, 2015; pp. 1440–1448. [Google Scholar]
  7. Ren, S.; He, K.; Girshick, R.B.; Sun, J. Faster R-CNN: Towards Real-Time Object Detection with Region Proposal Networks. IEEE Trans. Pattern Anal. Mach. Intell. 2017, 39, 1137–1149. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  8. Redmon, J.; Farhadi, A. YOLO9000: Better, Faster, Stronger. arXiv 2016, arXiv:1612.08242. [Google Scholar]
  9. He, K.; Gkioxari, G.; Dollár, P.; Girshick, R.B. Mask R-CNN. In Proceedings of the IEEE International Conference on Computer Vision, Honolulu, HI, USA, 21–26 July 2017. [Google Scholar]
  10. Xu, M.; Zhang, Z.; Hu, H.; Wang, J.; Wang, L.; Wei, F.; Bai, X.; Liu, Z. End-to-End Semi-Supervised Object Detection with Soft Teacher. In Proceedings of the IEEE/CVF International Conference on Computer Vision, Montreal, QC, Canada, 11–17 October 2021. [Google Scholar]
  11. Wang, K.; Yan, X.; Zhang, D.; Zhang, L.; Lin, L. Towards Human-Machine Cooperation: Self-supervised Sample Mining for Object Detection. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Salt Lake City, UT, USA, 18–23 June 2018. [Google Scholar]
  12. Tang, P.; Ramaiah, C.; Xu, R.; Xiong, C. Proposal Learning for Semi-Supervised Object Detection. In Proceedings of the IEEE/CVF Winter Conference on Applications of Computer Vision, Montreal, QC, Canada, 11–17 October 2021. [Google Scholar]
  13. Rhee, P.; Erdenee, E.; Kyun, S.D.; Ahmed, M.U.; Jin, S. Active and semi-supervised learning for object detection with imperfect data. Cogn. Syst. Res. 2017, 45, 109–123. [Google Scholar] [CrossRef]
  14. Xie, Q.; Dai, Z.; Hovy, E.H.; Luong, T.; Le, Q. Unsupervised Data Augmentation for Consistency Training. In Proceedings of the Neural Information Processing Systems 33: Annual Conference on Neural Information Processing Systems 2020, NeurIPS 2020, Virtual, 6–12 December 2020; Available online: https://proceedings.neurips.cc/paper/2020/file/44feb0096faa8326192570788b38c1d1-Paper.pdf (accessed on 28 April 2022).
  15. Shafait, F.; Smith, R. Table detection in heterogeneous documents. In Proceedings of the The Ninth IAPR International Workshop on Document Analysis Systems, DAS 2010, Boston, MA, USA, 9–11 June 2010; Doermann, D.S., Govindaraju, V., Lopresti, D.P., Natarajan, P., Eds.; ACM: New York, NY, USA, 2010; pp. 65–72. [Google Scholar] [CrossRef] [Green Version]
  16. Kasar, T.; Barlas, P.; Adam, S.; Chatelain, C.; Paquet, T. Learning to Detect Tables in Scanned Document Images Using Line Information. In Proceedings of the 12th International Conference on Document Analysis and Recognition, ICDAR 2013, Washington, DC, USA, 25–28 August 2013; IEEE Computer Society: Piscataway, NJ, USA, 2013; pp. 1185–1189. [Google Scholar] [CrossRef] [Green Version]
  17. Cesarini, F.; Marinai, S.; Sarti, L.; Soda, G. Trainable Table Location in Document Images. In Proceedings of the 16th International Conference on Pattern Recognition, ICPR 2002, Quebec, QC, Canada, 11–15 August 2002; IEEE Computer Society: Piscataway, NJ, USA, 2002; pp. 236–240. [Google Scholar] [CrossRef]
  18. e Silva, A.C. Learning Rich Hidden Markov Models in Document Analysis: Table Location. In Proceedings of the 10th International Conference on Document Analysis and Recognition, ICDAR 2009, Barcelona, Spain, 26–29 July 2009; IEEE Computer Society: Piscataway, NJ, USA, 2009; pp. 843–847. [Google Scholar] [CrossRef]
  19. Kieninger, T.; Dengel, A. The T-Recs Table Recognition and Analysis System. In Proceedings of the Document Analysis Systems: Theory and Practice, Third IAPR Workshop, DAS’98, Nagano, Japan, 4–6 November 1998; Lee, S., Nakano, Y., Eds.; Springer: Berlin/Heidelberg, Germany, 1998; Volume 1655, pp. 255–269. [Google Scholar] [CrossRef] [Green Version]
  20. Hao, L.; Gao, L.; Yi, X.; Tang, Z. A Table Detection Method for PDF Documents Based on Convolutional Neural Networks. In Proceedings of the 12th IAPR Workshop on Document Analysis Systems, DAS 2016, Santorini, Greece, 11–14 April 2016; IEEE Computer Society: Piscataway, NJ, USA, 2016; pp. 287–292. [Google Scholar] [CrossRef]
  21. Schreiber, S.; Agne, S.; Wolf, I.; Dengel, A.; Ahmed, S. DeepDeSRT: Deep Learning for Detection and Structure Recognition of Tables in Document Images. In Proceedings of the 14th IAPR International Conference on Document Analysis and Recognition, ICDAR 2017, Kyoto, Japan, 9–15 November 2017; IEEE: Piscataway, NJ, USA, 2017; pp. 1162–1167. [Google Scholar] [CrossRef]
  22. Hashmi, K.A.; Pagani, A.; Liwicki, M.; Stricker, D.; Afzal, M.Z. CasTabDetectoRS: Cascade Network for Table Detection in Document Images with Recursive Feature Pyramid and Switchable Atrous Convolution. J. Imaging 2021, 7, 214. [Google Scholar] [CrossRef] [PubMed]
  23. Moll, M.A.; Baird, H.S. Segmentation-based retrieval of document images from diverse collections. In Proceedings of the Document Recognition and Retrieval XV, part of the IS&T-SPIE Electronic Imaging Symposium, San Jose, CA, USA, 29–31 January 2008; Yanikoglu, B.A., Berkner, K., Eds.; SPIE: Bellingham, WA, USA, 2008; Volume 6815, p. 68150L. [Google Scholar] [CrossRef] [Green Version]
  24. Nayef, N.; Ogier, J. Text zone classification using unsupervised feature learning. In Proceedings of the 13th International Conference on Document Analysis and Recognition, ICDAR 2015, Nancy, France, 23–26 August 2015; IEEE Computer Society: Piscataway, NJ, USA, 2015; pp. 776–780. [Google Scholar] [CrossRef]
  25. Tombre, K.; Tabbone, S.; Pélissier, L.; Lamiroy, B.; Dosch, P. Text/Graphics Separation Revisited. In Proceedings of the Document Analysis Systems V, 5th International Workshop, DAS 2002, Princeton, NJ, USA, 19–21 August 2002; Lopresti, D.P., Hu, J., Kashi, R.S., Eds.; Springer: Berlin/Heidelberg, Germany, 2002; Volume 2423, pp. 200–211. [Google Scholar] [CrossRef] [Green Version]
  26. Zhong, X.; Tang, J.; Jimeno-Yepes, A. PubLayNet: Largest dataset ever for document layout analysis. In Proceedings of the 2019 International Conference on Document Analysis and Recognition (ICDAR), Sydney, NSW, Australia, 20–25 September 2019. [Google Scholar]
  27. Zach, C.; Sánchez, A.P.; Pham, M. A dynamic programming approach for fast and robust object pose recognition from range images. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, CVPR 2015, Boston, MA, USA, 7–12 June 2015; IEEE Computer Society: Piscataway, NJ, USA, 2015; pp. 196–203. [Google Scholar] [CrossRef]
  28. Bhatt, J.; Hashmi, K.A.A.; Afzal, M.Z.; Stricker, D. A Survey of Graphical Page Object Detection with Deep Neural Networks. Appl. Sci. 2021, 11, 5344. [Google Scholar] [CrossRef]
  29. Hashmi, K.A.; Liwicki, M.; Stricker, D.; Afzal, M.A.; Afzal, M.A.; Afzal, M.Z. Current Status and Performance Analysis of Table Recognition in Document Images with Deep Neural Networks. IEEE Access 2021, 9, 87663–87685. [Google Scholar] [CrossRef]
  30. Nazir, D.; Hashmi, K.A.; Pagani, A.; Liwicki, M.; Stricker, D.; Afzal, M.Z. HybridTabNet: Towards better table detection in scanned document images. Appl. Sci. 2021, 11, 8396. [Google Scholar] [CrossRef]
  31. Baur, C.; Albarqouni, S.; Navab, N. Semi-supervised Deep Learning for Fully Convolutional Networks. In Proceedings of the Medical Image Computing and Computer Assisted Intervention - MICCAI 2017-20th International Conference, Quebec, QC, Canada, 11–13 September 1 2017; Descoteaux, M., Maier-Hein, L., Franz, A.M., Jannin, P., Collins, D.L., Duchesne, S., Eds.; Springer: Berlin/Heidelberg, Germany, 2017; Volume 10435, pp. 311–319. [Google Scholar] [CrossRef] [Green Version]
  32. Nie, D.; Gao, Y.; Wang, L.; Shen, D. ASDNet: Attention Based Semi-supervised Deep Networks for Medical Image Segmentation. In Proceedings of the Medical Image Computing and Computer Assisted Intervention - MICCAI 2018-21st International Conference, Granada, Spain, 16–20 September 2018; Frangi, A.F., Schnabel, J.A., Davatzikos, C., Alberola-López, C., Fichtinger, G., Eds.; Springer: Berlin/Heidelberg, Germany, 2018; Volume 11073, pp. 370–378. [Google Scholar] [CrossRef]
  33. Roß, T.; Zimmerer, D.; Vemuri, A.S.; Isensee, F.; Bodenstedt, S.; Both, F.; Kessler, P.; Wagner, M.; Müller, B.; Kenngott, H.; et al. Exploiting the potential of unlabeled endoscopic video data with self-supervised learning. Int. J. Comput. Assist. Radiol. Surg. 2018, 13, 925–933. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  34. Iscen, A.; Tolias, G.; Avrithis, Y.; Chum, O. Label Propagation for Deep Semi-supervised Learning. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Long Beach, CA, USA, 16–20 June 2019. [Google Scholar]
  35. Gan, Y.; Zhu, H.; Guo, W.; Xu, G.; Zou, G. Deep semi-supervised learning with contrastive learning and partial label propagation for image data. Knowl. Based Syst. 2022, 245, 108602. [Google Scholar] [CrossRef]
  36. Kiran, B.R.; Thomas, D.M.; Parakkal, R. An overview of deep learning based methods for unsupervised and semi-supervised anomaly detection in videos. J. Imaging 2018, 4, 36. [Google Scholar] [CrossRef] [Green Version]
  37. Papandreou, G.; Chen, L.; Murphy, K.P.; Yuille, A.L. Weakly-and Semi-Supervised Learning of a Deep Convolutional Network for Semantic Image Segmentation. In Proceedings of the 2015 IEEE International Conference on Computer Vision, ICCV 2015, Santiago, Chile, 7–13 December 2015; IEEE Computer Society: Piscataway, NJ, USA, 2015; pp. 1742–1750. [Google Scholar] [CrossRef]
  38. Olsson, V.; Tranheden, W.; Pinto, J.; Svensson, L. ClassMix: Segmentation-Based Data Augmentation for Semi-Supervised Learning. In Proceedings of the IEEE/CVF Winter Conference on Applications of Computer Vision, Montreal, QC, Canada, 11–17 October 2021. [Google Scholar]
  39. Jeong, J.; Lee, S.; Kim, J.; Kwak, N. Consistency-based Semi-supervised Learning for Object detection. In Proceedings of the Neural Information Processing Systems 32: Annual Conference on Neural Information Processing Systems 2019, NeurIPS 2019, Vancouver, BC, Canada, 8–14 December 2019; Wallach, H.M., Larochelle, H., Beygelzimer, A., d’Alché-Buc, F., Fox, E.B., Garnett, R., Eds.; 2019; pp. 10758–10767. Available online: https://papers.nips.cc/paper/2019/hash/d0f4dae80c3d0277922f8371d5827292-Abstract.html (accessed on 28 April 2022).
  40. Miyato, T.; Maeda, S.; Koyama, M.; Ishii, S. Virtual Adversarial Training: A Regularization Method for Supervised and Semi-supervised Learning. IEEE Trans. Pattern Anal. Mach. Intell. 2018, 41, 1979–1993. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  41. Sajjadi, M.; Javanmardi, M.; Tasdizen, T. Regularization With Stochastic Transformations and Perturbations for Deep Semi-Supervised Learning. 2016. Available online: https://proceedings.neurips.cc/paper/2016/file/30ef30b64204a3088a26bc2e6ecf7602-Paper.pdf (accessed on 28 April 2022).
  42. Grandvalet, Y.; Bengio, Y. Semi-supervised Learning by Entropy Minimization. In Proceedings of the Neural Information Processing Systems 17 Neural Information Processing Systems, NIPS 2004, Vancouver, BC, Canada, 13–18 December 2004; pp. 529–536. [Google Scholar]
  43. Berthelot, D.; Carlini, N.; Goodfellow, I.J.; Papernot, N.; Oliver, A.; Raffel, C. MixMatch: A Holistic Approach to Semi-Supervised Learning. 2019. Available online: https://proceedings.neurips.cc/paper/2019/file/1cd138d0499a68f4bb72bee04bbec2d7-Paper.pdf (accessed on 28 April 2022).
  44. Jeong, J.; Verma, V.; Hyun, M.; Kannala, J.; Kwak, N. Interpolation-based semi-supervised learning for object detection. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Seattle, WA, USA, 13–19 June 2020. [Google Scholar]
  45. Radosavovic, I.; Dollár, P.; Girshick, R.B.; Gkioxari, G.; He, K. Data Distillation: Towards Omni-Supervised Learning. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Honolulu, Hawaii, USA, 21–26 July 2017. [Google Scholar]
  46. Yang, Q.; Wei, X.; Wang, B.; Hua, X.; Zhang, L. Interactive Self-Training With Mean Teachers for Semi-Supervised Object Detection. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, CVPR 2021, Virtual, 19–25 June 2021; IEEE: Piscataway, NJ, USA, 2021; pp. 5941–5950. [Google Scholar]
  47. Tang, Y.; Chen, W.; Luo, Y.; Zhang, Y. Humble Teachers Teach Better Students for Semi-Supervised Object Detection. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Nashville, TN, USA, 20–25 June 2021. [Google Scholar]
  48. Lin, T.; Maire, M.; Belongie, S.J.; Bourdev, L.D.; Girshick, R.B.; Hays, J.; Perona, P.; Ramanan, D.; Dollár, P.; Zitnick, C.L. Microsoft COCO: Common Objects in Context. In Proceedings of the European Conference on Computer Vision, Zurich, Switzerland, 6—12 September 2014. [Google Scholar]
  49. Sohn, K.; Berthelot, D.; Li, C.; Zhang, Z.; Carlini, N.; Cubuk, E.D.; Kurakin, A.; Zhang, H.; Raffel, C. FixMatch: Simplifying Semi-Supervised Learning with Consistency and Confidence. Adv. Neural Inf. Process. Syst. 2020, 33, 596–608. [Google Scholar]
  50. Mondal, A.; Lipps, P.; Jawahar, C.V. IIIT-AR-13K: A New Dataset for Graphical Object Detection in Documents. In Proceedings of the International Workshop on Document Analysis Systems, Wuhan, China, 26–29 July 2020. [Google Scholar]
  51. Li, M.; Xu, Y.; Cui, L.; Huang, S.; Wei, F.; Li, Z.; Zhou, M. DocBank: A Benchmark Dataset for Document Layout Analysis. arXiv 2020, arXiv:2006.01038. [Google Scholar]
  52. Sohn, K.; Zhang, Z.; Li, C.; Zhang, H.; Lee, C.; Pfister, T. A Simple Semi-Supervised Learning Framework for Object Detection. arXiv 2020, arXiv:2005.04757. [Google Scholar]
  53. Powers, D.M.W. Evaluation: From precision, recall and F-measure to ROC, informedness, markedness and correlation. arXiv 2020, arXiv:2010.16061. [Google Scholar]
  54. Lin, T.; Dollár, P.; Girshick, R.B.; He, K.; Hariharan, B.; Belongie, S.J. Feature Pyramid Networks for Object Detection. In Proceedings of the 2017 IEEE Conference on Computer Vision and Pattern Recognition, CVPR 2017, Honolulu, HI, USA, 21–26 July 2017; IEEE Computer Society: Piscataway, NJ, USA, 2017; pp. 936–944. [Google Scholar] [CrossRef] [Green Version]
  55. He, K.; Zhang, X.; Ren, S.; Sun, J. Deep Residual Learning for Image Recognition. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Boston, MA, USA, 7–12 June 2015. [Google Scholar]
  56. Liu, Z.; Lin, Y.; Cao, Y.; Hu, H.; Wei, Y.; Zhang, Z.; Lin, S.; Guo, B. Swin Transformer: Hierarchical Vision Transformer using Shifted Windows. In Proceedings of the IEEE/CVF International Conference on Computer Vision, Montreal, QC, Canada, 11–17 October 2021. [Google Scholar]
  57. Nguyen, P.; Ngo, L.; Truong, T.; Nguyen, T.T.; Vo, N.D.; Nguyen, K. Page Object Detection with YOLOF. In Proceedings of the 2021 8th NAFOSTED Conference on Information and Computer Science (NICS), Hanoi, Vietnam, 21–22 December 2021; pp. 205–210. [Google Scholar] [CrossRef]
Figure 1. Left: graphical objects Figure, Text and Title. Right: graphical objects Table, List Text and Title. Both the images are generated by the Soft Teacher model, which is trained on 10% of the labeled PubLayNet data.
Figure 1. Left: graphical objects Figure, Text and Title. Right: graphical objects Table, List Text and Title. Both the images are generated by the Soft Teacher model, which is trained on 10% of the labeled PubLayNet data.
Futureinternet 14 00176 g001
Figure 2. Left: graphical objects Table and Figure. Right: graphical objects Natural image and Signature. Both the images are generated by the the Soft Teacher model, which is trained on 5% of the labeled IIIT-AR-13K data.
Figure 2. Left: graphical objects Table and Figure. Right: graphical objects Natural image and Signature. Both the images are generated by the the Soft Teacher model, which is trained on 5% of the labeled IIIT-AR-13K data.
Futureinternet 14 00176 g002
Figure 3. The image represents the Soft Teacher process: (1) The training data are divided into labeled and unlabeled data. (2) Weak and strong augmentation are applied to unlabeled data. (3) The Teacher model uses the weakly augmented data to generate two sets of pseudo boxes: one for classification and the other for bounding box regression. (4) The Student model takes both the labeled and strongly augmented data with pseudo labels to perform detection training. (5) The Teacher model is an EMA of the Student model.
Figure 3. The image represents the Soft Teacher process: (1) The training data are divided into labeled and unlabeled data. (2) Weak and strong augmentation are applied to unlabeled data. (3) The Teacher model uses the weakly augmented data to generate two sets of pseudo boxes: one for classification and the other for bounding box regression. (4) The Student model takes both the labeled and strongly augmented data with pseudo labels to perform detection training. (5) The Teacher model is an EMA of the Student model.
Futureinternet 14 00176 g003
Figure 4. Left: Augmentation with scale, shift, and color change. Right: Augmentation with rotation. The images represent the classification pseudo boxes on two different augmented images. They are generated by the teacher model during the training process of the Soft Teacher model on 10% labeled data.
Figure 4. Left: Augmentation with scale, shift, and color change. Right: Augmentation with rotation. The images represent the classification pseudo boxes on two different augmented images. They are generated by the teacher model during the training process of the Soft Teacher model on 10% labeled data.
Futureinternet 14 00176 g004
Figure 5. Left: Augmentation with rotation and blur. Right: Augmentation with scale and shift. The images represent the regression pseudo boxes on two different augmented images. They are generated by the Teacher model during the training process of the Soft Teacher model on 10% labeled data.
Figure 5. Left: Augmentation with rotation and blur. Right: Augmentation with scale and shift. The images represent the regression pseudo boxes on two different augmented images. They are generated by the Teacher model during the training process of the Soft Teacher model on 10% labeled data.
Futureinternet 14 00176 g005
Figure 6. A visual representation of Precision, Recall and F1-Score at multiple IoUs. Our model is trained on 10% PubLayNet labeled data with Faster R-CNN and ResNet101.
Figure 6. A visual representation of Precision, Recall and F1-Score at multiple IoUs. Our model is trained on 10% PubLayNet labeled data with Faster R-CNN and ResNet101.
Futureinternet 14 00176 g006
Figure 7. A label-wise visual representation of Precision at multiple IoUs. Our model is trained on 10% PubLayNet labeled data with Faster R-CNN and ResNet101.
Figure 7. A label-wise visual representation of Precision at multiple IoUs. Our model is trained on 10% PubLayNet labeled data with Faster R-CNN and ResNet101.
Futureinternet 14 00176 g007
Figure 8. The results of our model with Faster R-CNN and ResNet101 that is trained on 10% of PubLayNet labeled data. Blue represents True Positives and Red represents False Positives. In this figure, (left) represents a couple of samples with True Positives, (right) depicts True Positives and False Positives.
Figure 8. The results of our model with Faster R-CNN and ResNet101 that is trained on 10% of PubLayNet labeled data. Blue represents True Positives and Red represents False Positives. In this figure, (left) represents a couple of samples with True Positives, (right) depicts True Positives and False Positives.
Futureinternet 14 00176 g008
Figure 9. The results of our model with Faster R-CNN and ResNet50 that is trained on 10% of IIIT-AR-13K labeled data. Blue represents True Positives and Red represents False Positives. In this figure, (left) represents a couple of samples with True Positives, (right) depicts True Positive and False Positive.
Figure 9. The results of our model with Faster R-CNN and ResNet50 that is trained on 10% of IIIT-AR-13K labeled data. Blue represents True Positives and Red represents False Positives. In this figure, (left) represents a couple of samples with True Positives, (right) depicts True Positive and False Positive.
Futureinternet 14 00176 g009
Figure 10. A visual representation of Precision, Recall and F1-Score at multiple IoUs. The model is trained on 10% IIIT-AR-13K labeled data and is equipped with Faster R-CNN and ResNet50.
Figure 10. A visual representation of Precision, Recall and F1-Score at multiple IoUs. The model is trained on 10% IIIT-AR-13K labeled data and is equipped with Faster R-CNN and ResNet50.
Futureinternet 14 00176 g010
Figure 11. A class-wise visual representation of precision at multiple IoUs. The model is trained on 10% IIIT-AR-13K labeled data and is equipped with Faster R-CNN and ResNet50.
Figure 11. A class-wise visual representation of precision at multiple IoUs. The model is trained on 10% IIIT-AR-13K labeled data and is equipped with Faster R-CNN and ResNet50.
Futureinternet 14 00176 g011
Figure 12. The results of out model equipped with Faster R-CNN and ResNet101 and trained on 5% of sub-PubLayNet labeled data. Blue represents True Positives and Red represents False Positives. In this figure, (left) represents a sub-DocBank sample with a True Positive and False Positives, (right) shows True Positives and False Positives on a sub-IIIT-AR-13K sample.
Figure 12. The results of out model equipped with Faster R-CNN and ResNet101 and trained on 5% of sub-PubLayNet labeled data. Blue represents True Positives and Red represents False Positives. In this figure, (left) represents a sub-DocBank sample with a True Positive and False Positives, (right) shows True Positives and False Positives on a sub-IIIT-AR-13K sample.
Futureinternet 14 00176 g012
Table 1. Label-wise summary of PubLayNet training dataset.
Table 1. Label-wise summary of PubLayNet training dataset.
PubLayNetNumber of ImagesNumber of Annotations
Table86,460102,514
Figure91,968109,292
Text334,5482,343,356
Title255,731627,125
List53,04980,759
Total335,7033,263,046
Table 2. Label-wise summary of IIIT-AR-13K training dataset.
Table 2. Label-wise summary of IIIT-AR-13K training dataset.
IIIT-AR-13KNumber of ImagesNumber of Annotations
Table690311,163
Figure12932004
Natural Image12581987
Logo165379
Signature208420
Total933315,953
Table 3. Actual datasets are used to create subset datasets that only consist of the Table and Figure annotations. The table represents the distribution of the Table and Figure images in the training datasets of sub-PubLayNet, sub-IIIT-AR-13K and sub-DOCBANK.
Table 3. Actual datasets are used to create subset datasets that only consist of the Table and Figure annotations. The table represents the distribution of the Table and Figure images in the training datasets of sub-PubLayNet, sub-IIIT-AR-13K and sub-DOCBANK.
DatasetTableFigureTotal
sub-PubLayNet86,46091,968102,514
sub-IIIT-AR-13K690312937837
sub-DocBank19,52889,612103,285
Table 4. The table represents the distribution of the Table and Figure annotations in the three subset training datasets.
Table 4. The table represents the distribution of the Table and Figure annotations in the three subset training datasets.
DatasetTablesFiguresTotal
sub-PubLayNet102,514109,292211,806
sub-IIIT-AR-13K22224812703
sub-DocBank25,991128,312154,303
Table 5. Performance mAP (0.5:0.95) comparison of supervised and ours models trained on 1 % ,   5 % and 10 % of PubLayNet dataset.
Table 5. Performance mAP (0.5:0.95) comparison of supervised and ours models trained on 1 % ,   5 % and 10 % of PubLayNet dataset.
TechniqueDetector + Backbone1%5%10%FPS
SupervisedFaster R-CNN + ResNet5082.583.383.412.3
OursFaster R-CNN + ResNet5084.987.287.316.7
Faster R-CNN + ResNet10187.488.288.915.8
Swin-T88.389.188.614.9
Table 6. Performance comparison of Faster R-CNN baseline model trained on fully labeled data with our semi-supervised model which is trained on 10% labeled PubLayNet dataset. The model uses Faster R-CNN and ResNet101.
Table 6. Performance comparison of Faster R-CNN baseline model trained on fully labeled data with our semi-supervised model which is trained on 10% labeled PubLayNet dataset. The model uses Faster R-CNN and ResNet101.
TechniqueDetector[email protected][email protected]mAPFPS
Zhong et al. [26]Supervised + Faster R-CNN93.791.190.2-
OursFaster R-CNN
(10% labeled data)
96.493.890.015.8
Table 7. Class-wise performance AP (0.5:0.95) comparison of Faster R-CNN baseline model trained on fully labeled data with our semi-supervised model which is trained on 10% labeled PubLayNet dataset. The model uses Faster R-CNN and ResNet101.
Table 7. Class-wise performance AP (0.5:0.95) comparison of Faster R-CNN baseline model trained on fully labeled data with our semi-supervised model which is trained on 10% labeled PubLayNet dataset. The model uses Faster R-CNN and ResNet101.
TechniqueDetectorTableFigureTextTitleList
Zhong et al. [26]Supervised + Faster R-CNN90.293.791.082.688.3
OursFaster R-CNN
(10% labeled data)
96.694.985.581.491.5
Table 8. A detailed representation of Precision, Recall and F1-Score at multiple IoUs. Our model is trained on 10% PubLayNet labeled data with Faster R-CNN and ResNet101.
Table 8. A detailed representation of Precision, Recall and F1-Score at multiple IoUs. Our model is trained on 10% PubLayNet labeled data with Faster R-CNN and ResNet101.
MethodIoUPrecisionRecallF1-Score
Semi-Supervised
Soft Teacher
Faster R-CNN + ResNet101
10% labeled
0.500.970.960.94
0.600.960.960.93
0.700.950.950.93
0.800.930.920.91
0.900.850.850.83
Table 9. Performance comparison in terms of mAP (0.5:0.95) between supervised and our semi-supervised model trained on 1 % , 5 % and 10 % of the IIIT-AR-13K dataset. Both the models use Faster-RCNN with ResNet 50 backbone.
Table 9. Performance comparison in terms of mAP (0.5:0.95) between supervised and our semi-supervised model trained on 1 % , 5 % and 10 % of the IIIT-AR-13K dataset. Both the models use Faster-RCNN with ResNet 50 backbone.
TechniqueDetector + Baseline1%5%10%FPS
SupervisedFaster R-CNN + ResNet5035.849.757.412.3
OursFaster R-CNN + ResNet5042.251.863.316.8
Table 10. Performance comparison of YOLO F baseline model trained on fully labeled data with our model which is equipped with Faster R-CNN and ResNet50 and is trained on 10% labeled IIIT-AR-13K dataset.
Table 10. Performance comparison of YOLO F baseline model trained on fully labeled data with our model which is equipped with Faster R-CNN and ResNet50 and is trained on 10% labeled IIIT-AR-13K dataset.
TechniqueDetector[email protected][email protected]mAPFPS
Nguyen et al. [57]Supervised + YOLO F81.264.958.8-
OursFaster R-CNN
(10% labeled data)
87.867.063.316.8
Table 11. Class-wise performance AP (0.5:0.95) comparison of YOLO F baseline model trained using fully labeled data with our model, which is equipped with Faster-RCNN and ResNet50 and is trained on 10% labeled data.
Table 11. Class-wise performance AP (0.5:0.95) comparison of YOLO F baseline model trained using fully labeled data with our model, which is equipped with Faster-RCNN and ResNet50 and is trained on 10% labeled data.
TechniqueDetectorTableFigureNatural ImageLogoSignature
Nguyen et al. [57]YOLO F88.363.773.018.350.6
OursFaster R-CNN
(10% labeled)
87.160.782.528.258.0
Table 12. A detailed representation of Precision, Recall and F1-Score at multiple IoUs. Our model is trained on 10% IIIT-AR-13K labeled data with Faster R-CNN and ResNet50 backbone.
Table 12. A detailed representation of Precision, Recall and F1-Score at multiple IoUs. Our model is trained on 10% IIIT-AR-13K labeled data with Faster R-CNN and ResNet50 backbone.
MethodIoUPrecisionRecallF1-Score
Semi-Supervised
Soft Teacher
Faster R-CNN + ResNet50
10% labeled
0.500.910.880.83
0.600.880.840.81
0.700.800.760.77
0.800.640.590.66
0.900.370.330.42
Table 13. The table represents the mAP@(0.5:0.95) of sub-PubLayNet, sub-IIIT-AR-13K and sub-DocBank validation datasets on a sub-PubLayNet trained model.
Table 13. The table represents the mAP@(0.5:0.95) of sub-PubLayNet, sub-IIIT-AR-13K and sub-DocBank validation datasets on a sub-PubLayNet trained model.
Sub-PubLayNetSub-IIIT-AR-13KSub-DocBank
Methods Labeled Data Table Figure mAP Table Figure mAP Table Figure mAP
ST_FRCNN_R505%95.992.194.059.723.041.472.749.060.8
10%95.992.294.060.224.242.272.349.460.8
ST_FRCNN_R1015%96.094.294.561.122.641.971.148.059.6
10%96.192.995.157.316.637.066.248.057.1
Table 14. The table represents the mAP@(0.5:0.95) of sub-DocBank, sub-IIIT-AR-13K and sub-PubLayNet validation datasets on a sub-DocBank trained model.
Table 14. The table represents the mAP@(0.5:0.95) of sub-DocBank, sub-IIIT-AR-13K and sub-PubLayNet validation datasets on a sub-DocBank trained model.
Sub-DocBankSub-PubLayNetSub-IIIT-AR-13K
Methods Labeled Data Table Figure mAP Table Figure mAP Table Figure mAP
ST_FRCNN_R505%82.464.573.488.571.279.852.314.333.3
10%82.565.373.989.072.080.560.628.040.3
ST_FRCNN_R1015%80.765.373.084.965.175.055.518.236.9
10%80.468.574.487.165.076.055.417.936.7
Table 15. The table represents the mAP@(0.5:0.95) of sub-IIIT-AR-13K, sub-PubLayNet and sub-DocBank validation datasets on a sub-IIIT-AR-13K trained model.
Table 15. The table represents the mAP@(0.5:0.95) of sub-IIIT-AR-13K, sub-PubLayNet and sub-DocBank validation datasets on a sub-IIIT-AR-13K trained model.
Sub-IIIT-AR-13KSub-PubLayNetSub-DocBank
Methods Labeled Data Table Figure mAP Table Figure mAP Table Figure mAP
ST_FRCNN_R505%86.351.568.973.125.849.433.620.527.0
10%88.054.871.475.731.053.354.329.041.7
ST_FRCNN_R1015%87.345.866.575.431.353.460.431.345.8
10%85.548.367.478.321.249.756.023.039.5
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Share and Cite

MDPI and ACS Style

Kallempudi, G.; Hashmi, K.A.; Pagani, A.; Liwicki, M.; Stricker, D.; Afzal, M.Z. Toward Semi-Supervised Graphical Object Detection in Document Images. Future Internet 2022, 14, 176. https://doi.org/10.3390/fi14060176

AMA Style

Kallempudi G, Hashmi KA, Pagani A, Liwicki M, Stricker D, Afzal MZ. Toward Semi-Supervised Graphical Object Detection in Document Images. Future Internet. 2022; 14(6):176. https://doi.org/10.3390/fi14060176

Chicago/Turabian Style

Kallempudi, Goutham, Khurram Azeem Hashmi, Alain Pagani, Marcus Liwicki, Didier Stricker, and Muhammad Zeshan Afzal. 2022. "Toward Semi-Supervised Graphical Object Detection in Document Images" Future Internet 14, no. 6: 176. https://doi.org/10.3390/fi14060176

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop