Next Article in Journal
Hardware Development and Safety Control Strategy Design for a Mobile Rehabilitation Robot
Previous Article in Journal
Seismic Response Study of L-Shaped Frame Structure with Magnetorheological Dampers
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Visual Object Detection with DETR to Support Video-Diagnosis Using Conference Tools

by
Attila Biró
1,2,3,
Katalin Tünde Jánosi-Rancz
4,
László Szilágyi
4,5,
Antonio Ignacio Cuesta-Vargas
2,3,6,
Jaime Martín-Martín
3,7 and
Sándor Miklós Szilágyi
1,*
1
Department of Electrical Engineering and Information Technology, George Emil Palade University of Medicine, Pharmacy, Science, and Technology of Targu Mures, Str. Nicolae Iorga, Nr. 1, 540088 Targu Mures, Romania
2
Department of Physiotherapy, University of Malaga, 29071 Malaga, Spain
3
Biomedical Research Institute of Malaga (IBIMA), 29590 Malaga, Spain
4
Computational Intelligence Research Group, Sapientia Hungarian University of Transylvania, 540485 Targu Mures, Romania
5
Physiological Controls Research Center, Óbuda University, 1034 Budapest, Hungary
6
Faculty of Health Science, School of Clinical Science, Queensland University Technology, Brisbane 4000, Australia
7
Legal and Forensic Medicine Area, Department of Human Anatomy, Legal Medicine and History of Science, Faculty of Medicine, University of Malaga, 29071 Malaga, Spain
*
Author to whom correspondence should be addressed.
Appl. Sci. 2022, 12(12), 5977; https://doi.org/10.3390/app12125977
Submission received: 4 May 2022 / Revised: 1 June 2022 / Accepted: 9 June 2022 / Published: 12 June 2022
(This article belongs to the Section Computing and Artificial Intelligence)

Abstract

:
Real-time multilingual phrase detection from/during online video presentations—to support instant remote diagnostics—requires near real-time visual (textual) object detection and preprocessing for further analysis. Connecting remote specialists and sharing specific ideas is most effective using the native language. The main objective of this paper is to analyze and propose—through DEtection TRansformer (DETR) models, architectures, hyperparameters—recommendation, and specific procedures with simplified methods to achieve reasonable accuracy to support real-time textual object detection for further analysis. The development of real-time video conference translation based on artificial intelligence supported solutions has a relevant impact in the health sector, especially on clinical practice via better video consultation (VC) or remote diagnosis. The importance of this development was augmented by the COVID-19 pandemic. The challenge of this topic is connected to the variety of languages and dialects that the involved specialists speak and that usually needs human translator proxies which can be substituted by AI-enabled technological pipelines. The sensitivity of visual textual element localization is directly connected to complexity, quality, and the variety of collected training data sets. In this research, we investigated the DETR model with several variations. The research highlights the differences of the most prominent real-time object detectors: YOLO4, DETR, and Detectron2, and brings AI-based novelty to collaborative solutions combined with OCR. The performance of the procedures was evaluated through two research phases: a 248/512 (Phase1/Phase2) record train data set, with a 55/110 set of validated data instances for 7/10 application categories and 3/3 object categories, using the same object categories for annotation. The achieved score breaks the expected values in terms of visual text detection scope, giving high detection accuracy of textual data, the mean average precision ranging from 0.4 to 0.65.

1. Introduction

Nowadays, the importance of multilingual translation has grown exponentially thanks to the impact of the COVID-19 pandemic [1,2]. Visual Object Detection [3,4] (moreover, Textual Object Detection), as a central topic of video conference-based remote diagnosis and consultancy [5] has intensified the digital transformation in the medical field. The issue generates a new ‘realtime multilingual data processing’ market to catalyze the elimination of existing communication barriers of multilingual video conferences. The adoption of Visual Object Detection (VOD) to the healthcare domain will increase the efficacy of remote diagnosis processes [2], reducing the reaction time in remote consultations. Furthermore, optical character recognition (OCR) is becoming more popular [6], since it is considered a central pillar of many advanced technologies, particularly related to real-time object detection.
The Textual and Visual Object Detection (TVOD) method analyzed popular algorythms and models in object detection (e.g., Fast R-CNN, Faster R-CNN, Region-based Convolutional Neural Networks, Region-based Fully Convolutional Network, as well as realtime object detectors such as YOLO, DETECTRON and DETR) to syntetize the outcomes and to provide reliable outcomes for a new type of collaborative working model. Our research results can be easily adapted to the Sports Safety field using video conference tools or platforms.
This research validates the feasibility of VOD [3,4] topics, using several ‘realtime object detection Artificial Intelligence’ (AI) models and architectures to support the non-supervised communication in Video-Diagnosis, using Conference Tools. Unfortunately, such systems, ready for multilingual interpretation support are not available to our best knowledge, and such attempts are not visible in the literature.
The outcomes of this paper are crucial in the sense that it applies a feasible approach for communication bridge to health scientists as an invisible technological layer (the ‘realtime multilingual translator engine’), using an artificial intelligence-based pipeline [7], to support collaborative video conferencing solutions.
This multilingual collaboration part of research [8] has significant importance and adapts technologies (OCR, AI) from other sectors (‘Information Technology’ (IT), Telecom, Industry) to Healthcare. The reason, we started the research: Polysemy, ambiguity, individual usage trends, and natural changes of importance over time contribute to the fact that languages are difficult to grasp. In some specific languages, it is not enough to identify just the phrases on the screen, because—based on the context—they might have different meanings. In the case of languages such as the 22 Indian languages [9], Chinese or Japanese, it will raise out the contextualization topic. While translating a specific word, phrase, or sentence, the system considers broader additional information that clarifies the role of understanding the text in question. However, it is not obvious what could and should be considered as context, when combined with visual information and beneficial context information.
These topics generate a central issue for human translators. To get a proper interpretation while attending a dedicated technical health-based conference call, attendees must submit their presentation with all relevant information a couple of days before the live session. In the case of multilingual (ad hoc) video consultancies, where every minute matters, there is no place for such preparation.
There are existing models [10,11] and methods for bidirectional translation [12,13], but in the case of multilingual and multiuser conferences, when the involved parties are speaking in languages that are not coming from the same language families, the complexity of translation is becoming high, increasing the latency time of translation as well as decreasing the translation quality.
Challenging topics are solved currently with human interpreters, but this kind of setup is expensive and requires extra arrangements. To substitute the human factor with a highly reliable solution [13,14], several research works have been conducted [15] providing various approaches for different setups. This paper proposes a (‘DEtection TRansformer’) DETR [16,17,18,19] based solution design and procedure for Textual (Visual) Object Detection and will search the threshold where we can achieve feasible accuracy to support remote multilingual video diagnostics.
For this research on the validity and feasibility checking side, deep learning based model structures has been applied, such as YOLO4 (‘You Only Look Once’) [20,21,22], the DETR (‘DEtection TRansformer’) [16,17,23] and the Detectron2 [24,25,26,27,28]. We conducted the experiments in two phases: the first round contains the technical minimum training data set for reasonable statistical learning models and analyses different model architectures’ hyperparameters. Further on, we updated the data collection and annotation protocol based on these experiments. Finally, we doubled the initial training and validation data sets to improve the experiments based on the lessons learned from the first phase.
We analysed other approaches on the market (such as in the case of [17]) where the experiments were they performed on COCO 2017 detections with 118 k training images and 5 k validation images. Our scope was to find a feasible solution on minimal required training data and accessible environment for healthcare sector. Our research provides solutions where the model requires a minimum training data set [29].
The research goal of both phases was to run limited validation-type experiment rounds—on each candidate group—and to summarize their results, and simultaneously to measure, compare, and analyze the outcomes, as well as to bring specific requirements about input data and preparation process efficacy. The main scope of the experiments is to prove that the detection and localization of textual and other visual element types are possible with high accuracy (even on limited training data sets), by application of screenshot images. Furthermore, using AI models already established in other sectors, we can provide outcomes with a high degree of precision to detect the textual bounding boxes for videoconference tools. These models might be part of a further cooperative communication layer between video diagnosis [5] attendees.
The novelty and innovative content of this research is the adaptation of already known AI solutions and Machine Learning algorithms from other fields to healthcare to save time, save human resources, improve the efficacy of multilingual collaboration tools, break multilingual limits and save lives. In this way, the outcomes and results can quickly cover other use cases with minimal effort.

2. Materials and Methods

Unlike traditional computer vision techniques, DETR approaches object detection as a direct set prediction problem. It relies on a set-based global loss, which forces unique predictions via bipartite matching, and a “Transformer Encoder-Decoder” architecture [18].
To validate DETR’s feasibility, we conducted additional experiments with three of the most promising real-time object detector models, YOLO4, Detectron2, and DETR, each with different model variations.
Given a fixed small set of learned object queries, DETR reasons about the relations of the objects and the global image context [30] to output the final set of predictions in parallel directly. Due to this parallel nature, DETR is very fast and efficient, and for this reason, we propose to employ the End-to-End Textual Object Detection with Transformers [18]. Figure 1 shows the concept of DETR.
DETR is a PyTorch model established by Facebook based on the Transformer methodology [31], where the model’s backbone is a Resnet50 image classification model, which is fine-tuned with a Transformer based classification head for detecting relevant bounding boxes (in our case textual boxes, such as word, sentence, phrase) and assigning them to the appropriate predefined category. Transformers were introduced by Vaswani et al. [32] as a new attention-based building block for machine translation [17]. For our use case, we implemented a custom-defined classification based on four class types. The input data does not have to be the same shape (same as Detectron2). The annotation format is JSON.

2.1. Data

We prepared data sets from real-life use cases for our research experiments, with data taken from real video conferences and applications for this requirement. Due to the lack of appropriate options, we did not use any predefined annotated data sets. Instead, a dedicated Python picture extractor has made the data acquisition process to gain relevant potential pictures from videos. It has been used for manual data collection as well.
There were collected, filtered, and classified thousands of raw pictures. Finally, we get the data set for two independent research phases. The first one was composed of 248 train data set records, with 55 sets of validation data instances for seven application categories and three object categories, providing the necessary minimum threshold number for a specific data science project. The second phase had 512 train data set records, with 110 sets of validation data instances for ten application categories and three object categories. Human experts have made the data labeling based on a dedicated annotation and data processing protocol. The input data has been adjusted for different models (YOLO4, DETR, Detectron2) using Python scripts. Human experts performed the annotation process using the Visual Object Tagging Tool (VOTT) [33] application on Windows10 and Mac systems, as well as the Label Image Tagging Tool (LabelImg) [34] on Linux or Mac. The experiments were implemented in Python on the Google Colab Pro platform.

2.2. Experimental Environment

Our experiments were run on Config 1 PC with the following configuration: MBO Gigabyte Z390 Aorus Pro, CPU INTEL Core i7-8700K 3.7 GHz 12 MB LGA1151, DDR4 32 GB 3600 MHz Kingston HyperX Predator Black CL17 KIT2, VGA MSI RTX 2080 Ventus 8 GB, SSD M.2 SAMSUNG 970 Pro 1 TB, Corsair RMx (2018) 750 W Modular 80+ Gold) as well as on Google Colab- and Google Colab Pro (to be able to use accelerated hardware) platform.

2.3. Data Classification

From an algorithmic perspective, there is no particular limit for the number of object categories or instances. From the task perspective, using “Visual Textual Object Detection” as an approach, different levels of object differentiation can be valid. For our use case, we proposed four categories as follows:
  • Class 1 (TEXT) are the text blocks, and contains text only;
  • Class 2 (STATIC) contains buttons or texts and need additional dictionaries for further processing;
  • Class 3 (NO TRANSLATE) contains mixed elements unnecessary to process. They prevent sections from being recognized as any other category, providing data processing according to GDPR;
  • Class 4 (CONTAINER) is defined as inter-dependency element. They are container objects to connect larger blocks of interdependent textual elements.

2.4. Architecture

DETR is a PyTorch model of Facebook using the Transformer methodology (see Section 2 and Figure 1), and matches Faster R-CNN [35] with a ResNet-50 [36] than with other options we tried during our experiments. The architectue consists of three main parts: a CNN backbone for extracting compact feature representations, a simple encoder-decoder transformer, and a simple FFN for predicting the final detection [17]. The image input size was defined arbitrarily. Experiments have shown that images resized between 800 and 1333 pixels produce the best results. Initially, we chose as image size 800 pixels, then increased it to 1333 pixels to improve the accuracy. Another important fact about the architecture is that it uses Transformer attention blocks.

2.5. Algorithm Logic

As it has been mentioned by N. Carion et al. “DETR model directly predicts (in parallel) the final set of detections by combining a common CNN with a transformer architecture” [17]. In a generic way, the Transformer contains five major parts: (1) Input transformation for Encoder, (2) Input Transformation for Decoder, (3) Encoder—to extract the contextual features from input sequence through multi-head self-attention mechanism, (4) Decoder—from encoder output to generate the next prediction, and the (5) Final Layer to map decoder’s output to the object probability space [19].
As is shown in Figure 1, the logic of the algorithm is structured as follows: (a) the backbone ResNet50 predicts image features from the image, (b) the algorithm calculates the spatial positional encoding for each block of the predicted image resized to the size range of 800 to 1333 pixels, (c) run Transformer module on the blocks with the belonging positional encoding, (d) each Transformer head predicts the bounding box and the object within it, and finally (e) the Transformer predicts the number of objects in the image [19].
There are two pillars of the direct set predictions in detection algorithm as shown in Figure 1: (1) a set prediction loss, which enables unique matching between predicted and ground truth boxes, as well as (2) a dedicated architecture which enables the prediction of a set of objects and models the relations in one step [17].
DETR’s transformer uses positional encodings at every attention layer, as shown in Figure 2. From the CNN backbone, the image features are passed through the transformer encoder, in parallel with spatial positional encoding. These features are then added to queries and keys within the ‘multi-head self-attention’ layers. After that, the decoder gets the queries, output (i) positional encoding, and (ii) encoder memory. Finally there are generated the final set of ‘predicted class labels’ and the ’textual bounding boxes’, through the ‘multiple multihead self-attention’ and ‘decoder-encoder’ attention [17].

2.5.1. Advantages

From a textual object detection perspective, DETR offers a simpler, more flexible pipeline architecture that requires fewer heuristics and casts the object detection (in our case text detection) task as an image-to-set problem [17]. Traditional computer vision models use complex pipelines with custom layers to localize objects in an image and extract the features. Conversely, DETR replaces this approach with a simpler neural network (NN), resulting in deep end-to-end learning (DL) solution to the object detection problem [17]. In TVOD research, the most important advantages are as follows: (a) Fast model—12 frames per second on GPU with the most accurate model configuration, (b) State-of-the-art accuracy as of 2020, (c) it has a straightforward implementation thanks to the PyTorch wrapper.
In our experiments we used a modest testing environment (see Section 2.2), which can be accessible by anyone currently on market.

2.5.2. Disadvantages of DETR from the Textual Object Detection Perspective

According to DETR research paper [19], one of the main disadvantages of DETR from the “textual object detection” perspective is that the detection efficacy on small targets is usually poor. In order to solve this issue, the model needs significant improvements [37]. In the case of videoconference tools, this deficiency is not present. In this research, we faced the following: (a) it requires heavy computation power (minimum 8 GB GPU memory required for the model), (b) the Transformer is a black box for interactive interoperability, (c) the Transformer network requires the largest amount of training data.

2.6. Pre-Processing

Nowadays, the Visual Transformers are the subject of dedicated focus because of their ability to comprehensively interpret the image’s information. These model types have high fine tunability and higher accuracy than traditional image processing neural networks.
“DETR approaches object detection as a direct set prediction problem. The model uses a set-based global loss, which forces unique predictions via bipartite matching, and a Transformer encoder-decoder architecture” [30]. Each experiment and validation research (see Section 3.1 and Section 3.2) in the same research phase was conducted on the same training-and validation dataset, application categories, and object categories, resulting in: (i) Phase 1: 248 images with 55 validation, 7 application categories, and 3 object categories, (ii) Phase 2: 512 annotated images, with 11 validation, 10 application categories, and 3 object categories. In addition, the prepared dataset contains manually prepared, multi-oriented scene text with quadrangle bounding box annotation.
We tested the model’s limits using a minimum viable set of learned object queries (see Section 2.1) on minimalist test environment (see Section 2.2). We tested the most well-known algorithms (See Table 1 and Table 2) and their variations (See Table 3) to get the best performance index and to prove that if we choose the optimal setup, we will not face a well-known drawback: to run until infinity the training process. Our goal was to run some relatively limited experiment rounds and summarize the results. We made result elaborations based on average precision (AP) and mean average precision (mAP) metrics for evaluation. Furthermore, we also inspected some labeled validation examples visually for a better human evaluation. We used the following variations:
  • Different number of epochs: 50, 150, and 1000.
  • Different categorizations: all 4 predefined categories, or 2 categories: text + others (static, no translate, container categories merged to one category).
Table 1. Detectron Phase 1 Experiments and Results.
Table 1. Detectron Phase 1 Experiments and Results.
ModelIterations AP 1 AP 2 AP 3 mAPTotal Loss
faster_rcnn_R_50_C4_3x5000.070340.073670.092350.068771.320
faster_rcnn_R_50_DC5_3x10000.176590.233420.209210.177061.253
faster_rcnn_R_50_FPN_3x7000.309850.443420.360160.342551.154
faster_rcnn_R_50_FPN_3x10000.305600.423020.347450.341601.183
faster_rcnn_R_101_FPN_3x10000.297640.471870.437850.393411.129
faster_rcnn_X_101_32x8_FPN_3x30000.427990.560960.563310.493690.458
faster_rcnn_X_101_32x8_FPN_3x10,0000.452230.578210.557340.487580.343
retinanet_R_50_FPN_1x10000.310270.459980.473570.408720.495
retinanet_R_50_FPN_3x10000.321500.461140.469530.411030.477
retinanet_R_101_FPN_3x30000.369230.552300.552670.495890.250
retinanet_R_101_FPN_3x10,0000.379680.559380.572700.505130.189
rpn_R_50_C4_1x1000 0.320
rpn_R_50_FPN_1x3000 0.180
Table 2. Detectron Phase 2 Experiments and Results.
Table 2. Detectron Phase 2 Experiments and Results.
ModelTotal LossmAPWorkersBATCH/IMGIMS/BATCHIterations
faster_rcnn_R_50_C4_1x0.9630.29753412861000
faster_rcnn_R_50_DC5_1x0.9670.31446412861000
faster_rcnn_R_50_FPN_1x0.9120.33348412861000
faster_rcnn_R_50_C4_3x0.9120.34338412861000
faster_rcnn_R_50_DC5_3x0.8860.34033412861000
faster_rcnn_R_50_FPN_3x0.9120.36219412861000
faster_rcnn_R_50_FPN_3x0.7070.39064412863000
faster_rcnn_R_101_C4_3x0.8570.32096412861000
faster_rcnn_R_101_DC5_3x0.8820.30924412861000
faster_rcnn_R_101_FPN_3x0.8690.32056412861000
faster_rcnn_X_101_32x8_FPN_3x0.9980.3045526431000
retinanet_R_50_FPN_1x0.6610.24791412861000
retinanet_R_50_FPN_3x0.5680.25690412861000
retinanet_R_101_FPN_3x0.5640.28165412861000
rpn_R_50_C4_1x0.3500.24046412861000
rpn_R_50_FPN_1x0.3760.35200412861000
Table 3. YOLO4 experiments and results.
Table 3. YOLO4 experiments and results.
YOLO4 TINY 640 (16)YOLO4 TINY 416 (32)YOLO4 TINY 800 (16)
Epochs AP 1 AP 2 AP 3 mAP AP 1 AP 2 AP 3 mAP AP 1 AP 2 AP 3 mAP
10 0.0290.0400.1300.0660.1450.1990.3980.247
200.1640.1530.3520.2230.0510.0390.1170.0690.1720.1410.3120.208
300.1580.1360.3520.2150.0690.0470.1310.0820.2800.3440.2910.305
400.2720.1500.3430.2550.0640.0510.1270.0810.2810.2650.3920.312
500.2550.1560.3720.2610.0700.0630.1180.0840.2570.3320.4450.345
600.2700.1790.3370.2620.0680.0650.1130.0840.2690.2220.4720.321
700.3080.1290.3390.2590.0780.0580.1310.0890.2900.4000.4840.392
800.2960.1050.3560.2520.0870.0620.1300.0930.3260.3590.4850.390
900.3430.2310.3640.3130.0800.0590.1380.0920.3470.4050.4850.412
1000.3500.1520.4180.3070.0870.0690.1330.0960.3390.3790.4810.400
1100.3260.1800.4080.3050.0900.0690.1370.0990.3350.3950.4860.405
1200.3260.2360.3930.3480.0920.0720.1370.1000.3550.3990.4790.411
1300.3000.1750.4100.295 0.3430.4110.4860.413
140nannannannan 0.3400.4110.4830.411
150 0.3450.4090.4830.412
YOLO4 LARGE 480 (16)YOLO4 TINY 416 (16)YOLO4 TINY 800 (16, 0.005)
Epochs AP 1 AP 2 AP 3 mAP AP 1 AP 2 AP 3 mAP AP 1 AP 2 AP 3 mAP
100.1400.0230.1900.1180.0350.0080.1040.0490.3150.3370.4510.368
200.2060.0610.2630.1770.0490.0260.0930.0560.3730.2710.4320.359
300.2140.0240.2660.1680.0430.0300.1090.0610.4400.2470.3720.353
400.2470.0410.2880.1920.0640.0150.1020.060nannannannan
500.2280.0240.3000.1840.0620.0310.1110.068
600.2680.0570.3710.2320.0490.0280.1180.065
700.2520.0420.3440.2130.0610.0310.1050.065
800.2720.0570.3230.2170.0760.0270.1200.074
900.2580.0120.2830.1840.0790.0180.1420.080
1000.2810.0400.3160.2120.0840.0290.1100.074
1100.2350.0110.2980.1810.0780.0340.1320.081
1200.2620.0340.2990.1980.0800.0270.1400.082
130 0.0870.0230.1300.080
140 0.0860.0220.1350.081
150 0.0870.0240.1330.081

2.7. Decision Making

Based on the initial training results, we recognized two bottlenecks to be managed: (a) the results do not meet the expected thresholds, and (b) the training still has some reserves. These are visible in Figure 3 and Figure 4.
Based on our findings, we redesigned the system to produce results despite the training data set limitation. Experimental findings showed that categorizing four classes for textual object detection requires much more training data than is proposed by standards. A broader complexity of options and outcomes will not be enough to support a feasible multilingual VC consultancy solution. Therefore, we modified the initial training data set by merging three categories into one, yielding the following model with the results shown in Figure 5, which can be compared with the non-clustered solution (see Figure 6). Following are the changes to the classes:
  • New Class 1, which corresponds to the previous Class 1 and requires translation,
  • New Class 2, which corresponds to the union of previous Classes 2, 3, and 4, and requires no translation.

3. Data Analysis

For evaluation, we used average precision (AP) and mean average precision (mAP) [38], computed as follows
mAP = 1 n k = 1 n AP k ,
where AP k represents the average precision of class k, and n is the number of classes.

3.1. Alternative No. 1: YOLO4 Model and Analysis

YOLO offers a wide array of model variations, versions, architecture types, implementations, and underlying AI frameworks. Figure 7 shows the YOLO4 architectures, with Input (like Images, Patches or Image Pyramid), Backbone (it could be: VGG16, ResNet-50, SpineNet, EfficientNet, ResNeXt-101, Darknet53, DenseNet, SqueezeNet [4], MobileNet, ShuffleNet), Neck (e.g., FPN, PANet, Bi-FPN, etc.), Head which could have Dense Prediction (RPN, YOLO, SSD, RetinaNet, FCOS, CornerNet, CenterNet, MatrixNet, etc.) in the case of One-Stage type Detector as well as Sparse Prediction (as Faster R-CNN, R-FCN, Mask R-CNN, Libra R-CNN, etc.) in the case of Two-Stage Detector [39].
We started the training from randomly initialized weights or some weight sets pre-trained from scratch. This research focused on YOLO v4, using Python and TensorFlow. Data annotation has been done by VOTT [33] and LabelImg [34]. Later ones always lead to better convergence. YOLOv4 has two basic variations: standard (large [257 MB]) and tiny (small [23 MB]). These were involved in evaluation with the following configurations over 120–150 epochs: (i) Different input image sizes (between 416 × 416 and 800 × 800 pixels), (ii) Different batch sizes (8–32), (ii) Different learning rates (0.005–0.02), (iii) Different optimizers (SGD, Adam, Nadam).

3.2. Alternative No. 2: Detectron2 Model and Analysis

Detectron2 is a PyTorch based framework by Facebook that implements several of the widespread object detection/segmentation algorithms, such as Mask R-CNN, RetinaNet, Faster R-CNN, RPN, Fast R-CNN etc. As the meta-architecture shown in the Figure 8, we can distinguish three blocks: (1) Backbone Network to extract the feature maps from the input image, (2) the Region Proposal Network which detects the object regions from features and the (3) Box Head which is one of the subclasses of ROI heads [40].
In our study, only algorithms with model zoo [41] were involved in object detection. Input data were not in the same shape, unlike in the case of YOLO. Therefore, we transformed the annotations into COCO JSON format. We tried out and evaluated all proposed (best practices) relevant algorithms with the following variations: (i) different number of iterations (1000 to 10,000), (ii) different number of images per batch (4–16), (iii) different batch size per image (8–128), (iv) different warm-up iteration count (50 to 150), (v) different learning rates (0.001–0.02).

4. Results

4.1. DETR Models

For the DETR experiment, the minimum required VRAM is 6 GB. However, the minimum VRAM is a minimum standard requirement for efficient training. One of the advantages of this model is that it could not overfit with prepared data set (see Section 2.1 for details). At approximately 500 epochs, all experiments reached their peak performance (see Figure 5 and Figure 6). The global loss values are a combination of bounding box detection accuracy and categorization accuracy. DETR had a reasonable training time: around 9.5 h for 1000 epochs (for environment see Section 2.2) on the available training data set. Based on our experiments, the number of categories affects the model’s final performance, and we can reach the top performance by using a clustering approach (synergizing the relevant classes). The best results with four categories were gained in mAP > 0.5 (see Figure 6). Nevertheless, we reached the best overall research outcome ( mAP = 0.65 ) using a limited number of categories by clustering (see Figure 5); as a result, we obtained a better mAP score in almost the same number of epochs.
The results of the experiments were analyzed and evaluated through the values of the loss function for the training dataset (solid line) and the validation dataset (dashed line) as well as mean average precision scores for the validation dataset (solid line). In the results figures (see. Figure 3, Figure 4, Figure 5 and Figure 6), the left graph shows the loss function and the right one the mAP score, both plotted against the epoch number. Figure 3, Figure 4 and Figure 6 show the results of the first, second, and third experiment, with 50, 150, and 1000 epochs, respectively, using all four predefined categories—text, static, no translate, container—and DETR specific N/A categories for disputable results.
Figure 5 shows the 4th experiment results with 1000 epochs using the clustered approach. Subsequently, we merged three of the original categories into one, which resulted two new categories: (i) ‘New Class 1’ with original Class 1 (text) and (ii) ‘New Class 2’ for ‘no translation object holder’ composed from Class 2 (static) with Class 3 (no translate) with Class 4 (container) from original mode. Finally, Figure 9 shows the OCR output connected to the 4th experiment and confirms that the algorithm is feasible even with mAP = 0.65 .

4.2. Alternative No. 1: YOLO4 Models

It is visible in Table 3 that YOLOv4 (tiny) reached mAP values up to 0.41. We used the following settings: the pre-trained initial weights, 800 × 800 input shape, 120 epochs, SGD optimizer, initial learning rate set to 0.01, and a fixed scheduled rate decay during the run epochs. Based on the experiments we formulated the following remarks:
  • The performance does not strictly correlate with training time (epochs).
  • The performance almost always plateaued before terminating the learning process.
  • Accuracy does not always correlate with the overall mAP and class-specific AP values.
  • In comparison with other experiment groups such as Detectron2 or DETR, YOLO requires more training time (around 3–6 min per epoch), which makes the hyperparameter tuning process very slow.
  • A very long training–over 150 epochs–becomes uncertain due to the vanishing gradient phenomenon. At this moment, the training is interrupted irrespective of the original expected length of the experiment. ‘YOLO large’ and ‘YOLO tiny’ have the same performance, ‘YOLO tiny’ actually tends to give better results. We realized that there is no reason to choose YOLO large for further experiments. Larger input image size provided better results than expected (up to the examined 800 × 800 , at least).
  • With batch size up to 8 the relevant models were unable to learn, so we had to set the batch size to 16 as the minimum value, while 32 was the maximum value we employed due to GPU memory constraints (see Section 2.2).
According to our investigation, the optimizer type does not seem to have a significant impact in terms of performance. While advanced ones (such as Adam, Nadam) are initially faster, they tend to have the same overall accuracy as vanilla Stochastic Gradient Descent in the long run. However, Adam and Nadam apparently lead to vanishing or exploding gradients more often.

4.3. Alternative No. 2: Detectron2 Models

In our experimental environment described in Section 2.2, 1000 iterations last around 20–30 min, and the “iterations” here cannot correspond to YOLO epochs. The more advanced model types have higher memory requirements. In the case of Detectron2, the warm-up iteration number does not affect the outcome. The optimal learning rate is around 0.005–0.01, where the larger ones tend to explode gradients. In the case of a more advanced model, this type provides the most accurate outcome. According to our research results, performance is not strictly correlated with training time (epochs): performance almost always plateaued before the learning process was terminated. According to the experiments, as indicated in Table 1, RetinaNet and Faster-RCNN detectron models provided the best results, reaching around 0.5 in mAP.

4.4. Detectron Phase 2 Research

Detectron2 benefits from models that have much more rapid training than YOLO. Using the same model architecture hyperparameters involved in the earlier experiments, we renewed the research by doing the same experiments and refining our model architecture hyperparameters. However, a more suitable image processing algorithm was applied on the double amount of training and validation images, moreover improving the annotation protocol and the review process. The outcome of connected experiments gives a much better loss value at the same mAP in the case of original classes. The second phase resulted better accuracy by better training data adjustment. Table 2 shows the results of Detectron2 obtained in Phase 2. The result is quite impressive, even with four object categories. In this way, we achieved the goal of this paper, having the optimal minimal set of training and validation data, and the proposed feasible models for a multilingual use case.

5. Discussion

The achieved overall score breaks the expected values in a specific OCR scope, resulting in a high textual data detection accuracy with mAP = 0.65 .
As it is visible on the charts of Figure 3 and Figure 4, during first experiment with DETR ( mAP = 0.174 ) and second experiment with DETR ( mAP = 0.39 ), there is room for improvement by increasing the number of epochs since the model did not yet produce overfitting. During the third experiment with DETR ( mAP = 0.5 ) depicted in Figure 6, we can see that the model reaches its best after 400 epochs, and there is only a tiny increment after that point. The model is not overfitting, but it cannot improve its performance. We observed very similar results in the fourth experiment with DETR ( mAP = 0.65 ) exhibited in Figure 5, but with a limited number of categories. However, a considerably higher mAP score has been achieved almost in the same number of epochs. We checked the results for our Textual Object Detection at this point of research. The system had not reached the 0.90 value with mAP, just 0.65 (see Figure 9); nonetheless, the focused objects were identified accurately.
In Phase 1 of this research, we applied the DETR DL-based model structure. It used only 248 manually collected and annotated training examples and 55 validation examples to fit the technical fminimum viable model for practical statistical learning. To achieve the theoretical maximum with mAP exceeding 0.90 (e.g., in the case of context-sensitive language detection), the model requires a minimum set of 1500 training with 200 validation data, 1700 instances combined, according to our calculations based on research experiments.
The original goal of the research was to prove that detecting textual elements with Visual Transformers and fully customizable ML models is possible with suitable accuracy by applying screenshot images.
Figure 3 shows that there is room for refinement by enriching the number of epochs because the model is not overfitted yet. Overfitting can be determined from loss functions: we can reach the starting point of overfit when the training loss decreases, but the validation loss starts to increase. Figure 4 shows the 2nd experiment results, and it is observable that we can still boost the epoch number. Figure 6 shows the 3rd experiment results. The model is reaching its best after 400 epochs, and there is only a small increment after that point. We have to change the type of experiment if we want to improve the model. The model is not overfitted yet, but it cannot improve anymore via learning. At this point, we obtained mAP = 0.5 . Figure 5 shows the 4th experiment results with 1000 epochs with a clustered approach, and this experiment brings the real breakthrough; a much better score ( mAP = 0.65 ) was reached in almost the same number of epochs (see Figure 5). Figure 9 shows the OCR output connected to the 4th experiment.
The feasibility checking and validation experiment emphasize that the models recognize text boxes and categories with high accuracy, specific to the requirements of the textual topic approach. Each experiment group gave the same level of performance (mAP between 0.4 and 0.65) irrespective of the actual model types. The main bottleneck for further improvement is the limited data quantity and quality and the coverage of all use cases. For this reason, we improved the annotation protocols, annotation review process, and image acquisition and selection guidelines to start the second research phase.
YOLO reached its best ( mAP = 0.413 ) with YOLOv4 Tiny (see Table 3). YOLO’s disadvantage was the slow training process. Detectron’s best score of mAP = 0.5 is coming from Retinanet R101 FPN 3x (see Table 1). Detectron2’s disadvantage is the complex framework. Finally, DETR’s best of mAP = 0.65 came from DETR RESNET50 (see Figure 5). Disadvantages of this model are the slow training, separated detection, and classification model. As a further investigation, it will be compared with the results of the most promising architectures. According to the results, it can be provided a new solution to support the multilingual, collaborative distant diagnostics, breaking the language barriers with AI-supported solutions. The settings used by the best models are as follows:
  • YOLO4 Tiny: 800 × 800 input size, 120 epochs, SGD optimizer, learning rate 0.01, batch size 16;
  • Detectron2 RetinaNet R 101 FPN 3x: 10,000 epochs, learning rate 0.01, 6 images per batch, 128 batch per images, 150 warm-ups;
  • DETR RESNET50: variable input size, 500 epochs, learning rate 0.0001, batch size 2, LR decay 0.0001 × every 200 epochs.
During the second phase experiments, we obtained almost the same level of mAP but with much better loss values (see Table 2). Suppose we apply the clustering approach to our data classification, in other words: instead of four different classes (text, static, no translate, container), we focus just on textual objects—then our models reach the best performance on a moderate number of training data with validation and in this way we can adapt real-time object detection models to real-time consultations [5,8] and multilingual collaborations tools and solutions. From the investigated models, DETECTRON and DETR can serve best as these approaches because they will provide the same level of accuracy, but the best results will provide DETR.

5.1. DETR Insights on Textual Object Detection

The most important outcomes from the experiments can be summarized as follows:
  • About 6 GB of VRAM is required (at least) for efficient training;
  • The model could not be over-fitted during the experiments, and this is promising;
  • At approximately 500 epochs, all experiments reached their peak performance;
  • The global loss values are a combination of bounding box detection accuracy and categorization accuracy;
  • Moderate training time (9.5 h/1000 epochs with the available training data);
  • The number of categories affects the model’s final performance—object detection with clustering [42] was a line of research connected to our scope.

5.2. AI for OCR and Multilingual Translations for Video Conference Consultancies

Object detection with transformers (DETR) reaches competitive performance with Faster R-CNN via a transformer encoder-decoder architecture [43]. Our experiment shows that even with a low number of pictures and moderate performance ( mAP = 0.65 ), we can build models that recognize text box elements, e.g., phrases for collaborative work. Multilingual real-time neural translators can easily deploy this approach as an interface for collaborative and interactive translation in real-time. As a consequence of our study, to fit the model to a wide variety of languages, we can define precise requirements for a highly accurate model, together with annotation guidelines and preparation policies.

6. Conclusions

The paper proposed DETR for visual, textual data recognition to support future translation of multiuser videoconferences and to boost the efficacy of distance diagnosis. The Textual and VOD project analyzed different scenarios to provide reliable outcomes for a new collaborative working model. Results will be applied to various sectors (e.g., health, sports) that use remote consultations or e-collaboration tools and videoconference platforms (like Zoom, Google Meet, Microsoft Teams, Skype, Google Hangouts, GoToMeeting, Cisco Webex Meetings, Bluejeans, Intermedia Anymeeting, RingCentral Video, Zoho Meeting, Dialpad Meetings, TrueConf Online, Lifesize Go or even Youtube Live, Facebook Live, Slack Video Calls, etc.). The research validated, the feasibility of the Textual Object Detection topic, using several different models and architectures to support video diagnosis using conference tools.
The main goal of the investigation was to run some limited validation experiment rounds on each candidate group, summarize their results, compare the outcomes, and bring specific requirements about the input data and the preparation process, and the efficacy of the new approach.
In the case of multilingual conferences, to support further data processing, such as neural multilingual translation, a feasible visual (textual) object detection model can be prepared by a modest number of specific training datasets using the DETR model. When it comes to adjusting the efficacy level, it depends on clustering or object classes, quality- and quantity of training data. The solution provides satisfactory results, starting from 0.50 mAP for this use case, which is an acceptable outcome connected to data science and DL projects. Using DETR with the clustering approach for textual object detection, the prediction of bounding boxes give high accuracy at approximately 500 epochs (see Figure 5 and Figure 6), where all experiments reach their peak performance. The results of mAP = 0.65 represent high accuracy in textual object detection, as shown in Figure 9, being promising from the collaborative, multilingual solutions perspective. It could bring a new line of development for specific applications. In VOD projects, we usually expect mAP values above 0.9 to prove the feasibility of different use cases. However, if we are only looking for texts and phrases, the expected criteria are already met above 0.5 mAP. Experiments have demonstrated that using the DETR model for a restricted number of training datasets with real-time collaboration tools is possible to strengthen multilingual collaboration with multilingual translation. The textual data object detection is feasible having a mean average precision of 0.65.

Author Contributions

Conceptualization, A.B., S.M.S. and A.I.C.-V.; methodology, A.B., A.I.C.-V., J.M.-M. and S.M.S.; software, A.B. and K.T.J.-R.; validation, A.B., K.T.J.-R., L.S. and S.M.S.; formal analysis, A.B., J.M.-M. and S.M.S.; investigation, A.B. and K.T.J.-R.; resources, A.B. and S.M.S.; data curation, A.B. and K.T.J.-R.; writing—original draft preparation, A.B.; writing—review and editing, A.B. and L.S.; visualization, A.B. and L.S.; supervision, A.I.C.-V., J.M.-M. and S.M.S.; project administration, A.B.; funding acquisition, A.B. and L.S. All authors have read and agreed to the published version of the manuscript.

Funding

This research was supported by ITware, Hungary. The work of K.T. Jánosi-Rancz and L. Szilágyi was supported by Sapientia Foundation—Institute for Scientific Research.

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

Data are available upon request. The Figures (DETR architecture, Detailed DETR transformer architecture, YOLO4 architecture, Detectron detailed architecture) and the results charts of this study (Figure 3, Figure 4, Figure 5 and Figure 6, DETR OCR output) can be found in: https://doi.org/10.6084/m9.figshare.19699993 (accessed on 31 May 2022).

Conflicts of Interest

The authors declare no conflict of interest.

Abbreviations

The following abbreviations are used in this manuscript:
Adaman alternative optimization algorithm
AIartificial intelligence
APaverage precision
Bi-FPNweighted bi-directional feature pyramid network
CenterNetmachine learning model for Anchorless Object Detection
CNNconvolutional neural network
COCOcommon objects in context
CornerNetnew approach to object detection
COVID-19coronavirus disease
CPUcentral processing unit
Darknet53is a convolutional neural network architecture
DDRDocument Domain Randomization
DETRDetection Transformer
DenseNeta type of convolutional neural network architecture
EfficientNeta convolutional neural network architecture
FCOSfully convolutional one-stage object detection
FPNfeature pyramid network
GBgigabyte
GDPRGeneral Data Protection Regulation
ITinformation technology
JSONJavaScript object notation
mAPmean average precision
MatrixNeta proprietary machine learning algorithm
MobileNeta convolutional neural network architecture
NadamNesterov-accelerated Adaptive Moment Estimation
NNneural network
OCRoptical character recognition
PANetPath Aggregation Network
Resnet50a convolutional neural network architecture
ResNeXt-101a model introduced in the Aggregated Residual Transformations
RetinaNetis a one-stage object detection model
ROIRegion of Interest
RPNRegion Proposal Network
R-CNNregion-based convolutional neural network
R-FCNregion-based fully convolutional network
SGDstochastic gradient descent
SSDSingle Shot Multi-Box Detector
ShuffleNetan extremely efficient CNN
SpineNetlearning scale-permuted backbone
SqueezeNeta convolutional neural network
TVODtextual visual object detection
VCvideo consultation
VGAVideo Graphics Array
VGGVisual Geometry Group
VGG16a convolutional neural network
VODvisual object detection
VOTTVisual Object Tagging Tool
VRAMvideo RAM, video random access memory
YOLOYou Only Look Once

References

  1. Ozili, P.K.; Arun, T. Spillover of COVID-19: Impact on the Global Economy. SSRN 2020. [Google Scholar] [CrossRef]
  2. Pogue, M.; Raker, E.; Hampton, K.; Saint Laurent, M.; Mishori, R. Conducting remote medical asylum evaluations in the United States during COVID-19: Clinicians’ perspectives on acceptability, challenges and opportunities. J. Forensic Leg. Med. 2021, 84, 102255. [Google Scholar] [CrossRef] [PubMed]
  3. Zhao, Z.-Q.; Zheng, P.; Xu, S.-T.; Wu, X. Object detection with deep learning: A review. IEEE Trans. Neural Netw. Learn. Syst. 2019, 30, 3212–3232. [Google Scholar] [CrossRef]
  4. Wu, Y.; Zhang, D.; Yin, F.; Zhang, Y. Salient object detection based on global to local visual search guidance. Signal Process. Image Commun. 2022, 102, 116618. [Google Scholar] [CrossRef]
  5. Krause, M.; Brado, M.; Schosser, R.; Bartsch, F.-R.; Gerneth, M.; Kauffmann, G. Diagnostic accuracy in remote expert consultation using standard video-conference technology. Eur. Radiol. 1996, 6, 932–938. [Google Scholar] [CrossRef]
  6. Mori, S.; Suen, C.Y.; Yamamoto, K. Historical review of OCR research and development. Proc. IEEE 1992, 80, 1029–1058. [Google Scholar] [CrossRef]
  7. Smith, R.; Antonova, D.; Lee, D.-S. Adapting the Tesseract open source OCR engine for multilingual OCR. In Proceedings of the International Workshop on Multilingual OCR, Barcelona, Spain, 25 July 2009; pp. 1–8. [Google Scholar] [CrossRef]
  8. Gu, J.T.; Neubig, G.; Cho, K.H.; Li, V.O.K. Learning to translate in real-time with neural machine translation. In Proceedings of the 15th Conference of the European Chapter of the Association for Computational Linguistics, Valencia, Spain, 3–7 April 2017; Volume 1, pp. 1053–1062. [Google Scholar]
  9. Das, H.S.; Roy, P. A CNN-BiLSTM based hybrid model for Indian language identification. Appl. Acoustics 2021, 182, 108274. [Google Scholar] [CrossRef]
  10. Ding, L.A.; Wu, D.; Tao, D.C. Improving neural machine translation by bidirectional training. In Proceedings of the Conference on Empirical Methods in Natural Language Processing, Punta Cana, Dominican Republic, 7–11 November 2021; pp. 3278–3284. [Google Scholar] [CrossRef]
  11. Qi, J.W.; Peng, Y.X. Cross-modal bidirectional translation via reinforcement learning. In Proceedings of the Twenty-Seventh International Joint Conference on Artificial Intelligence (IJCAI-18), Stockholm, Sweden, 13–19 July 2018; pp. 2630–2636. [Google Scholar] [CrossRef]
  12. Shin, J.H.; Georgiou, P.G.; Narayanan, S. Towards modeling user behavior in interactions mediated through an automated bidirectional speech translation system. Comput. Speech Lang. 2010, 24, 232–256. [Google Scholar] [CrossRef]
  13. Ataman, D.; Firat, O.; Di Gangi, M.A.; Federico, F.; Birch, A. On the importance of word boundaries in character-level neural machine translation. In Proceedings of the 3rd Workshop on Neural Generation and Translation, Hong Kong, 4 November 2019; pp. 187–193. [Google Scholar] [CrossRef]
  14. Bulut, S.Ö. Integrating machine translation into translator training: Towards ‘Human Translator Competence’? Translogos Transl. Stud. J. 2019, 2, 1–26. [Google Scholar] [CrossRef]
  15. Bizzoni, Y.; Juzek, T.S.; España-Bonet, C.; Chowdhury, K.D.; van Genabith, J.; Teich, E. How human is machine translationese? Comparing human and machine translations of text and speech. In Proceedings of the 17th International Conference on Spoken Language Translation, Online, 9–10 July 2020; pp. 280–290. [Google Scholar] [CrossRef]
  16. Briones, J. Object Detection with Transformers. Available online: https://medium.com/swlh/object-detection-with-transformers-437217a3d62e (accessed on 28 April 2022).
  17. Carion, N.; Massa, F.; Synnaeve, G.; Usunier, N.; Kirillov, A.; Zagoruyko, S. End-to-end object detection with transformers. In Proceedings of the European Conference on Computer Vision, Glasgow, UK, 23–28 August 2020; pp. 213–229. [Google Scholar] [CrossRef]
  18. Wang, J.F.; Song, L.; Li, Z.M.; Sun, H.B.; Sun, J.; Zheng, N.N. End-to-end object detection with fully convolutional network. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Nashville, TN, USA, 19–25 June 2021; pp. 15844–15853. [Google Scholar] [CrossRef]
  19. Wieniawska, H.W. What Exactly is Happening Inside the Transformer. Available online: https://medium.com/swlh/what-exactly-is-happening-inside-the-transformer-b7f713d7aded (accessed on 28 April 2022).
  20. Du, J. Understanding of object detection based on CNN family and YOLO. J. Phys. Conf. Ser. 2018, 1004, 012029. [Google Scholar] [CrossRef]
  21. Redmon, J.; Farhadi, A. YOLOv3: An incremental improvement. arXiv 2018, arXiv:1804.02767. [Google Scholar]
  22. TensorFlow 2 YOLOv4. Available online: https://bit.ly/3kgtaff (accessed on 28 April 2022).
  23. DETR: End-to-End Object Detection with Transformers. 2020. Available online: https://github.com/facebookresearch/detr (accessed on 28 April 2022).
  24. Lin, T.Y.; Goyal, P.; Girshick, R.; He, K.M.; Dollár, P. Focal loss for dense object detection. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Honolulu, HI, USA, 21–26 July 2017; pp. 2980–2988. [Google Scholar] [CrossRef]
  25. He, Y.H.; Zhu, C.C.; Wang, J.R.; Savvides, M.; Zhang, X.Y. Bounding box regression with uncertainty for accurate object detection. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Long Beach, CA, USA, 16–17 June 2019; pp. 2888–2897. [Google Scholar] [CrossRef]
  26. Li, Z.M.; Peng, C.; Yu, G.; Zhang, X.Y.; Deng, Y.D.; Sun, J. DetNet: Design backbone for object detection. In Lecture Notes in Computer Science; Springer: Berlin/Heidelberg, Germany, 2018; Volume 11213, pp. 339–354. [Google Scholar] [CrossRef]
  27. Shao, S.; Li, Z.M.; Zhang, T.Y.; Peng, C.; Yu, G.; Zhang, X.Y.; Li, J.; Sun, J. Objects365: A large-scale, high-quality dataset for object detection. In Proceedings of the IEEE/CVF International Conference on Computer Vision, Seoul, Korea, 27 October–2 November 2019; pp. 8430–8439. [Google Scholar] [CrossRef]
  28. Zhang, H.K.; Chang, H.; Ma, B.P.; Wang, N.Y.; Chen, X.L. Dynamic R-CNN: Towards high quality object detection via dynamic training. In Lecture Notes in Computer Science; Springer: Berlin/Heidelberg, Germany, 2020; Volume 12360, pp. 260–275. [Google Scholar] [CrossRef]
  29. Tarimoradi, H.; Karami, H.; Gharehpetian, G.B.; Tenbohlen, S. Sensitivity analysis of different components of transfer function for detection and classification of type, location and extent of transformer faults. Measurement 2022, 187, 110292. [Google Scholar] [CrossRef]
  30. GitHub—George Kalitsios: Face Detection Using DETR End-to-End Object Detection with Transformers. 2021. Available online: https://github.com/george-kalitsios/Face-Detection-using-DETR-End-to-End-Object-Detection-with-Transformers (accessed on 28 April 2022).
  31. Ma, T.L.; Mao, M.Y.; Zheng, H.H.; Gao, P.; Wang, X.D.; Han, S.M.; Ding, E.R.; Zhang, B.C.; Doermann, D. Oriented object detection with transformer. arXiv 2021, arXiv:2106.03146v1. [Google Scholar]
  32. Vaswani, A.; Shazeer, N.; Parmar, N.; Uszkoreit, J.; Jones, L.; Gomez, A.N.; Kaiser, L.; Polosukhin, I. Attention is all you need. In Proceedings of the 31st Conference on Neural Information Processing Systems, Long Beach, CA, USA, 4–9 December 2017; pp. 6000–6010. [Google Scholar] [CrossRef]
  33. VoTT—Visual Object Tagging Tool 2020. Available online: https://github.com/microsoft/VoTT (accessed on 28 April 2022).
  34. LabelImg. 2018. Available online: https://github.com/tzutalin/labelImg (accessed on 28 April 2022).
  35. Girshick, R. Fast r-cnn. In Proceedings of the IEEE International Conference on Computer Vision, Santiago, Chile, 7–13 December 2015; pp. 1440–1448. [Google Scholar] [CrossRef]
  36. Nath, U.; Kushagra, S. Better Together: Resnet-50 accuracy with 13× fewer parameters and at 3× speed. arXiv 2020, arXiv:2006.05624v3. [Google Scholar]
  37. Gao, P.; Zheng, M.; Wang, X.; Dai, J.; Li, H. Fast convergence of DETR with spatially modulated co-attention. In Proceedings of the IEEE/CVF International Conference on Computer Vision, Montreal, BC, Canada, 11–17 October 2021; pp. 3621–3630. [Google Scholar] [CrossRef]
  38. Tan, R.J. Breaking Down Mean Average Precision (mAP)—Another Metric for Your Data Science Toolkit. 2019. Available online: https://towardsdatascience.com/breaking-down-mean-average-precision-map-ae462f623a52#1a59 (accessed on 28 April 2022).
  39. Bochkovskiy, A.; Wang, C.Y.; Liao, H.Y.M. YOLOv4: Optimal speed and accuracy of object detection. arXiv 2020, arXiv:2004.10934v1. [Google Scholar]
  40. Honda, H. Digging into Detectron 2—Part 1. Available online: https://medium.com/@hirotoschwert/digging-into-detectron-2-47b2e794fabd (accessed on 28 April 2022).
  41. FAIR’s Research Platform for Object Detection Research, Implementing Popular Algorithms Like MASK R-CNN and RetinaNet. Available online: https://github.com/facebookresearch/Detectron (accessed on 28 April 2022).
  42. Zheng, M.H.; Gao, P.; Zhang, R.R.; Li, K.C.; Wang, X.G.; Li, H.S.; Dong, H. End-to-end object detection with adaptive clustering transformer. In Proceedings of the 32nd British Machine Vision Conference, Online, 22–25 November 2021; pp. 1–13. [Google Scholar]
  43. Dai, Z.G.; Cai, B.L.; Lin, Y.G.; Chen, J.Y. UP-DETR: Unsupervised pre-training for object detection with transformers. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Nashville, TN, USA, 20–25 June 2021; pp. 1601–1610. [Google Scholar] [CrossRef]
Figure 1. DETR architecture as presented in [17].
Figure 1. DETR architecture as presented in [17].
Applsci 12 05977 g001
Figure 2. Detailed DETR transformer architecture as presented in [17].
Figure 2. Detailed DETR transformer architecture as presented in [17].
Applsci 12 05977 g002
Figure 3. DETR results: (a) loss function obtained for the training set (solid line) and the validation set (dashed line) during 50 epochs, for all categories (1st experiment); (b) mAP scores obtained for the validation set.
Figure 3. DETR results: (a) loss function obtained for the training set (solid line) and the validation set (dashed line) during 50 epochs, for all categories (1st experiment); (b) mAP scores obtained for the validation set.
Applsci 12 05977 g003
Figure 4. DETR results: (a) loss function obtained for the training set (solid line) and the validation set (dashed line) during 150 epochs, for all categories (2nd experiment); (b) mAP scores obtained for the validation set.
Figure 4. DETR results: (a) loss function obtained for the training set (solid line) and the validation set (dashed line) during 150 epochs, for all categories (2nd experiment); (b) mAP scores obtained for the validation set.
Applsci 12 05977 g004
Figure 5. DETR results: (a) loss function obtained for the training set (solid line) and the validation set (dashed line) during 1000 epochs, for two categories (4th experiment); (b) mAP scores obtained for the validation set.
Figure 5. DETR results: (a) loss function obtained for the training set (solid line) and the validation set (dashed line) during 1000 epochs, for two categories (4th experiment); (b) mAP scores obtained for the validation set.
Applsci 12 05977 g005
Figure 6. DETR results: (a) loss function obtained for the training set (solid line) and the validation set (dashed line) during 1000 epochs, for all categories (3rd experiment); (b) mAP scores obtained for the validation set.
Figure 6. DETR results: (a) loss function obtained for the training set (solid line) and the validation set (dashed line) during 1000 epochs, for all categories (3rd experiment); (b) mAP scores obtained for the validation set.
Applsci 12 05977 g006
Figure 7. Yolo4 Architecture [39].
Figure 7. Yolo4 Architecture [39].
Applsci 12 05977 g007
Figure 8. Detectron—Detailed Architecture of Base-RCNN-FPN [40].
Figure 8. Detectron—Detailed Architecture of Base-RCNN-FPN [40].
Applsci 12 05977 g008
Figure 9. DETR: OCR Output.
Figure 9. DETR: OCR Output.
Applsci 12 05977 g009
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Share and Cite

MDPI and ACS Style

Biró, A.; Jánosi-Rancz, K.T.; Szilágyi, L.; Cuesta-Vargas, A.I.; Martín-Martín, J.; Szilágyi, S.M. Visual Object Detection with DETR to Support Video-Diagnosis Using Conference Tools. Appl. Sci. 2022, 12, 5977. https://doi.org/10.3390/app12125977

AMA Style

Biró A, Jánosi-Rancz KT, Szilágyi L, Cuesta-Vargas AI, Martín-Martín J, Szilágyi SM. Visual Object Detection with DETR to Support Video-Diagnosis Using Conference Tools. Applied Sciences. 2022; 12(12):5977. https://doi.org/10.3390/app12125977

Chicago/Turabian Style

Biró, Attila, Katalin Tünde Jánosi-Rancz, László Szilágyi, Antonio Ignacio Cuesta-Vargas, Jaime Martín-Martín, and Sándor Miklós Szilágyi. 2022. "Visual Object Detection with DETR to Support Video-Diagnosis Using Conference Tools" Applied Sciences 12, no. 12: 5977. https://doi.org/10.3390/app12125977

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop