Next Article in Journal
Interface Crack Approaching a Three-Material Joint
Previous Article in Journal
Probiotic Characteristics and Antifungal Activity of Lactobacillus plantarum and Its Impact on Fermentation of Italian Ryegrass at Low Moisture
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Object Detection Based on Multiple Information Fusion Net

1
College of Information Science and Technology, Northeast Normal University, Changchun 130000, China
2
Institute for Intelligent Elderlycare, College of Humanities and Sciences, Northeast Normal University, Changchun 130000, China
*
Authors to whom correspondence should be addressed.
Appl. Sci. 2020, 10(1), 418; https://doi.org/10.3390/app10010418
Submission received: 23 November 2019 / Revised: 29 December 2019 / Accepted: 1 January 2020 / Published: 6 January 2020
(This article belongs to the Section Computing and Artificial Intelligence)

Abstract

:
Object detection has been playing a significant role in computer vision for a long time, but it is still full of challenges. In this paper, we propose a novel object detection framework based on relationship among different objects and the scene-level information of the whole image to cope with the problem that some strongly correlated objects are difficult to be recognized. Our motivation is to enrich the semantics of object detection feature by a scene-level information branch and a relationship branch. There are three important changes of our framework over traditional detection methods: representation of relationship, scene-level information as the prior knowledge and the fusion of the above two information. Extensive experiments are carried out on PASCAL VOC and MS COCO databases. The experimental results show that the detection performance can be improved by introducing relationship and scene-level information, and our proposed model achieve better performance than several classical and state-of-the-art methods.

1. Introduction

Object detection is a hot topic in the field of computer vision and machine learning due to their widely applications in autonomous driving, robots, video surveillance, pedestrian detection, and so on. The classical object detection techniques are mainly based on the use of manual features, which can be divided into three steps: (1) target area selection; (2) feature extraction; (3) classification. In the first step, sliding-window strategy [1] which utilizes the sliding-windows with different dimensions and length-width ratios is widely adopted to search for candidate regions exhaustively. In the second step, the candidate regions obtained in the first step are analyzed. Several techniques can be used in this step for feature extraction, such as scale-invariant feature transform (SIFT) [2], histogram of oriented gradients (HOG) [3] and speeded-up robust features (SURF) [4]. In the third step, the candidate regions are classified according to the features extracted in the previous step by using classifiers such as support vector machine (SVM) [5] and AdaBoost [6]. Although the classical methods have been adopted in some object detection problems, there are still some limitations that hinder their breakthrough in speed and accuracy. Firstly, since the sliding-window strategy will capture many candidate regions in the original image, and the feature of regions needs to be extracted one by one, the classical object detection approaches are time-consuming. Secondly, the classical object detection methods may lack robustness because artificially designed features are sensitive to the variance in morphology, illumination and occlusion of object.
Recently, some deep learning techniques have been applied to object detection to overcome the limitations of traditional approaches [7,8,9,10,11,12,13]. The current state-of-the-art detector based on the deep learning can be roughly divided into two categories. One is two-stage methods which first form a series of candidate object proposals by Selective Search [8], EdgeBoxes [9], DeepMask [12] or region proposal network (RPN) [7], and then input the proposals into convolutional neural network for classification. The other is one-stage methods which straightforwardly predict confidences and locations of multiple objects on the whole feature map without generating candidate object proposals.
Region-based convolutional network (R-CNN) [14], as the beginning of combining object detection and deep learning, is a representative two-stage based approach. It achieves excellent object detection accuracy by extracting CNN features from the candidate regions and applies linear SVMs as the classifier. However, since the ConvNet forward pass is performed for each object proposal independently, the computational cost of R-CNN is high. Furthermore, the multi-stage training strategy which contains feature extraction, fine-tuning network, training SVMs, and bounding-box regression also makes the training of R-CNN be slow. In [15], a spatial pyramid pooling network (SPPNet) was proposed. Although SPPNet can speed up R-CNN by sharing computation, its training is also a multi-stage pipeline. Besides, the fine-tuning algorithm proposed in SPPNet cannot update the convolutional layer, which limits its accuracy when the networks are very deep. For the sake of further decreasing the computational cost and improving the accuracy of object detection, Ross et al. proposed a fast region-based convolutional network (Fast R-CNN) [16]. The Fast R-CNN utilizes a novel RoI-pooling operation to extract feature vectors for each candidate region from shared convolutional feature map, which greatly improves the processing speed. In Fast R-CNN, the detection accuracy can also be enhanced by updating all network layers during training. Although SPPNet and Fast R-CNN have effectively reduced the training time of object detection networks, the region proposal computation is still considered as a bottleneck in them. To deal with this issue, Ren et al. proposed a Faster R-CNN [7] which replaces the Selective Search method with RPN to achieve end-to-end training. RPN is a kind of fully convolutional network (FCN) [17]. By sharing full-image convolutional features with the detection network, RPN enables nearly cost-free region proposals to solve the time-consuming problem of Fast R-CNN. However, the multiple scale proposals generated by sliding a fixed set of filters over a fixed set of convolutional feature maps in RPN may be inconsistent with the sizes of objects. Thus, Cai et al. proposed a multi-scale CNN (MS-CNN) [18] to match the receptive fields to different scales of objects and employed a multiple output layer for object detection. Recently, for the purpose of improving the detection performance, some more state-of-the-art techniques (such as Resnet [19] and Inception series [20,21,22]) were employed to replace the standard CNN as the backbone networks of the two-stage based object detection methods, which can be found in the object detection API from google [23].
Different from the aforementioned methods, the one-stage approaches can achieve complete single network training under the premise of guaranteeing a certain accuracy rate. You only look once (YOLO) [24], YOLO9000 [25], an iterative grid based object detector (G-CNN) [26], and single shot multibox detector) [27] are representative techniques in this category. Through treating the object detection task as a regression problem, YOLO spatially separates bounding boxes and associated class probabilities. Since the whole detection pipeline of YOLO is a single network, an end-to-end optimization of the network can be directly performed. SSD combines predictions of multiple feature maps with different resolutions to detect objects of various sizes. Since the proposal generation, subsequent pixel and feature resampling stages are eliminated in SSD, it can be easily trained. Although the running speed of one-stage methods can be significantly improved, their accuracy is always inferior to the two-stage approaches [27]. To address this issue, the Resnet and Inception have also been utilized [23]. Furthermore, Lin et al. replaced the standard cross entropy loss with a novel Focal loss and proposed a RetinaNet to solve the class imbalance problem in one-stage based object detection [28].
No matter the approach belongs to one-stage or two-stage, most the aforementioned algorithms do not effectively utilize the relationship among objects, but only use the feature associated with the object itself for detection. Recently, some researchers have realized the importance of relation, and proposed some methods [29,30,31] to achieve better detection results by exploring the relationships between objects. In ION [32], Bell et al. proposed a spatial recurrent neural networks (RNNs) for exploring contextual information across the entire image. Xu et al. put forward a scene graph generation approach by iterative message passing [33]. The network regards a single object as a point in topology, and the relationships of objects are considered as edges connecting points. Through passing information between the edges and points, it is proved that the relationship between objects has a positive impact on detection. Georgia et al. proposed a human-centric based model called InteractNet [34], in which human is regarded as the main clue to establish a relationship with other surrounding objects. The InteractNet indicates that a person’s external behavior can provide powerful information to locate the objects they are interacting with. Liu et al. proposed a structure inference net (SIN) [35] which explores the structure relationship between objects for detection. However, SIN only takes the spatial coordinates of object proposals into account, while the appearance feature of object proposals is neglect. Han et al. presented a relation network [36], which considers both the appearance and geometry feature of object proposals for relation construction. Nevertheless, the scene-level feature which could provide a lot of context information for object detection [37] is ignored in relation Nntwork.
This paper proposes a novel object detection algorithm based on multiple information fusion net (MIFNet). Compared with the existing techniques, our algorithm not only adaptively establishes relationships between objects through attention mechanism [38], but also introduces scene-level information to make the proposed approach richer in semantics. In MIFNet, the relationships between an object and all other objects are got by relation channel modules. Besides, by introducing the scene-level context [21,39,40], the proposed network can enrich the object feature with scene information. The experimental results on PASCAL VOC [41] and MS COCO [42] databases demonstrate the effectiveness of the proposed algorithm.
The paper is structured as follows. The related work is introduced in Section 2. The proposed MIFNet is described in Section 3. The experimental results are given in Section 4. The conclusion is provided in Section 5.

2. Related Work

Context information: In real life, it is unlikely that an object can exist alone. Visual objects occur in particular environments and usually coexist with other related objects [43]. When the object’s appearance feature is insufficient because of small object size, object occlusion, or poor image quality, a proper modeling of context will facilitate object detection and recognition task. Context information has been applied in many methods to enhance the performance of object detection [44,45,46,47,48,49], which can be roughly divided into two categories [49,50]: global information [32,51] (refers to the image level or scene level information), local information [35,36] (considers the object relationship or the interaction between the object and its surrounding area). It is proved that both the global and local context information have a positive impact on the object detection. Our proposed MIFNet has the capability of utilizing both global context (scene-level information) and local information (object relationship) to make the object’s appearance feature richer.
Attention mechanism: The attention mechanism in deep learning is inspired by the mode of human attention thinking and has been widely used in natural language processing [52]. In attention module, an individual element can be influenced by aggregating information from other elements and the dependency between elements is modeled without excessive assumptions on their locations and feature distributions. The aggregation weights can be learned automatically, which is driven by the task goal. Recently, attention mechanism has been successfully applied in vision problems [37,53].
Attention mechanism can be represented as follows:
Attention   ( Q ,   K ,   V )   =   softmax   ( Q K T d V ) ,
where Q, K, V are three feature matrices with the same shape. In natural language processing, Q, K, V represent sets of queries, keys, and values, respectively. In our work, Q denotes the object feature, K denotes all surrounding object feature, V represents the all image feature (with object feature and location feature). In Equation (1), given a query qQ and all keys K, the dot products of q with K will be calculated to get the similarity between them. Then we divide dot products by a scaling factor d and the softmax function is applied to obtain the weights on all image feature (i.e., it can obtain the influence of each object on the current object). For more detailed information about the attention mechanism, the readers can refer to [38]. In our work, the attention mechanism is utilized to get the relationship between objects.

3. The Proposed Method

The framework of proposed multiple information fusion net (MIFNet) is shown in Figure 1. In our MIFNet, the feature map of an input image is first obtained through a feedforward convolutional network (VGG or Resnet). In the next stage, the network feature map is divided into two parts. One is as a part of the input of the first branch and the other is utilized to get the region proposals through RPN and then served as the input of the second branch. In the first branch network (I), a series of operations is performed on the feature map of the entire image to get the scene-level information as the input of scene GRU (Gated Recurrent Unit in III up). In the second branch network (II), the attention mechanism is utilized to establish object relationships adaptively. For the purpose of classifying and regressing regions of interest (RoIs), the second branch network not only utilizes the appearance feature extracted by convolutional layers and the coordinate information of the object, but also the information of all surrounding objects as the input of relation GRU (in III below). In the message passing module (III), scene GRUs and relation GRUs communicate information to each other in order to keep up with new information. In the last stage, we concatenate the information obtained by these two GRUs to refine the position of the corresponding RoI and predict the category of objects.

3.1. Scene-Level Information Processing Module

Contextual information is important for accurate object recognition. To extract the scene-level information, the image feature is firstly obtained by convolutional network (VGG or Resnet) as the input of the first branch. Secondly, the image feature obtained by RoI-pooling layer and the feature obtained by RPN (without scene information) are concatenated as the input of a convolutional layer. By concatenation, the information of potential object is richer. Besides, the weight of potential object can also be increased by training. In the end, the output fs of the first network branch, which called scene feature, is input to the scene GRU to choose information and update object feature.

3.2. Relationship Module

In most previous object detection methods based on the convolutional neural network [7,16], each object is identified independently, and the relationship of objects is neglected. To overcome this limitation, the proposed approach models the relationship of objects by groups. That is, the feature vector of an object is obtained by fusing the features of itself and other objects to enrich the information, as shown in Figure 2.
In Figure 2, given an input set of N objects { ( f t n ,   f c n ) } n   =   1 N , in which f t n is the original appearance feature of the nth object extracted by convolutional neural network, f c n denotes the location feature of the nth object composed by the 4-dimensional feature of object bounding box. The bounding box feature comprises width (w), height (h) and the center coordinates (x, y) of the box in our study. The relationship channel is a module that handles relationships among different objects, Nr is the number of channels (Nr = 64). By object relation module, the f c 1 ( n ) ,   f c 2 ( n ) f c N r ( n ) which fuse the location information of all surrounding objects can be gained. For the purpose of obtaining the output f t n that is finally sent into the relation GRU, we concatenate the vectors on all channels f c 1 ( n ) , f c 2 ( n ) f c N r ( n ) . Because the processing mechanisms of relation channel modules are the same, we take one relation channel module as an example to explain how relation channel works.
Figure 3 shows the process of one relation channel module. Firstly, the dot product operation is applied to obtain the appearance weight w t m n between the mth and nth objects, as shown in Equation (2).
w t m n   =   ( W K f t m ) · ( W Q f t n ) d ,
where W K , W Q are matrices which map the original appearance f t m and f t n into subspaces, · denotes the operation of dot product to obtain the degree of matching between W K f t m and   W Q f t n .
Secondly, the location weight w g m n is calculated by Equation (3).
w g m n   =   Relu ( W g · ε g ( f c m , f c n ) ) ,
where f c m and f c n are geometry features that contain six relative position information ( log ( | x m x n | w n ) , log ( | y m y n | h n ) , log ( w m w n ) , log ( h m h n ) , x m x n w n , y m y n h n ), in which wn, hn, xn and yn are the width, height and center coordinates of the nth object, ε g is a function based on sine and cosine to embed the geometry features into high-dimensional space [38]. Then we use W g to convert the embedded vector to scalar weight. The Relu activation function is utilized to ensure that only objects with certain geometric relationship can participate in this relationship calculation.
Next, the relationship weight is obtained by Equation (4).
w m n   =   softmax ( w g m n · exp ( w t m n ) ) ,
where the relationship weight w m n represents the impact of the mth object to the nth object, softmax is employed for normalization. Finally, Equation (5) can be utilized to get a feature f c N r ( n ) that has the influence of surrounding objects on it.
f c N r ( n )   =   m w m n · ( W v f t n ) ,  
where W v is used to transform the original appearance feature f t n linearly. Equation (5) is the process of integrating the information of the object and other objects into the original appearance feature. The output f c N r ( n ) is the weighted sum of the initial appearance features from other objects, which contains both its original appearance feature and the feature of all objects around it.
In the end, by the relation channel module, the feature f t n which merges features of multiple channels can be gained by Equation (6).
f t n   =   f t n + [ f c 1 ( n ) , , f c N r ( n ) ] , f o r   a l l   n .
where the fusion feature f t n includes the extracted original appearance feature f t n (the initial appearance feature after convolutional layers) and the relationship feature ( f c 1 ( n ) , , f c N r ( n ) ) (fusing the location information of all surrounding objects under a particular channel). In the relation channel, the feature of other objects can be mixed together to identify the relationship between the current object and other objects, and finally merged with the original appearance feature through the fully connected network. The final output f t n is the input of the Relation GRU.

3.3. Message Passing Module

As we have discussed previously, context information is important for accurate object detection. For example, in Figure 4a, if road is considered as the global or scene-level information, the objects in this image are hardly to be detected as ships and planes since it is generally impossible for them to appear in the road scene. Similarly, in Figure 4b, when a dinner table appears, the probability of detecting chairs increases, because the dinner tables and chairs always appear in pairs. Thus, the Gated Recurrent Unit (GRU) [54] is utilized in this study. Similar to the long short-term memory (LSTM) model [55], GRU unit also has the function of adjusting the information flow in the unit, but it is lightweight and effective [35]. In the message passing module, information is continuously passed between the scene GRU and the relation GRU so that the useful information can be preserved.
GRU only has two gates. It combines the input gate and the forget gate in the LSTM into one, and the combined gate is called update gate, which determines how much information from the previous time and the current time is to be passed on. The other gate in GRU is reset gate, which controls how much past information is forgotten.
In our work, two parallel GRUs are applied to pass information to each other, one is the scene GRU and the other is the relation GRU. The scene GRU receives the whole image information fs as the input. The input of relation GRU is the integrated object information f t n , which includes the object’s own information and the influence of surrounding objects on it. We represent the initial state h i of the network with the original appearance feature (without any scene information or relation information). Here, since the processing mechanisms of scene GRU and relation GRU are identical, we take the relation GRU as an example to show how GRU works. Firstly, the reset gate of the tth moment r t is calculated as follows:
r t   =   σ ( W r [ f t n , h i ] ) ,
where σ is the logistic sigmoid function, [,] denotes the concatenation of vectors, and W r is a weight matrix learned through the convolutional neural network. The output of reset gate r t determines whether the previous state is forgotten. When r t is close to zero, the status information h i of the previous moment will be forgotten, and the hidden status is reset to the current input. Similarly, the update gate of the tth moment z t is computed by
z t   =   σ ( W z [ f t n , h i ] ) ,
where z t is used to determine how much past information can continue to be passed on, W z is a weight matrix. If the value of the update gate is larger, the state information introduced at the previous moment is more, and vice versa. In GRU, the new hidden state h ˜ t can be obtained through Equation (9):
h ˜ t   =   t a n h ( W [ r t h i , f t n ] ) ,  
where the new hidden state h ˜ t is determined by the value of the reset gate, W is a weight matrix, denotes the element-wise multiplication. The actual output h i + 1 is then computed by
h i + 1   =   1 z t h i + z t h ˜ t ,
where some of the previous state hi will be passed, and the new hidden state h ˜ t will be selectively updated. Through GRU, the scene module and the relation module can pass information to each other and constantly update new information. In this way, useful information will continue to be passed, and useless information will be ignored. Finally, richer information can be obtained through Equation (11).
h i   =   [ h i + 1 s , h i + 1 r ] 2 ,
where h i + 1 s represents the information obtained by the scene GRU, and h i + 1 r denotes the information obtained by the relation GRU. The integrated information h i will be sent into the next GRUs as the new initial state.

4. Experimental Results

4.1. Experimental Settings

Databases and evaluation metrics: Our model was evaluated on two databases: PASCAL VOC [41] and MS COCO [42]. PASCAL VOC is a widely used image database for object detection and classification. In our work, VOC 2007 and VOC 2012 subsets were utilized. VOC 2007 data set contained 9963 annotated images and 24,640 annotated objects, which were composed of three parts: train, validation (val), and test sets. VOC 2012 is an updated version of VOC 2007 data set, which included 11,530 pictures with 20 categories including people, animals (such as cats, dogs and birds), vehicles (such as cars, ships and planes), and furniture (such as chairs, tables, and sofas). Some examples of PASCAL VOC database can be seen in Figure 5. MS COCO database is built by Microsoft, which contained 328,000 images with 2000 object labels. Compared with PASCAL VOC, MS COCO includes natural images and common object images in daily life with more complex background, larger number of targets and smaller size. Thus, the object detection task on MS COCO database was more difficult and challenging. The sample images in MS COCO are shown in Figure 6. In this study, we followed Ref. [35] to adopt average precision (AP) and mean of average precision (mAP) as our evaluation metrics to compare the performances of different approaches. AP, which derived from precision and recall, is one of the popular metrics to measure the accuracy of object detectors. It computes the average precision value for recall value over 0 to 1. The precision and the recall rates are defined as follows:
p r e c i s i o n   =   T P T P + F P ,
r e c a l l   =   T P T P + F N ,
where T P represents the number of true positive examples, F P represents the number of fault positive examples, and F N represents the number of fault negative examples. mAP is the average of AP on all categories.
Implementation details: In our experiments, the proposed model was implemented based on Faster R-CNN [7], an open-source framework for object detection built on Tensorflow [56] platform. The VGG-16 [57] and Resnet-101 [19] pre-trained on ImageNet [58] were adopted as backbone networks in our model to extract image feature. When adding the newly fully connected and convolutional layers, they were randomly initialized with a zero-mean Gaussian distribution with standard deviations of 0.01 and 0.001. The message passing module contained two parallel GRU units with shared parameters. All the parameters of the GRU units were initialized based on SIN [35]. Non-maximum suppression (NMS) with intersection over union (IOU) were used for duplicate removal in all experiments.
Stochastic gradient descent (SGD) was applied to fine tune our network. Each SGD mini batch was composed of 256 randomly sampled object proposals from two randomly chosen images. In each mini-batch, 25% of the RoIs were selected as foreground from object proposals, which had IOU overlapped with a ground-truth bounding box of at least 0.5. We sampled the remaining RoIs from object proposals which had a maximum IOU with ground truth in the interval [0:1; 0:5]. We trained our model on a single NVIDIA GeForce GTX TITAN X GPU with 12 GB memory. The experimental parameters, training and test time of our MIFNet can be seen in Table 1.

4.2. Performance Comparisons

PASCAL VOC Database: The performance of our MIFNet was compared with some classical and state-of-the-art object detection methods, including Fast R-CNN [16], Faster R-CNN [7], SIN [35], ION [32], CPF [40], and so on. The experimental results on data set VOC 2007 test and VOC 2012 test are shown in Table 2 and Table 3 respectively. All the experimental results of comparison approaches are quoted from their corresponding literatures. From these tables, the following points can be observed. Firstly, Fast R-CNN and Faster R-CNN are classical two-stage approaches, while SSD is a classical one-stage approach. Since the relationship between objects and context information were neglected in them, their performances were inferior to other approaches. Secondly, by utilizing the spatial recurrent neural networks and semantic segmentation, ION and CPF took global contextual information into account. Therefore, they outperformed the classical approaches such as Fast R-CNN, Faster R-CNN, and SSD. Thirdly, SIN considered both the scene context information and object relationships. However, the relationship in SIN was established only by geometric structure of the objects, which neglected the objects’ appearance information. As a result, its performance was still worse than the proposed MIFNet. At last, the proposed approach leveraged the attention mechanism to adaptively establish the relationship between objects, which considered both geometric and appearance information. Besides, the scene-level information was also introduced into the model. Thus, our MIFNet greatly improved the detection accuracy of some small and highly correlated objects (such as chair, boat, plant, tv) and achieved the best performance on PASCAL VOC database.
MS COCO database: In order to further verify the effectiveness of our proposed method, MS COCO database was utilized. The object detection results of different methods on this database are tabulated in Table 4. In this table, AP was averaged precision across all object categories and multiple intersection over union (IOU) values from 0.5 to 0.95, AP50 denotes the mAP at IOU = 0.50, AP70 denotes the mAP at IOU = 0.70, average recall (AR) represents the average recall rate averaged over all categories and IOU thresholds. AR1, AR10 and AR100 denote the maximum recall rate of the fixed number (1, 10, 100) of objects detected in each image, ARS, ARM and ARL represent the recall rate of small (area smaller than 322), medium (area between 322 and 962) and large (area bigger than 962) objects respectively. From Table 4, we can get the following observations. Firstly, since ION takes global context information into consideration, its performance is better than the classical approaches such as Fast R-CNN, Faster R-CNN and YOLOv2. Secondly, SIN outperforms ION, which indicates the object relationship is important for object detection. Finally, the proposed MIFNet performs best on MS COCO database because it establishes the relationship between objects by both geometric and appearance information adaptively and takes the scene-level information into account. In summary, these observations are consistent with the experimental results of PASCAL VOC database.
Effectiveness of scene-level information: For the purpose of verifying the effectiveness of each part in our proposed method, some ablation experiments are carried out. Here, we employ the VGG-16 as the backbone of our MIFNet. In the first experiment, only the scene-level information is considered to update the object feature. As shown in Table 5 and Table 6, applying scene-level information achieves a better mAP of 75.8% compared with the baseline (Faster R-CNN without scene-level information and object relationships) on PASCAL VOC 2007. On the bigger database MS COCO, a better mAP of 23.5% can be obtained. We find that introducing the scene-level information can improve the detection performances in certain categories, including bike, bottle, chair, plant, TV, and so on. Especially, the average accuracy of plant has increased by more than 10%. These results are not surprising since these categories are usually highly relevant to the context of scene. From Figure 7, we can clearly see that the detection result in (b) with scene-level information is more accurate than the detection result in (a) without scene-information. This may because the probability of plant appearing increases with the introducing of balcony information.
Effectiveness of Relation and Relation Settings: In the second ablation experiment, the validity of relationship information is evaluated. Here, the scene-level information is ignored in our model and we only use a set of Relation GRUs for object detection. Experiments are performed on the PASCAL VOC and MS COCO databases, respectively. From the experimental results shown in Table 7 and Table 8, we can see that the performance of our model with only relationship information is still superior to the baseline (Faster R-CNN), especially for highly correlated objects. Taking the detection results in Figure 8 as an example. It is clear that due to the introduction of relation information, the tables and chairs which are strongly correlated with each other can be more accurately detected. This indicates the object relationship is very important for detection.
Through the above experiments, we can clearly know that both the scene-level information and object relationship are beneficial for detection, and they are indispensable. Figure 9 shows the detection results of some algorithms and our model. It can be seen that our model performs better when detecting small and highly correlated objects (such as driver, table, chair) due to the message passing between scene-level information and object relationship. At the same time, for some objects with strong correlation with the scene, the detection result is also well (such as the boats in the sea scene and aeroplanes in the sky scene).
In order to test whether the number of relation modules will influence the detection result of our MIFNet, we also conduct an experiment to compare the performances of the proposed model with different numbers of relation modules. As shown in Table 9, we find that with the increase of the number of relation modules, the detection accuracy of our model will gradually decrease. This may because too many relation modules will make the network over-associate two objects. For example, once an object appears near the table, it will be detected as a chair regardless its feature. Therefore, we choose one module in the experiment.
Effectiveness of GRU Settings and the inputs of GRU: In our network, multiple parallel GRU units are used to fuse the information of scene-level context and object relationship. In order to study the effectiveness of different GRU settings, several experiments are conducted. Firstly, we build the message passing module with different numbers (1 to 3) of GRU units and test their performances. From the experimental results in Table 10, it can be found that when the number of stacked GRU units increases from 1 to 2, the mAP decreases. In addition, when the number of stacked GRU units increases from 2 to 3, no significant performance change can be observed. This indicates that one stacked GRU is enough for our proposed MIFNet.
Then, for the purpose of verifying the effectiveness of the message passing module in our MIFNet, we compare the experimental performances of two methods. The first method uses the scene-level information and the object relationship information as inputs to the different GRUs, which is the strategy employed in our approach. The second method is to concatenate the scene-level information and the object relationship information as one vector and then input this vector to only one GRU. From the experimental results in Table 11, the detection performances of two different methods are 77.6% and 76.2%, respectively. It is clear that the first method obtains better detection results since different information can be effectively transmit to each other through the two groups of GRUs in it. Nevertheless, the second method which directly concatenates different information cannot accomplish information transmission.

5. Conclusions

This paper proposed a network that fuses both the scene-level and relationship information for object detection in images. Compared with other methods, the most important advantage of our approach is that we leverage the attention mechanism to model the relationship between objects adaptively. Besides, the relationship weights are obtained using not only the geometric structure, but also the appearance feature of the objects. At last, the scene-level information is also considered in our model. Two widely used databases are employed in our experiment. From the experimental results, we can see that through fusing the scene-level and object relationship information, our proposed MIFNet outperforms some classical and state-of-the-art approaches. Furthermore, some ablation experiments are also carried out to test the effectiveness of our MIFNet.

Author Contributions

Conceptualization, Y.Z. and J.W.; methodology, Y.Z. and J.W.; validation, Y.Z., J.W., and J.K.; formal analysis, M.Q.; data curation, Y.Z. and J.W.; software, Y.Z. and Y.L. (Yunpeng Liu); writing—original draft preparation, Y.Z.; writing—review and editing, J.W. and M.Q.; supervision, Y.L. (Yinghua Lu); project administration, J.K. and Y.L. (Yinghua Lu); funding acquisition, J.K. and Y.L. (Yinghua Lu). All authors have read and agreed to the published version of the manuscript.

Funding

This work was supported in part by the National Natural Science Foundation of China (Nos. 61672150, 61907007, 61702092), by the Fund of the Jilin Provincial Science and Technology Department Project (Nos. 20180201089GX, 20190201305JC) and the Fundamental Research Funds for the Central Universities (No. 2412019FZ049).

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Felzenszwalb, P.F.; Girshick, R.B.; McAllester, D.; Ramanan, D. Object detection with discriminatively trained part-based models. IEEE Trans. Pattern Anal. Mach. Intell. 2009, 32, 1627–1645. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  2. Lowe, D.G. Distinctive image features from scale-invariant keypoints. Int. J. Comput. Vis. 2004, 60, 91–110. [Google Scholar] [CrossRef]
  3. Dalal, N.; Triggs, B. Histograms of Oriented Gradients for Human Detection. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, San Diego, CA, USA, 20–25 June 2005; pp. 886–893. [Google Scholar]
  4. Bay, H.; Tuytelaars, T.; Van Gool, L. Surf: Speeded up robust features. In European Conference on Computer Vision; Springer: Berlin, Germany, 2006; pp. 404–417. [Google Scholar]
  5. Suykens, J.A.; Vandewalle, J. Least squares support vector machine classifiers. Neural Process. Lett. 1999, 9, 293–300. [Google Scholar] [CrossRef]
  6. Freund, Y.; Schapire, R.E. A decision-theoretic generalization of on-line learning and an application to boosting. J. Comput. Syst. Sci. 1997, 55, 119–139. [Google Scholar] [CrossRef] [Green Version]
  7. Ren, S.; He, K.; Girshick, R.; Sun, J. Faster r-cnn: Towards Real-Time Object Detection with Region Proposal Networks; Neural Information Processing Systems Foundation, Inc.: La Jolla, CA, USA, 2015; pp. 91–99. [Google Scholar]
  8. Uijlings, J.R.; Van De Sande, K.E.; Gevers, T.; Smeulders, A.W. Selective search for object recognition. Int. J. Comput. Vis. 2013, 104, 154–171. [Google Scholar] [CrossRef] [Green Version]
  9. Zitnick, C.L.; Dollár, P. Edge boxes: Locating object proposals from edges. In European Conference on Computer Vision; Springer: Cham, Switzerland, 2014; pp. 391–405. [Google Scholar]
  10. Shi, W.; Bao, S.; Tan, D. FFESSD: An Accurate and Efficient Single-Shot Detector for Target Detection. Appl. Sci. 2019, 9, 4276. [Google Scholar] [CrossRef] [Green Version]
  11. He, W.; Huang, Z.; Wei, Z.; Li, C.; Guo, B. TF-YOLO: An Improved Incremental Network for Real-Time Object Detection. Appl. Sci. 2019, 9, 3225. [Google Scholar] [CrossRef] [Green Version]
  12. Pinheiro, P.O.; Collobert, R.; Dollár, P. Learning to segment object candidates. In Proceedings of the Conference on Neural Information Processing Systems, Montreal, QC, Canada, 7–12 December 2015; pp. 1990–1998. [Google Scholar]
  13. Jiang, Y.; Peng, T.; Tan, N. CP-SSD: Context Information Scene Perception Object Detection Based on SSD. Appl. Sci. 2019, 9, 2785. [Google Scholar] [CrossRef] [Green Version]
  14. Girshick, R.; Donahue, J.; Darrell, T.; Malik, J. Rich feature hierarchies for accurate object detection and semantic segmentation. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Columbus, OH, USA, 23–28 June 2014; pp. 580–587. [Google Scholar]
  15. He, K.; Zhang, X.; Ren, S.; Sun, J. Spatial pyramid pooling in deep convolutional networks for visual recognition. IEEE Trans. Pattern Anal. Mach. Intell. 2015, 37, 1904–1916. [Google Scholar] [CrossRef] [Green Version]
  16. Girshick, R. Fast r-cnn. In Proceedings of the IEEE International Conference on Computer Vision, Santiago, Chile, 7–13 December 2015; pp. 1440–1448. [Google Scholar]
  17. Dai, J.; Li, Y.; He, K.; Sun, J. R-fcn: Object detection via region-based fully convolutional networks. In Proceedings of the Conference on Neural Information Processing Systems, Barcelona, Spain, 5–10 December 2016; pp. 379–387. [Google Scholar]
  18. Cai, Z.; Fan, Q.; Feris, R.S.; Vasconcelos, N. A unified multi-scale deep convolutional neural network for fast object detection. In European Conference on Computer Vision; Springer: Cham, Switzerland, 2016; pp. 354–370. [Google Scholar]
  19. He, K.; Zhang, X.; Ren, S.; Sun, J. Deep residual learning for image recognition. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Las Vegas, NV, USA, 27–30 June 2016. [Google Scholar]
  20. Ioffe, S.; Szegedy, C. Batch normalization: Accelerating deep network training by reducing internal covariate shift. arXiv 2015, arXiv:1502.03167. [Google Scholar]
  21. Szegedy, C.; Vanhoucke, V.; Ioffe, S.; Shlens, J.; Wojna, Z. Rethinking the inception architecture for computer vision. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Las Vegas, NV, USA, 27–30 June 2016; pp. 2818–2826. [Google Scholar]
  22. Szegedy, C.; Ioffe, S.; Vanhoucke, V.; Alemi, A.A. Inception-v4, inception-resnet and the impact of residual connections on learning. arXiv 2016, arXiv:1602.07261. [Google Scholar]
  23. Huang, J.; Rathod, V. Supercharge your Computer Vision Models with the TensorFlow Object Detection API. Google AI Blog, 15 June 2017. Available online: https://ai.googleblog.com/2017/06/supercharge-your-computervision-models.html (accessed on 4 January 2020).
  24. Redmon, J.; Divvala, S.; Girshick, R.; Farhadi, A. You only look once: Unified, real-time object detection. In Proceedings of the International Conference on Computer Vision & Pattern Recognition, Las Vegas, NV, USA, 27–30 June 2016; pp. 779–788. [Google Scholar]
  25. Redmon, J.; Farhadi, A. YOLO9000: Better, faster, stronger. arXiv 2017, arXiv:1612.08242. [Google Scholar]
  26. Najibi, M.; Rastegari, M.; Davis, L.S. G-cnn: An iterative grid based object detector. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Las Vegas, NV, USA, 27–30 June 2016; pp. 2369–2377. [Google Scholar]
  27. Liu, W.; Anguelov, D.; Erhan, D.; Szegedy, C.; Reed, S.; Fu, C.Y.; Berg, A.C. Ssd: Single shot multibox detector. In European Conference on Computer Vision; Springer: Cham, Switzerland, 2016; pp. 21–37. [Google Scholar]
  28. Lin, T.Y.; Goyal, P.; Girshick, R.; He, K.; Dollár, P. Focal loss for dense object detection. In Proceedings of the IEEE International Conference on Computer Vision, Venice, Italy, 22–29 October 2017; pp. 2980–2988. [Google Scholar]
  29. Jae Hwang, S.; Ravi, S.N.; Tao, Z.; Kim, H.J.; Collins, M.D.; Singh, V. Tensorize, factorize and regularize: Robust visual relationship learning. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Salt Lake City, UT, USA, 18–22 June 2018; pp. 1014–1023. [Google Scholar]
  30. Krishna, R.; Chami, I.; Bernstein, M.; Fei-Fei, L. Referring relationships. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Salt Lake City, UT, USA, 18–22 June 2018; pp. 6867–6876. [Google Scholar]
  31. Chen, X.; Li, L.J.; Fei-Fei, L.; Gupta, A. Iterative visual reasoning beyond convolutions. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Salt Lake City, UT, USA, 18–22 June 2018; pp. 7239–7248. [Google Scholar]
  32. Bell, S.; Lawrence Zitnick, C.; Bala, K.; Girshick, R. Inside-outside net: Detecting objects in context with skip pooling and recurrent neural networks. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Las Vegas, NV, USA, 27–30 June 2016; pp. 2874–2883. [Google Scholar]
  33. Xu, D.; Zhu, Y.; Choy, C.B.; Fei-Fei, L. Scene graph generation by iterative message passing. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Honolulu, HI, USA, 21–26 July 2017; pp. 5410–5419. [Google Scholar]
  34. Gkioxari, G.; Girshick, R.; Dollár, P.; He, K. Detecting and recognizing human-object interactions. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Salt Lake City, UT, USA, 18–22 June 2018; pp. 8359–8367. [Google Scholar]
  35. Liu, Y.; Wang, R.; Shan, S.; Chen, X. Structure inference net: Object detection using scene-level context and instance-level relationships. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Salt Lake City, UT, USA, 18–22 June 2018; pp. 6985–6994. [Google Scholar]
  36. Hu, H.; Gu, J.; Zhang, Z.; Dai, J.; Wei, Y. Relation networks for object detection. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Salt Lake City, UT, USA, 18–22 June 2018; pp. 3588–3597. [Google Scholar]
  37. Bahdanau, D.; Cho, K.; Bengio, Y. Neural machine translation by jointly learning to align and translate. arXiv 2014, arXiv:1409.0473. [Google Scholar]
  38. Vaswani, A.; Shazeer, N.; Parmar, N.; Uszkoreit, J.; Jones, L.; Gomez, A.N.; Kaiser, Ł.; Polosukhin, I. Attention is all you need. In Proceedings of the Conference and Workshop on Neural Information Processing Systems, Long Beach, CA, USA, 4–9 December 2017; pp. 5998–6008. [Google Scholar]
  39. Zeng, X.; Ouyang, W.; Yang, B.; Yan, J.; Wang, X. Gated bi-directional cnn for object detection. In European Conference on Computer Vision; Springer: Cham, Switzerland, 2016; pp. 354–369. [Google Scholar]
  40. Shrivastava, A.; Gupta, A. Contextual priming and feedback for faster r-cnn. In European Conference on Computer Vision; Springer: Cham, Switzerland, 2016; pp. 330–348. [Google Scholar]
  41. Everingham, M.; Van Gool, L.; Williams, C.K.; Winn, J.; Zisserman, A. The pascal visual object classes (voc) challenge. Int. J. Comput. Vis. 2010, 88, 303–338. [Google Scholar] [CrossRef] [Green Version]
  42. Lin, T.Y.; Maire, M.; Belongie, S.; Hays, J.; Perona, P.; Ramanan, D.; Dollár, P.; Zitnick, C.L. Microsoft coco: Common objects in context. In European Conference on Computer Vision; Springer: Cham, Switzerland, 2014; pp. 740–755. [Google Scholar]
  43. Liu, L.; Ouyang, W.; Wang, X.; Fieguth, P.; Chen, J.; Liu, X.; Pietikäinen, M. Deep learning for generic object detection: A survey. arXiv 2018, arXiv:1809.02165. [Google Scholar] [CrossRef] [Green Version]
  44. Torralba, A. Contextual priming for object detection. Int. J. Comput. Vis. 2003, 53, 169–191. [Google Scholar] [CrossRef]
  45. Oliva, A.; Torralba, A. The role of context in object recognition. Trends Cognit. Sci. 2007, 11, 520–527. [Google Scholar] [CrossRef]
  46. Chen, L.C.; Papandreou, G.; Kokkinos, I.; Murphy, K.; Yuille, A.L. Semantic image segmentation with deep convolutional nets and fully connected crfs. arXiv 2014, arXiv:1412.7062. [Google Scholar]
  47. Chen, L.C.; Papandreou, G.; Kokkinos, I.; Murphy, K.; Yuille, A.L. Deeplab: Semantic image segmentation with deep convolutional nets, atrous convolution, and fully connected crfs. IEEE Trans. Pattern Anal. Mach. Intell. 2017, 40, 834–848. [Google Scholar] [CrossRef]
  48. Divvala, S.K.; Hoiem, D.; Hays, J.H.; Efros, A.A.; Hebert, M. An empirical study of context in object detection. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Miami Beach, FL, USA, 20–25 June 2009; pp. 1271–1278. [Google Scholar]
  49. Galleguillos, C.; Belongie, S. Context based object categorization: A critical survey. Comput. Vis. Image Underst. 2010, 114, 712–722. [Google Scholar] [CrossRef] [Green Version]
  50. Zhang, X.; Yang, Y.H.; Han, Z.; Wang, H.; Gao, C. Object class detection: A survey. ACM Comput. Surv. 2013, 46, 10. [Google Scholar] [CrossRef]
  51. Ouyang, W.; Wang, X.; Zeng, X.; Qiu, S.; Luo, P.; Tian, Y.; Li, H.; Yang, S.; Wang, Z.; Loy, C.-C.; et al. Deepid-net: Deformable deep convolutional neural networks for object detection. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Boston, MA, USA, 7–12 Jun 2015; pp. 2403–2412. [Google Scholar]
  52. Hu, D. An introductory survey on attention mechanisms in NLP problems. In Proceedings of SAI Intelligent Systems Conference; Springer: Cham, Switzerland, 2019; pp. 432–448. [Google Scholar]
  53. Xu, K.; Ba, J.; Kiros, R.; Cho, K.; Courville, A.; Salakhudinov, R.; Zemel, R.; Bengio, Y. Show, attend and tell: Neural image caption generation with visual attention. In Proceedings of the International Conference on Machine Learning, Ithaca, NY, USA, June 2015; pp. 2048–2057. [Google Scholar]
  54. Cho, K.; Van Merriënboer, B.; Bahdanau, D.; Bengio, Y. On the properties of neural machine translation: Encoder-decoder approaches. arXiv 2014, arXiv:1409.1259. [Google Scholar]
  55. Hochreiter, S.; Schmidhuber, J. Long short-term memory. Neural Comput. 1997, 9, 1735–1780. [Google Scholar] [CrossRef] [PubMed]
  56. Abadi, M.; Agarwal, A.; Barham, P.; Brevdo, E.; Chen, Z.; Citro, C.; Corrado, G.S.; Davis, A.; Dean, J.; Devin, M.; et al. Tensorflow: Large-scale machine learning on heterogeneous distributed systems. arXiv 2016, arXiv:1603.04467. [Google Scholar]
  57. Simonyan, K.; Zisserman, A. Very deep convolutional networks for large-scale image recognition. arXiv 2014, arXiv:1409.1556. [Google Scholar]
  58. Deng, J.; Dong, W.; Socher, R.; Li, L.J.; Li, K.; Fei-Fei, L. Imagenet: A large-scale hierarchical image database. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Miami Beach, FL, USA, 20–25 June 2009; pp. 248–255. [Google Scholar]
Figure 1. The framework of our approach. I: Scene-level information processing module. II: Relationship module. III: Message passing module.
Figure 1. The framework of our approach. I: Scene-level information processing module. II: Relationship module. III: Message passing module.
Applsci 10 00418 g001
Figure 2. Object relation module.
Figure 2. Object relation module.
Applsci 10 00418 g002
Figure 3. Relation channel module.
Figure 3. Relation channel module.
Applsci 10 00418 g003
Figure 4. Some detection results of our model. (a) is a picture of road scene, (b) is a picture of dining room.
Figure 4. Some detection results of our model. (a) is a picture of road scene, (b) is a picture of dining room.
Applsci 10 00418 g004
Figure 5. Some sample images in PASCAL VOC database.
Figure 5. Some sample images in PASCAL VOC database.
Applsci 10 00418 g005
Figure 6. Some sample images in MS COCO database.
Figure 6. Some sample images in MS COCO database.
Applsci 10 00418 g006
Figure 7. The detection results with scene-level information or without scene-level information. (a): without scene-level information. (b): with scene-level information.
Figure 7. The detection results with scene-level information or without scene-level information. (a): without scene-level information. (b): with scene-level information.
Applsci 10 00418 g007
Figure 8. The detection results with relation information or without relation information. (a): without relation information. (b): with relation information.
Figure 8. The detection results with relation information or without relation information. (a): without relation information. (b): with relation information.
Applsci 10 00418 g008
Figure 9. Comparison of detection results.
Figure 9. Comparison of detection results.
Applsci 10 00418 g009aApplsci 10 00418 g009bApplsci 10 00418 g009c
Table 1. Experimental parameter setting.
Table 1. Experimental parameter setting.
Train DatabaseTest DatabaseBackboneIterations SettingLearning RateRun Time
(Sec/Img)
Test Time
(Sec/Img)
VOC 2007 trainval + VOC 2012 trainvalVOC 2007 testVGG16first 80k iterations0.00050.350.17
last 50k iterations0.00005
Resnet-101first 80k iterations0.0010.460.19
last 30k iterations0.0001
VOC 2007 trainval + VOC 2012 trainval + VOC 2007 testVOC 2012 testVGG16first 100k iterations0.00050.350.15
last 70k iterations0.00005
Resnet-101first 100k iterations0.0010.460.18
last 50k iterations0.0001
MS COCO 2014 train + MS COCO 2014 valMS COCO 2014 minivalVGG16first 350k iterations0.00050.320.16
last 200k iterations0.00005
Resnet-101first 350k iterations0.0010.460.19
last 150k iterations0.0001
Table 2. Detection results on PASCAL VOC 2007 test. Train set: 07 trainval + 12 trainval. “V” and “R” denote the model uses VGG-16 and Resnet-101 as backbone networks, respectively. The bold characters represent the best result for each column.
Table 2. Detection results on PASCAL VOC 2007 test. Train set: 07 trainval + 12 trainval. “V” and “R” denote the model uses VGG-16 and Resnet-101 as backbone networks, respectively. The bold characters represent the best result for each column.
MethodNetmAPAeroBikeBirdBoatBottleBusCarCatChairCowTableDogHorseMbikePersonPLANTSheepSofaTrainTv
Fast R-CNN [16]V70.077.078.169.359.438.381.678.686.742.878.868.984.782.076.669.931.870.174.880.470.4
Faster R-CNN [7]V73.276.579.070.965.552.183.184.786.452.081.965.784.884.677.576.738.873.673.983.072.6
SSD500 [24]V75.179.879.574.563.451.984.985.687.256.680.170.085.484.980.978.249.078.472.484.675.5
ION [32]V75.679.283.177.665.654.985.485.187.054.480.673.885.382.282.274.447.175.872.784.280.4
SIN [35]V76.077.580.175.067.162.283.286.988.657.784.570.586.685.677.778.346.677.674.782.377.1
Faster R-CNN [19]R76.479.880.776.268.355.985.185.389.856.787.869.488.388.980.978.441.778.679.885.372.0
SSD321 [24]R77.176.384.679.364.647.285.484.088.860.182.676.986.787.285.479.150.877.282.687.376.6
MIFNet(ours)V77.679.279.877.471.463.386.387.189.463.185.172.486.786.878.279.050.777.975.484.877.6
MIFNet(ours)R80.681.686.480.972.670.287.788.588.566.887.173.889.087.583.882.555.283.179.685.581.6
Table 3. Detection results on PASCAL VOC 2012 test. Train set: 07trainval + 12trainval + 07test. “V” and “R” denote the model uses VGG-16 and Resnet-101 as backbone networks, respectively. The bold characters represent the best result for each column.
Table 3. Detection results on PASCAL VOC 2012 test. Train set: 07trainval + 12trainval + 07test. “V” and “R” denote the model uses VGG-16 and Resnet-101 as backbone networks, respectively. The bold characters represent the best result for each column.
MethodNetMapAeroBikeBirdBoatBottleBusCarCatChairCowTableDogHorseMbikePersonPlantSheepSofaTrainTv
Fast R-CNN [16]V68.482.378.470.852.338.777.871.689.344.273.055.087.580.580.872.035.168.365.780.464.2
SSD300 [24]V70.384.276.369.653.240.878.573.688.050.573.561.785.880.681.277.544.373.266.781.165.8
Faster R-CNN [7]V70.484.979.874.353.949.877.575.988.545.677.155.386.981.780.979.640.172.660.981.261.5
CPF [40]V72.684.081.275.960.451.881.277.490.950.277.658.788.483.682.080.441.575.064.282.965.1
SIN [35]V73.184.879.574.559.755.779.578.889.951.976.858.287.882.981.881.651.275.263.981.867.8
SSD321 [24]R75.487.982.973.761.545.381.475.692.657.478.365.090.886.885.881.550.378.175.385.272.5
R-FCN [17]R77.686.983.481.563.862.481.681.193.158.083.860.892.786.084.684.459.080.868.686.172.9
MIFNet(ours)V74.486.281.976.460.358.080.278.490.453.878.558.388.383.083.782.753.076.366.582.670.5
MIFNet(ours)R78.487.784.382.866.560.486.082.293.759.076.562.291.387.684.885.958.381.372.186.479.2
Table 4. Detection results on MS COCO 2014 minival. Train set: trainval35k: MS COCO train + 35k val. “V” and “R” denote the model uses VGG-16 and Resnet-101 as backbone networks, respectively. The bold characters represent the best result for each column.
Table 4. Detection results on MS COCO 2014 minival. Train set: trainval35k: MS COCO train + 35k val. “V” and “R” denote the model uses VGG-16 and Resnet-101 as backbone networks, respectively. The bold characters represent the best result for each column.
MethodNetAPAP50AP70APSAPMAPLAR1AR10AR100ARSARMARL
Fast R-CNN [16]V20.539.919.44.120.035.821.329.530.17.332.152.0
Faster R-CNN [7]V21.140.919.96.722.532.321.530.430.89.933.449.4
YOLOv2 [27]V21.644.019.25.022.435.520.731.633.39.836.554.4
ION [32]V23.042.023.06.023.837.323.032.433.09.737.053.5
SIN [35]V23.244.522.07.324.536.322.631.632.010.534.751.3
SSD321 [24]R28.045.429.36.228.349.325.937.839.911.543.364.9
DSSD321 [24]R28.046.129.27.428.147.6------
R-FCN [17]R29.951.9-10.832.845.0------
SSD513 [24]R31.250.433.310.234.549.828.342.144.417.649.265.8
MIFNet(ours)V29.353.629.210.334.746.626.336.937.513.543.759.3
MIFNet(ours)R32.155.333.515.836.847.728.442.443.424.249.061.6
Table 5. Ablation study on PASCAL VOC 2007 test. All methods are trained on PASCAL VOC 2007trainval+2012trainval. Baseline: faster region-based convolutional network (faster R-CNN). Scene: only using scene information. The bold characters represent the best result for each column.
Table 5. Ablation study on PASCAL VOC 2007 test. All methods are trained on PASCAL VOC 2007trainval+2012trainval. Baseline: faster region-based convolutional network (faster R-CNN). Scene: only using scene information. The bold characters represent the best result for each column.
MethodMapAeroBikeBirdBoatBottleBusCarCatChairCowTableDogHorseMbikePersonPlantSheepSofaTrainTv
Baseline73.276.579.070.965.552.183.184.786.452.081.965.784.884.677.576.738.873.673.983.072.6
scene75.875.979.876.861.863.284.987.187.259.881.373.486.385.978.677.947.975.875.982.375.1
Table 6. Ablation study on MS COCO 2014 minival. All methods are trained on MS COCO train set. Baseline: Faster R-CNN trained. Scene: only using scene information. The bold characters represent the best result for each column.
Table 6. Ablation study on MS COCO 2014 minival. All methods are trained on MS COCO train set. Baseline: Faster R-CNN trained. Scene: only using scene information. The bold characters represent the best result for each column.
MethodAPAP50AP70APSAPMAPLAR1AR10AR100ARSARMARL
Baseline21.140.919.96.722.532.321.530.430.89.933.449.4
scene23.546.022.18.228.137.522.732.533.111.838.052.2
Table 7. Ablation study on PASCAL VOC 2007 test. All methods are trained on PASCAL VOC 2007trainval + 2012trainval. Baseline: Faster R-CNN. Relation: only using object-object relationships. The bold characters represent the best result for each column.
Table 7. Ablation study on PASCAL VOC 2007 test. All methods are trained on PASCAL VOC 2007trainval + 2012trainval. Baseline: Faster R-CNN. Relation: only using object-object relationships. The bold characters represent the best result for each column.
MethodMapAeroBikeBirdBoatBottleBusCarCatChairCowTableDogHorseMbikePersonPlantSheepSofaTrainTv
Baseline73.276.579.070.965.552.183.184.786.452.081.965.784.884.677.576.738.873.673.983.072.6
Relation76.477.980.075.167.262.586.086.488.661.084.772.986.887.377.478.746.676.372.982.276.7
Table 8. Ablation study on MS COCO 2014 minival. All methods are trained on MS COCO train set. Baseline: Faster R-CNN our trained. Relation: only using object-object relationships. The bold characters represent the best result for each column.
Table 8. Ablation study on MS COCO 2014 minival. All methods are trained on MS COCO train set. Baseline: Faster R-CNN our trained. Relation: only using object-object relationships. The bold characters represent the best result for each column.
MethodAPAP50AP70APSAPMAPLAR1AR10AR100ARSARMARL
Baseline21.140.919.96.722.532.321.530.430.89.933.449.4
Relation24.547.423.18.729.738.523.133.333.912.639.552.5
Table 9. Comparison of different relation module settings. K × Relation denotes with K connected relation modules. The bold characters represent the best result for each column.
Table 9. Comparison of different relation module settings. K × Relation denotes with K connected relation modules. The bold characters represent the best result for each column.
DatabaseGRU SettingsmAP
PASCAL VOC 1 × Relation77.6
2 × Relation75.2
3 × Relation74.7
MS COCO1 × Relation29.3
2 × Relation24.0
3 × Relation24.0
Table 10. Comparison of different GRU settings utilized in the attention-based global context sub-module. The experiments are conducted on PASCAL VOC 2007. (K × GRU) denotes that there are K stacked GRU units. The bold characters represent the best result for each column.
Table 10. Comparison of different GRU settings utilized in the attention-based global context sub-module. The experiments are conducted on PASCAL VOC 2007. (K × GRU) denotes that there are K stacked GRU units. The bold characters represent the best result for each column.
GRU SettingsmAPAeroBikeBirdBoatBottleBusCarCatChairCowTableDogHorseMbikePersonPlantSheepSofaTrainTv
1 × GRU77.679.279.877.471.463.386.387.189.463.185.172.486.786.878.279.050.777.975.484.877.6
2 × GRU76.977.479.876.868.563.785.687.287.160.485.070.886.386.279.379.348.376.476.785.377.2
3 × GRU76.976.580.177.267.962.886.786.587.263.184.071.987.086.177.878.747.876.877.684.176.1
Table 11. Comparison of different inputs of GRU. The experiments are conducted on PASCAL VOC 2007. All methods are trained on PASCAL VOC 2007 trainval + 2012 trainval. GRU1 represents the inputs are the scene-level information and relationship with objects. GRU2 represents the mixed information. The bold characters represent the best result for each column.
Table 11. Comparison of different inputs of GRU. The experiments are conducted on PASCAL VOC 2007. All methods are trained on PASCAL VOC 2007 trainval + 2012 trainval. GRU1 represents the inputs are the scene-level information and relationship with objects. GRU2 represents the mixed information. The bold characters represent the best result for each column.
SettingsmAPAeroBikeBirdBoatBottleBusCarCatChairCowTableDogHorseMbikePersonPlantSheepSofaTrainTv
GRU177.679.279.877.471.463.386.387.189.463.185.172.486.786.878.279.050.777.975.484.877.6
GRU276.276.480.774.965.962.183.787.387.861.682.467.787.284.778.778.449.977.677.281.677.1

Share and Cite

MDPI and ACS Style

Zhang, Y.; Kong, J.; Qi, M.; Liu, Y.; Wang, J.; Lu, Y. Object Detection Based on Multiple Information Fusion Net. Appl. Sci. 2020, 10, 418. https://doi.org/10.3390/app10010418

AMA Style

Zhang Y, Kong J, Qi M, Liu Y, Wang J, Lu Y. Object Detection Based on Multiple Information Fusion Net. Applied Sciences. 2020; 10(1):418. https://doi.org/10.3390/app10010418

Chicago/Turabian Style

Zhang, Yanni, Jun Kong, Miao Qi, Yunpeng Liu, Jianzhong Wang, and Yinghua Lu. 2020. "Object Detection Based on Multiple Information Fusion Net" Applied Sciences 10, no. 1: 418. https://doi.org/10.3390/app10010418

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop