Next Article in Journal / Special Issue
Pixel-Wise Crack Detection Using Deep Local Pattern Predictor for Robot Application
Previous Article in Journal
CNN-Based Multimodal Human Recognition in Surveillance Environments
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Statistic Experience Based Adaptive One-Shot Detector (EAO) for Camera Sensing System

1
Information Security Center, Beijing University of Posts and Telecommunications, Beijing 100876, China
2
Guizhou Provincial Key Laboratory of Public Big Data, Guizhou University, Guiyang 550025, China
*
Author to whom correspondence should be addressed.
Sensors 2018, 18(9), 3041; https://doi.org/10.3390/s18093041
Submission received: 22 July 2018 / Revised: 7 September 2018 / Accepted: 8 September 2018 / Published: 11 September 2018
(This article belongs to the Special Issue Semantic Representations for Behavior Analysis in Robotic System)

Abstract

:
Object detection in a camera sensing system has been addressed by researchers in the field of image processing. Highly-developed techniques provide researchers with great opportunities to recognize objects by applying different algorithms. This paper proposes an object recognition model, named Statistic Experience-based Adaptive One-shot Detector (EAO), based on convolutional neural network. The proposed model makes use of spectral clustering to make detection dataset, generates prior boxes for object bounding and assigns prior boxes based on multi-resolution. The model is constructed and trained for improving the detection precision and the processing speed. Experiments are conducted on classical images datasets while the results demonstrate the superiority of EAO in terms of effectiveness and efficiency. Working performance of the EAO is verified by comparing it to several state-of-the-art approaches, which makes it a promising method for the development of the camera sensing technique.

1. Introduction

Cameras are used to capture scalar phenomena in the form of images or videos, which provide more detailed and impressive data of the physical world [1]. In recent years, the camera sensing system has tended to be more automatic and intelligent. Compared with traditional cameras, a camera sensing system contains a communication interface, memory, an operating system and a processor. Aiming at facilitating the follow-up image inspection and judgement, the current development trend is to integrate the image sensor and detection algorithm into the same system in order for it to become modular as a camera sensor [2]. In most situations, imaging sensors are sensitive, accurate and well responding to objects. As a result, the detection of objects is important for the improvement of camera performance. This issue is most pronounced in the field of image processing and recognition. Due to the progress in machine learning methodologies, object detection models dramatically outperform manual detection, which exploit the best detection strategy by applying different algorithms. Neural network is one such principle with the potential to be beneficial for object detection optimizing.
The architecture of a convolutional neural network (CNN) was initially designed to take advantage of the 2D structure of an input signal [3,4]. Nevertheless, in line with the grand step in ImageNet Large Scale Visual Recognition Challenge (ILSVRC), the re-utilization of CNN is most pronounced due to its high precision in object detection and [5]. Thereafter, studies on CNN models were conducted and the recognition accuracy kept updating simultaneously [6,7,8,9]. Concretely, the mean average precision ( m A P ) and detection speed as the target detection parameters for open source dataset like PASCAL VOC and COCO keep improving continually. Researchers prefer to give the first priority to the advance in selective search (SS) [10] and region-based convolutional network (R-CNN) [11]. Fast R-CNN, which is a state-of-the-art detection algorithm, is capable of providing real-time processing via very deep convolutional networks regardless of the region proposal [12,13]. Further, Region Proposal Network (RPN) is proposed where a multimodal convolutional network is applied to jointly predict objectness and localization on a fusion of image and temporal motion cues [14]. Accordingly, for each Faster R-CNN, it is necessary to hold both a proposal network and a detection network, which is too sophisticated to optimize the model [15]. Specifically, You Only Look Once (YOLOv2) is a current well-employed method, with recent publications exploring the promise of using a hierarchical view of object classification and combining distinct datasets together. The basic idea of YOLO is to divide the image into smaller grid cells with each grid cell predicting bounding boxes and confidence for those boxes and class probabilities [16,17]. Unlike YOLO, the SSD architecture combines predictions from multiple feature maps at different resolutions in the network, which naturally handles objects of different sizes and improves detection quality [18].
The use of current-proposed detection model is, however, still limited, primarily because the single dataset, integrating difficulty, slow processing speed and low accuracy. For these reasons, the research is still ongoing to mitigate the deficiencies. In this research, we propose an object detection model EAO, statistic experience based adaptive one-shot detector, with the property of end-to-end detection. To enlarge the current detection dataset, a strategy for making a detection sample from a classification sample is proposed. Meanwhile, a detection dataset, namely ImageNet iLOC, for image processing models training and testing is constructed. The spectral clustering and ResNet methodology are integrated for image processing. The remaining part of this paper is organized as follows:
The background knowledge of spectral clustering, ResNet and stochastic gradient descent is depicted in Section 2. Section 3 describes the framework of EAO as well as the working principle in detail. Section 4 shows the results achieved in object detection experiments and the analysis of the model. The research findings and future planning for camera sensing system are presented in Section 5.

2. Preliminaries

2.1. Spectral Clustering

Clustering is one of the important techniques in the machine learning portfolio, particularly in unsupervised settings [19]. Spectral clustering is characterized with spectral graph theory—a field dedicated to graph studying via the eigenvalues and eigenvectors of matrices naturally associated with them [20]. In other words, spectral clustering splits data by means of graph partition using the weighted adjacency matrix and its spectrum [21,22]. Therefore, by exploiting the major components in the data revealed by the spectrum, spectral clustering results in a more powerful representation of data in feature space, so as to facilitate data clustering [23].
Assuming that we are given n samples, each datum x 1 , x 2 , , x n is in a sample space x i X . The symmetric weighted adjacency matrix W R n × n encodes the pairwise similarity where 0 w i j 1 . Aiming at obtaining a symmetric normalized weight matrix, the normalized graph Laplacian is computed as follows.
L = D 1 2 W D 1 2
Seeing that w i j W , we shall also define
L i j = w i j D i i D j j 1 2
where D is a diagonal matrix and D i i = j = 1 n w i j . Suppose that there are n orthonormal eigenvectors corresponding to the eigenvalues λ 1 , λ 2 , , λ n ( λ 1 λ 2 λ n ) of L.
Considering the eigenvectors selection process, a new matrix U R n × k is formed by stacking the first largest k eigenvectors from L. Thus, we have:
L ( k ) = U U T
Mathematically, the eigenvectors inherit information from the adjacency matrix as the new expression enhances the ample properties in the feature space. The samples lay in feature space are either clustered closer or repulsed further apart, which lead to a more prominent cluster in the feature space [24,25]. Thus, the data dimension is significantly reduced.

2.2. ResNet

To the best of our knowledge, features from deeper networks tend to obtain a better working performance than that extracted from shallow networks [26]. Whereas, the gradient fading with the network depth increasing comes to be the greatest technical challenge. In 2015, He et al. proposed the deep residual framework named ResNet [27]. Currently, the depth of the ResNet has already reached an astonishing level of 1001 layers and apparently, the depth can also be increased [28].
Since it is impossible to fit the output mapping with the stacked layer, He et al. focus on fitting the residual mapping and pushing the residual to zero. In [29], the idea of “shortcut connections” is first put forward for making the inputs identity mapping. Meanwhile, to improve the processing speed, He et al. carefully tune the layer sizes to balance the computation between various sub-networks [30].

2.3. Stochastic Gradient Descent

Stochastic gradient-based optimization is one of the important issues in mathematical resolution of scientific studies. Stochastic gradient descent (SGD), which is the core part of all the stochastic methods, is capable of processing large-scale problems in a way that clearly outperforms classical optimization methods [31,32]. Specifically, SGD proves its efficiency and reliability in recent advances in neural networks [33,34]. Generally, a convolutional network contains alternating layers of convolution and pooling (i.e., subsampling). Stochastic-pooling, which is one of the wide-spread pooling methods, shows the capability of the deep learning process to prevent the issue of over-fitting [35].
ADAM, proposed by D. P. Kingma and J. L. Ba is an algorithm for automatically tuning the learning rate in SGD [36]. Commonly, the loss function is defined for a single training example for SGD and is shown in Equation (4):
L S G D ( x p ) = 1 2 ( y p y ( x p ) ) 2 , p = 1 , 2 , , P
The weight updating of the model is expressed in Equations (5)–(8).
w n e w = w o l d α · y i · ( y p y ( x p ) )
c j i n e w = c j i o l d β · h j i j = 1 m i h j i · ( y p y ( x p ) )
a j i n e w = a j i o l d γ · h j i j = 1 m i h j i · ( c j i y i ) · ( y p y ( x p ) ) · 2 ( x p i a j i o l d ) b j i
b j i n e w = b j i o l d η · w i · ( c j i y i ) · h j i j = 1 m i h j i · ( y p y ( x p ) ) · 2 ( x p i a j i o l d ) b j i
where α , β , γ and η indicate learning rates of SGD, which influence the stability of the learning process. Nevertheless, based on the deployment of ADAM, the parameter update is simplified as:
θ n e w = θ o l d κ · m ^ v ^ + ε
where κ is the learning rate, θ is the model parameter replacing the a, b, c and w and ∇ θ is the gradient from the loss function. ε is a hyper-parameter specifically for ADAM. Considering the computation efficiency, m and v are initialized as the 1st moment vector and the 2nd moment vector which are applied to be the exponential moving average of θ and ( θ ) 2 separately. The m ^ and v ^ designate the parameters after deviation correction. Computation of m and v are facilitated by using the following formula:
m ^ = m 1 β 1 , v ^ = v 1 β 2
where β 1 and β 2 are exponential decay rates for the moment estimating, which are set during the model initialization.

3. EAO Object Detection Methodology

In this section, we describe EAO for image capture and processing, which can effectively address the difficulties in object detection tasks. In the proposed algorithm, all the samples are taken from the open access image dataset ImageNet, PASCAL VOC and COCO.

3.1. From Classification Dataset to Detection Dataset

Current object detection samples only occupy a small part of the datasets for other purposes (e.g., classification and tagging) [37,38]. Previous works have been focused primarily on obtaining an even higher accuracy of classification samples [6,39]. In such work, images are presented for machine learning model training and further testing [7,8]. Alternatively, object detection datasets are used as a secondary source due to their finite applications. For this reason, the object classification dataset holds great promise for effective learning of given targets after the labeling and recognizing process. Consequently, a strategy for converting a classification sample into a detection sample is devised. Consequently, we made the detection dataset, namely ImageNet iLOC, for network training and put forward the strategy for converting classification samples into detection samples.
Both ResNet and Spectral cluster provide good performances in the field of image classification and image segmentation. On the other hand, cropping [33,40] and warping [11,41] will inevitably bring deviation as well. According to the basic theory of deep learning and the experimental outcomes, the improvement in working performance and normalization definitely outweighs the working error. Spectral clustering is a kind of unsupervised learning method. In the case of image processing, spectral clustering is used to outline the object and further determine its relative position. In addition, the ResNet approach is employed for object classifying and matching.
Generally, an object set is defined as X i = { c x , c y , w , h , c a t e g o r y } where c x and c y indicate the coordinate of the ground truth box center, the width w and height h characterize the size of ground truth box and c a t e g o r y designate the object label. The ground truth box of the detection object is determined by spectral clustering. Specifically, we shall define the coordinate of left border L and bottom border D as the feature vectors for each sample. To start with, an input image sample is segmented into m × n pixels while the feature vectors can be expressed as L = { l 1 , l 2 , , l m } and D = { d 1 , d 2 , , d n } . The spectral clustering algorithm is deployed on the aforementioned pixels to acquire each cluster, which represents one detection object O j and O j O = { O 1 , O 2 , , O M } where O stand for the object set in the image. In line with each cluster, the coordinate vector of each pixel point of the object is defined as C j = { c 1 j , c 2 j , , c K j } . Therefore, each pixel point, with the maximum and minimum distance to the image boundary { L , D } can be calculated. Let t u j , t d j , t l j and t r j be the upper, bottom, left and right vertex of object contour respectively, we shall thus define
T j = { t u j , t d j , t l j , t r j }
Moreover, the coordinates of the anchor box vertexes can be represented by ( x 1 , y 1 ) , ( x 2 , Δ y ) , ( Δ x , y 3 ) and ( x 4 , y 4 ) which are furthest to the image boundary on each side.
t u j = κ K max d i s t a n c e ( D , c k j ) t d j = κ K min d i s t a n c e ( D , c k j ) t l j = κ K max d i s t a n c e ( L , c k j ) t r j = κ K min d i s t a n c e ( L , c k j )
In Equation (12), the function d i s t a n c e ( ) stands for calculating the distance of the vertex. Accordingly, the basis parameters w, h and ( c x , c y ) are obtained, i.e.,
w = x 4 Δ x h = y 1 Δ y ( c x , x y ) = Δ x + w , Δ y + h
Hereafter, the image label is determined by the value of confidence via ResNet. In addition, following the results from Faster R-CNN, SSD, YOLO, etc., we conduct a considerable number of experiments on object detection. The confidence value is adjusted based on experimental outcomes. We find that 85% for the confidence threshold results in a best working capability. If and only if c o n f i d e n c e 0.85 the target ground truth outputs the label of { ( c x j , c y j ) , w , h , c } . Figure 1 presents an example of two target objects recognition for a particular image. The object recognition process with ImageNet iLOC is described as Algorithm 1.
Algorithm 1 Generating object detection dataset based on spectral clustering
Input: Classification sample X and the size of the dataset is n.
Output: Detection object D
1:
functionClassificationToDetection(X)
2:
    Initialization k c l u s t e r = 6
3:
    for i = 1 to n do
4:
        Compute the left and bottom coordinates of sample X i
5:
        Update the coordinate set { L , D }
6:
        Spectral clustering for detection object O based on sample X i
7:
        for j = 1 to O M do
8:
             Compute the upper vertex t u j , bottom vertex t d j , left vertex t l j and right vertex t r j of object  O j
9:
             Update the object coordinate T j = { t u j , t d j , t l j , t r j } base on Equation (12)
10:
             Compute the width w and height h of the anchor box base on Equation (13)
11:
             Compute the center coordinate ( c x j , c y j ) of the anchor box base on Equation (13)
12:
             Update the positioning information of O j as b j = { ( c x j , c y j ) , w , h }
13:
             Compute the parameter c o n f i d e n c e of object O j as c = R e s N e t ( b j )
14:
             if max c ( c i ) > 0 . 85 then
15:
                 Compute the ground-truth box of object O j as g t b j = { ( c x j , c y j ) , w , h , c }
16:
                 Update the detection dataset with g t b j
17:
           else
18:
                 continue
19:
           end if
20:
        end for
21:
    end for
22:
    return the new detection dataset D
23:
end function

3.2. Prior Box Generating

As a rule, objects are sought to be delineated before their identification [10]. In order to get a unique partitioning of the object, most bounding boxes are defined manually based on experience. In most cases, the bounding boxes lack statistical analysis, which restrain the intersection over union (IOU) overlapping [11,13]. Previous work has paid less attention to both the anchor box (Faster) and the default boxes (SSD). Box shape and number are taken from subjective estimation. For this reason, the ground-true box cannot be approached and the object cannot be detected. In this research, K-means++ is employed for box shape clustering. The objects can be better outlined. Further, the efficiency of bounding boxes is improved by revising the parameter k. Thereupon, we focus on the prior data from the novel datasets to accurately construct the bounding box. Providing the exploratory nature of the figure composition, the images contain different configurations regarding the color, shape, as well as texture of the objects. The goal of this stage is to take K-means++ clustering to detect the shape and size of the object prior box, and thus to improve the convergence speed and detection accuracy. Unlike K-means clustering employed in YOLO and DSSD, K-means++ gets an increasing number of cluster centers, which reduces the uncertainty of random selection and improves the clustering speed and accuracy [18,42,43].
Considering the model complexity and high recall, K = 6 is adopted as a good tradeoff for further computing (Figure 2). Normally, the Euclidean distance is used to denote the distance between an element and the cluster center in the clustering algorithm. Note that the objects are of different sizes; the Euclidean distance cannot exactly reflect the object location in the image. We shall thus use a more appropriate form to demonstrate the distance, which is shown in Equation (14):
d ( b o x , c e n t r o i d ) = 1 I O U ( b o x , c e n t r o i d )
where b o x represents the samples and c e n t r o i d the cluster center. Function I O U outputs the overlapping ratio of this object to the cluster. Figure 3 shows the clustering result based on dataset ImageNet and COCO while most bounding boxes are of slender shapes. This involves boxing the object with the clustering outcome, which compares the same image via different bounding approaches. The bounding boxes from K-means++ clustering outcomes show better working performance, evidenced by a higher IOU overlapping proportion.

3.3. Multi-Resolution Feature Mapping

Object capturing on real images is both laborious and time-consuming. The single-shot detector (SSD), which significantly outperforms other methods, combines the standard architecture and the auxiliary structure for high quality image classification. For each feature layer of size m × n , SSD assigns default boxes of different scales to every single cell. However, the bounding boxes cannot always accurately detect the object. The box in the red solid line is the specific ground truth. As depicted by an example in Figure 4a, for objects of different sizes, default boxes of high-resolution from SSD fail to cover the targets, which in turn increase the computing amount. Meanwhile, in the layers of low resolution, objects are missed because the boxes are too large. To best fit a specific layer, the resolution largely affects the number of boxes: for a high-resolution cell the default boxes are redundant while for a low-resolution one, more boxes need to be stacked and revise the prior boxes into slender shapes handled carefully. Note that the prior boxes are assigned due to the clustering results; the total number of boxes decrease with the overlapping ratio increases in Figure 4b.
In this research, we employ the feed-forward convolutional neural network to generate a fixed-size set of bounding boxes and a non-maximum suppression approach to determine the final detections, which optimizes the current SSD. According to Section 3.2, computation with K-means++ clustering for object detection is facilitated by taking K = 6 . For every single image, the number of target objects is so limited that the cost can be significantly reduced by assigning bounding boxes more precisely. Within one layer, each cell maps to a specific region of the original image, the size of which varies based on different resolution of the convolutional layer. Considering the image segmentation principle, bounding boxes can be defined due to layer resolution and the cell property. For this reason, we come up with the strategy that more bounding boxes should be assigned to the convolutional layer of lower resolution, and vice versa.
The model architecture is shown in Figure 5. EAO is initiated with standard VGG 16 as base network, which is a classical network for feature mapping [44]. Instead of using max-pooling, we employ the stochastic-pooling algorithm to address the issue of overfitting [45]. The multi-resolution detection layers are then integrated to seed the detection algorithm. The feature mapping is on the foundation of activation function Rectified Linear Unit (ReLU) to prevent gradient disappearance and results in faster learning [46]. By using the classical VGG 16 version, the basic layer conv4_3 is kept and the two fully connected layers, FC6 and FC7 are converted into typical 19 × 19 convolution layers [47]. The latter is utilized as the detection layer. As a result, the network is deployed by multiple feature maps, whose sizes are 38 × 38 × 512 , 19 × 19 × 1024 , 10 × 10 × 512 , 5 × 5 × 256 , 3 × 3 and 1 × 1 × 256 . The convolutional layers decrease in size progressively. In each detection layer, a 3 × 3 convolution is applied to extract the feature of prior boxes. At each feature map cell, the offsets relative to the ground-truth box in the cell, as well as the conditional probabilities of the object category are picked. For a given cell with k prior boxes, we focus here on the four offsets involved with the ground-truth box. These layers are followed by a softmax classifier and a linear regressor, which predicts the category of the prior box and calculates the offset between the prior box and ground-truth box, respectively. Final detections are produced through non-maximum suppression step. Within the model, the random optimizer ADAM is taken for end-to-end training optimization.
Following the K-means++ clustering outcome, the shapes of bounding boxes for each layer are dedicatedly devised. The average IOU of a different detection model is presented (Table 1). With different bounding boxes assigned, a better average IOU of EAO is observed.

3.4. Training

The training process is originated from minimizing the loss of multitask objective function [48]. Supposing x is the convolution result for matching the predicted box p and the ground-truth box g, we introduce the classification loss ( c l s ) and the regression loss ( r e g ). Considering the multiple object categories, the overall objective loss function is a weighted sum of the classification loss and the regression loss:
L ( x , c , p , g ) = 1 N ( L c l s ( x , c ) + α L r e g ( x , p , g ) )
where N is the number of matched boxes, c is the multiple-class confidence and α is the weight term for controlling detection error. Typically, the regression loss is determined by s m o o t h L 1 which characterizes the localization accuracy between the predicted box p and the ground truth [12].
One of the key processes gleaned from the Faster R-CNN is that the offset of the width (w), the height (h) and the central point ( c x , c y ) can be obtained as the solution of regression. This idea is developed in our strategy for model evolution. We thus define the width, height and center of the predicted box, prior box and the ground-truth box. The offset vector between the prior box and the predicted box is t ^ i m while that of the predicted box is p and the ground-truth box is g. Thereby, we have
L r e g ( x , p , g ) = i P o s N m { c x , c y , w , h } x i j k s m o o t h L 1 ( t ^ i m t ^ j m )
in line with
t ^ i c x = p i c x d i c x d i w , t ^ i c y = p i c y d i c y d i h
t ^ i w = log p i w d i w , t ^ i h = log p i h d i h
t j c x = g j c x d i c x d i w , t j c y = g j c y d i c y d i h
t i w = log g j w d i w , t j h = log g j h d i h
Note that the confidence loss is the softmax loss over multiple-class confidence c, the classification loss is defined as
L c l s ( x , c ) = i P o s N x i j p log ( c ^ i p ) i N e g log ( c ^ i 0 )
together with
c ^ i p = exp ( c i p ) p exp ( c i p )
After the prior boxes generated, each of the prior boxes with IOU overlap higher than 0.7 are selected as a positive proposal. On the other hand, negative training samples imply the overlap is lower than 0.3. The rest of the prior boxes are removed from inputs of the training model. Since most of the prior boxes are negative in practical use, the min-batch method is used to keep the ratio of matched to unmatched boxes to 1:1. For an image lacking positive samples, we employ the negative ones for supplementing. In this case, the ADAM algorithm is employed to improve the model accuracy based on offset revising.

4. Experiments

Experiments are conducted on three challenging datasets: PASCAL VOC 2007, PASCAL VOC 2012 and COCO to evaluate the working performance of EAO. All the state-of-the-art algorithms, i.e., SSD, YOLO and Faster R-CNN, are trained with image detection datasets. The dataset, namely Imagenet, is taken for model training. In this manuscript, we introduce the self-made dataset ImageNet iLOC for EAO training after training with traditional detection samples. The detection accuracy outperforms other methods. The performance evolution is presented in Section 4.4.2. All the testing datasets in this research are the same.

4.1. Dataset

PASCAL VOC2007: The PASCAL VOC project provides standardized labelled images for object recognition. Meanwhile, the evaluation of recognition method on these datasets can be achieved through its evaluation server. There are images of 20 classes in PASCAL VOC2007. In order to detect objects from a number of visual object classes in realistic scenes, we take 16,551 images from the VOC2007 training collection and validation collection for training and 4952 pieces of test data for testing.
PASCAL VOC2012: Compared to PASCAL VOC2007, the size of image dataset increased substantially. In PASCAL VOC2012, each training image is associated to an annotation file providing object class label for each object. In this stage, all the 10,991 images from the dataset are used to evaluate the capability of EAO.
MS COCO: MS COCO consists of over 10,000 image samples of 91 object classes which aims at gathering images of complex daily scenes containing common objects in their natural context. However, objects in COCO tend to be smaller than those in PASCAL VOC, detection boxes assigning can therefore be adjusted for different layers.

4.2. Evaluation Protocol

Mean average precision( m A P ): For a given task, a classical evaluation protocol is to compute the p r e c i s i o n / r e c a l l curve. Recall stands for the proportion of all positive examples ranked above a given rank while precision for the proportion of all examples above that rank from the positive class [49]. Formally, the m A P that indicates the shape of p r e c i s i o n / r e c a l l outcome is proposed to evaluate the detection performance of PASCAL VOC. The m A P is specified by 11-points interpolated average precision, which is the average of the maximum precision for recall levels at a fixed set of uniformly-spaced recall values [ 0 , 0.1 , 0.2 , , 1.0 ] [50]. The precision at each recall level r is interpolated by taking the maximum precision. The m A P is expressed as
m A P = 1 11 r { 0 , 0 . 1 , , 1 } P i n t e r p ( r )
together with
P i n t e r p ( r ) = max r ~ : r ~ r p ( r ~ )
where is p( r ~ ) the measured precision at recall r ~ .
IOU: IOU is taken as a standardized metric for characterizing the performance of an object detector on MS COCO. Considering that more small objects appear in images of MS COCO, we evaluate the model detection precision for IOU [ 0.5 : 0.05 : 0.9 ] [11].

4.3. Implementation Details

We evaluate EAO in comparison to SSD, YOLO and Faster R-CNN. To start with, we fine-tune the model with ADAM optimizer and set β 1 = 0.9 , β 2 = 0.999 , ε = 1 × 10 8 . The initial learning rate is 0.001. The exponential decay rates vary slightly for different datasets after the first-step iteration. Working parameters for distinguished datasets are given in Table 2.
For each dataset, the EAO with two different types of input, 300 × 300 and 512 × 512 , is taken for training and testing. To further improve the recognition precision, we make and release the object classification dataset ImageNet iLOC with the proposed method in Section 3.1 and pre-train EAO with it. Subsequently, the model is applied to PASCAL VOC2012. The iterations for the pre-training process is fixed at 150k. However, this number differs in the second training step which are conducted on each dataset’s own training images. The differences depend on the size of dataset.

4.4. Results

4.4.1. PASCAL VOC2007

The network is first evaluated on PASCAL VOC2007 dataset, which is a basic task for object detection. We use the initialization parameters for 40k iterations, then continue training for 20k iterations with revised parameters in Table 2. Table 3 summarises all the results comparing with some state-of-the-art approaches, in terms of m A P on all categories.
According to the testing outcome, our model trained with input 300 × 300 , which keeps the same aspect ratio with SSD300, is of a higher accuracy ( 74.9 % vs. 74.3 % for SSD300). Our results denote EAO outperforms SSD when improving the original input size to 512 × 512 . A 78.1 % m A P is gained and the recall reaches 85–90%. In contrast with R-CNN, EAO is particularly effective due to the classifier and the regressor. To further assess the property of EAO from different resolution, we record the working process by employing analysis tool Tensorboard provided by Google tensorflow. With the multi-resolution network applied, a higher IOU can be obtained. Compared to SSD, EAO is highlighted with the convergence speed and the detection accuracy, which is expressed in Figure 6.
Most object recognition methods are limited in object detection since the object information is prone to be filtered through multi-layer convolution. On the one hand, Table 3 shows the detection precision for a small object is improved from a higher resolution network. On the other hand, prior boxes of various size are applied to different layers based on the pre-processing step in Section 3.1. Consequently, EAO provides higher robustness and detection precision than SSD (Figure 7). From prior boxes proposed to a high precision model, the oscillation processes stop after 60K and 80K times iteration for EAO512 and SSD512 separately. Crucially, EAO is amenable to efficiently stabilize the detection outcomes.

4.4.2. PASCAL VOC2012

Aiming at obtaining a comprehensive evaluation of the proposed model, we assign to each model the training set provided by PASCAL VOC2012. Distinctively, we enlarge the training set of EAO, which contains both ImageNet iLOC data sets (produced in Section 3.1) and PASCAL VOC2012 trainval sets. Then all models are evaluated on PASCAL VOC2012 test set. The outcome similar to that from PASCAL VOC2007 is acquired (Table 4). According to the test results, the detection precision of EAO512 attains a 75.8 percent, which outperforms that of Faster R-CNN, YOLO and SSD via training from basic dataset. Moreover, by adopting the training sets from ImageNet, iLOC yields a 3.4 percent and a 3.6 percent improvement for EAO300 and EAO512, respectively. We also note that the performance of EAO300, with the new expansion data augmentation, is better than that of original EAO512 trained by PASCAL VOC2012. This happens because, given more exact detection objects, the model will be trained to work in a more accurate and faster mode. Incorporating the spectral clustering on objects and the bounding box prediction, our proposed object detection dataset is shown to greatly boost the detection accuracy. Likewise, it is noteworthy that the working performance on objects that are difficult to recognize is improved, like the boat and the bird.

4.4.3. MS COCO

For the purpose of identifying our network on a more general, large-scale image dataset, we apply the models to MS COCO. Because objects in MS COCO tend to be smaller than those from PASCAL VOC, we train EAO with the initialization parameter for the first 140k iterations, followed by 40k iterations with the specifications in Table 2. Considering the bounding box generating strategy in Section 3.2, we use smaller prior boxes for all layers. The numbers and the sizes of the prior boxes for various layers are then determined based on the feature map principle.
We take the whole trainval set for model training and the whole test set for testing. The results can be visualized in Table 5. The working performance of EAO300 is close to that of SSD512 while EAO512 exceeds the state-of-the-art methods on MS COCO dataset in all criteria. Specifically, both EAO300 and EAO512 significantly outperform other methods on m A P @ 0.5 . The reason for this is most objects in MS COCO are relatively small while SSD and Faster R-CNN tend to fails to capture the delicate boundaries of small objects. The recognition results and the convergence process are presented in Figure 8 and Figure 9, respectively.

4.4.4. Inference Time

As long as a large number of prior boxes are generated from our method, it is advisable to perform non-maximum suppression during inference. We filter most boxes with a confidence threshold of 0.05. The top 150 detections of each image are maintained by integrated non-maximum suppression and jaccard overlap of 0.45 per class. Note that 20 object classes belong to the PASCAL VOC dataset; the best result on each image is 1.3 ms with a 300 × 300 input while the total runtime on all detection layers is 1.9 ms. A quantitative comparison with other competing approaches is depicted in Table 6.
As long as a large number of prior boxes are generated from our method, it is advisable to perform non-maximum suppression during inference. We filter most boxes with a confidence threshold of 0.05. The top 150 detections of each image are maintained by integrated non-maximum suppression and jaccard overlap of 0.45 per class. Note that 20 object classes belong to PASCAL VOC dataset; the best result on each image is 1.3 ms. On 300 × 300 images, it takes all the six detection layers 1.9 ms on non-maximum suppression when running on a Xeon-E7-8800 v3 CPU with 2 GTX1080 Ti GPU and a Dell R740 server. A qualitative comparison with other competing approaches is depicted in Table 6. For the same image dataset, the Fast YOLO shows a decent inference time of 155 FPS but a poor detection precision. With the additional feature mapping layer applied to the region proposal network, the Fast R-CNN is slowed down by feature downsampling. Besides, outcomes of EAO are slightly better than SSD on both running time and detection precision. In particular, EAO300 achieves a m A P of 76.6 while maintaining a real-time speed of 61 FPS.

5. Discussion and Conclusions

This paper introduces EAO as an end-to-end convolutional network for object recognition, which is efficient and reliable to detect objects with a high precision for camera sensing. Aiming at facilitating the image processing, we make the detection dataset, namely ImageNet iLOC, via the spectral clustering method. This paper also proposes a detailed study of the prior box generating principle. Comprehensively, a multi-resolution network model for object detection is constructed and trained. Experiments are carried out on classical image datasets. By using the pretraining dataset ImageNet iLOC, an improved working performance is obtained. All experimental results are carefully analyzed. Compared to some state-of-the-art methods, the results validate the effectiveness of EAO and demonstrate the high efficiency in both the runtime and recognition accuracy.
Future work will address more complex situations where objects are presented in the camera-video form. The current model can be extended to a recurrent neural network by integrating it with other algorithms.

Author Contributions

The work presented here was carried out in collaboration among all authors. All authors have contributed to, seen and approved the manuscript. Conceptualization, X.Z. and B.D.; Data curation, B.D.; Funding acquisition, L.G.; Investigation, Q.M.; Methodology, X.Z. and Y.Y.; Project administration, Q.M., L.G. and Y.Y.; Resources, Q.M.; Software, B.D.; Supervision, L.G.; Validation, Y.Y.; Visualization, B.D.; Writing—original draft, X.Z.

Funding

This research was funded by The National Science and Technology Major Project (Grant no. 2017YFB0803001), the National Natural Science Foundation of China (Grant no. 61502048), Beijing Science and Technology Planning Project (Grant no. Z161100000216145), the National “242” Information Security Program (2015A136) and The APC was funded by Lize Gu.

Acknowledgments

The authors would like to thank the reviewers for their helpful advice.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Wang, Y. Coverage Problems in Camera Sensor Networks. Ph.D. Thesis, The Pennsylvania State University, London, UK, 2013. [Google Scholar]
  2. Lee, S.H.; Yang, C.S. A Real Time Object Recognition and Counting System for Smart Industrial Camera Sensor. IEEE Sens. J. 2017, 17, 2516–2523. [Google Scholar] [CrossRef]
  3. Chua, L.O.; Roska, T.; Venetianer, P.L. The CNN is universal as the Turing machine. IEEE Trans. Circuits Syst. I 1993, 40, 289–291. [Google Scholar] [CrossRef]
  4. Hanggi, M.; Moschytz, G.S. An exact and direct analytical method for the design of optimally robust CNN templates. IEEE Trans. Circuits Syst. I 1999, 46, 304–311. [Google Scholar] [CrossRef] [Green Version]
  5. Russakovsky, O.; Deng, J.; Su, H.; Krause, J.; Satheesh, S.; Ma, S.; Huang, Z.; Karpathy, A.; Khosla, A.; Bernstein, M.; et al. ImageNet Large Scale Visual Recognition Challenge. Int. J. Comput. Vis. 2015, 115, 211–252. [Google Scholar] [CrossRef] [Green Version]
  6. He, K.; Zhang, X.; Ren, S.; Sun, J. Spatial Pyramid Pooling in Deep Convolutional Networks for Visual Recognition. IEEE Trans. Pattern Anal. Mach. Intell. 2015, 37, 1904–1916. [Google Scholar] [CrossRef] [PubMed]
  7. Szegedy, C.; Liu, W.; Jia, Y.; Sermanet, P.; Reed, S.; Anguelov, D.; Erhan, D.; Vanhoucke, V.; Rabinovich, A. Going deeper with convolutions. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Boston, MA, USA, 12–17 June 2015; pp. 1–9. [Google Scholar]
  8. He, K.; Zhang, X.; Ren, S.; Sun, J. Deep Residual Learning for Image Recognition. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Las Vegas, NV, USA, 27–30 June 2016; pp. 770–778. [Google Scholar]
  9. Zhang, B.; Gu, J.; Chen, C.; Han, J.; Su, X.; Cao, X.; Liu, J. One-two-one network for compression artifacts reduction in remote sensing. ISPRS J. Photogramm. Remote Sens. 2018. [Google Scholar] [CrossRef]
  10. Uijlings, J.R.R.; Sande, K.E.A.; Gevers, T.; Smeulders, A.W.M. Selective Search for Object Recognition. Int. J. Comput. Vis. 2013, 104, 154–171. [Google Scholar] [CrossRef] [Green Version]
  11. Girshick, R.; Donahue, J.; Darrell, T.; Malik, J. Rich feature hierarchies for accurate object detection and semantic segmentation. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Columbus, OH, USA, 23–28 June 2014; pp. 580–587. [Google Scholar]
  12. Girshick, R. Fast R-CNN. arXiv, 2015; arXiv:1504.08083. [Google Scholar]
  13. Ren, S.; He, K.; Girshick, R.; Sun, J. Faster R-CNN: Towards Real-Time Object Detection with Region Proposal Networks. IEEE Trans. Pattern Anal. Mach. Intell. 2017, 39, 1137. [Google Scholar] [CrossRef] [PubMed]
  14. Sarikaya, D.; Corso, J.J.; Guru, K.A. Detection and Localization of Robotic Tools in Robot-Assisted Surgery Videos Using Deep Neural Networks for Region Proposal and Detection. IEEE Trans. Med. Imaging 2017, 36, 1542–1549. [Google Scholar] [CrossRef] [PubMed]
  15. Zhang, B.; Li, Z.; Cao, X.; Ye, Q.; Chen, C.; Shen, L.; Perina, A.; Ji, R. Output constraint transfer for kernelized correlation filter in tracking. IEEE Trans. Syst. Man Cybern. 2017, 47, 693–703. [Google Scholar] [CrossRef]
  16. Redmon, J.; Divvala, S.; Girshick, R.; Farhadi, A. You Only Look Once: Unified, Real-Time Object Detection. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Las Vegas, NV, USA, 27–30 June 2016; pp. 779–788. [Google Scholar]
  17. He, K.; Gkioxari, G.; Dollár, P.; Girshick, R. Mask R-CNN. In Proceedings of the 2017 IEEE International Conference on Computer Vision (ICCV), Venice, Italy, 22–29 October 2017. [Google Scholar]
  18. Liu, W.; Anguelov, D.; Erhan, D.; Szegedy, C.; Reed, S.; Fu, C.Y.; Berg, A.C. SSD: Single Shot MultiBox Detector. In Proceedings of the European Conference on Computer Vision, Amsterdam, The Netherlands, 8–10 October 2016; pp. 21–37. [Google Scholar]
  19. Jebara, T.; Shchogolev, V. B-Matching for Spectral Clustering. In Proceedings of the Machine Learning: ECML 2006, Berlin, Germany, 18–22 September 2006; pp. 679–686. [Google Scholar]
  20. Spielman, D. Spectral Graph Theory, Combinatorial Scientific Computing; Chapman & Hall: London, UK, 2012. [Google Scholar]
  21. Von Luxburg, U. A tutorial on spectral clustering. Stat. Comput. 2007, 17, 395–416. [Google Scholar] [CrossRef] [Green Version]
  22. Vidal, R. Subspace clustering. IEEE Signal Process. Mag. 2011, 28, 52–68. [Google Scholar] [CrossRef]
  23. Zhang, C.; He, Y.; Fraser, C.S. Spectral Clustering of Straight-Line Segments for Roof Plane Extraction from Airborne LiDAR Point Clouds. IEEE Geosci. Remote Sens. Lett. 2018, 15, 267–271. [Google Scholar] [CrossRef]
  24. Ng, A.Y.; Jordan, M.I.; Weiss, Y. On spectral clustering: Analysis and an algorithm. In Proceedings of the Advances in neural information processing systems, Long Beach, CA, USA, 3–8 December 2001; pp. 849–859. [Google Scholar]
  25. Liu, R.; Zhang, H. Segmentation of 3D meshes through spectral clustering. In Proceedings of the 12th Pacific Conferece on omputer Graphics and Applications, Seoul, Korea, 6–8 October 2004; pp. 298–305. [Google Scholar] [Green Version]
  26. Belkin, M.; Niyogi, P. Laplacian eigenmaps for dimensionality reduction and data representation. Neural Comput. 2003, 15, 1373–1396. [Google Scholar] [CrossRef]
  27. Zhang, R.; Wang, Q.; Lu, Y. Combination of ResNet and Center Loss Based Metric Learning for Handwritten Chinese Character Recognition. In Proceedings of the 14th IAPR International Conference on Document Analysis and Recognition, Kyoto, Japan, 10–15 November 2017; pp. 25–29. [Google Scholar]
  28. Li, D.; Wang, Z. Video Superresolution via Motion Compensation and Deep Residual Learning. IEEE Trans. Comput. Imaging 2017, 3, 749–762. [Google Scholar] [CrossRef]
  29. Li, B.; He, Y. An Improved ResNet Based on the Adjustable Shortcut Connections. IEEE Access 2018, 6, 18967–18974. [Google Scholar] [CrossRef]
  30. Lin, Y.; Tu, Y.; Dou, Z.; Wu, Z. The Application of Deep Learning in Communication Signal Modulation Recognition. In Proceedings of the 2017 IEEE/CIC International Conference on Communications in China, Qingdao, China, 22–24 October 2017; pp. 1–5. [Google Scholar]
  31. Mu, Y. Averaging Projected Stochastic Gradient Descent for Large Scale Least Square Problem. Master’s Thesis, University of Massachusetts Boston, Boston, MA, USA, 2012. [Google Scholar]
  32. Graves, A.; Mohamed, A.R.; Hinton, G. Speech recognition with deep recurrent neural networks. In Proceedings of the 2013 IEEE International Conference on Acoustics, Speech and Signal Processing (Icassp), Vancouver, BC, Canada, 26–31 May 2013; pp. 6645–6649.
  33. Krizhevsky, A.; Sutskever, I.; Hinton, G.E. Imagenet classification with deep convolutional neural networks. In Advances in Neural Information Processing Systems 25; Pereira, F., Burges, C.J.C., Bottou, L., Weinberger, K.Q., Eds.; Curran Associates, Inc.: Red Hook, NY, USA, 2012; pp. 1097–1105. [Google Scholar]
  34. Hinton, G.E.; Salakhutdinov, R.R. Reducing the dimensionality of data with neural networks. Science 2006, 313, 504–507. [Google Scholar] [CrossRef] [PubMed]
  35. Johnson, J.; Alahi, A.; Li, F.F. Perceptual Losses for Real-Time Style Transfer and Super-Resolution. arXiv, 2016; arXiv:1603.08155. [Google Scholar]
  36. Kingma, D.P.; Ba, J. ADAM: A Method for Stochastic Optimization. In Proceedings of the International Conference on Learning Representations, San Diego, CA, USA, 7–9 May 2015; pp. 1–13. [Google Scholar]
  37. Deng, J.; Dong, W.; Socher, R.; Li, L.-J.; Li, K.; Fei-Fei, L. Imagenet: A large-scale hierarchical image database. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Miami, FL, USA, 20–25 June 2009; pp. 248–255. [Google Scholar]
  38. Thomee, B.; Shamma, D.A.; Friedland, G.; Elizalde, B.; Ni, K.; Poland, D.; Borth, D.; Li, L.-J. Yfcc100m: The new data in multimedia research. Commun. ACM 2016, 59, 64–73. [Google Scholar] [CrossRef]
  39. Zhang, B.; Liu, W.; Mao, Z.; Liu, J.; Shen, L. Cooperative and geometric learning algorithm (CGLA) for path planning of UAVs with limited information. Automatica 2014, 50, 809–820. [Google Scholar] [CrossRef]
  40. Zeiler, M.D.; Fergus, R. Visualizing and understanding convolutional neural networks. arXiv, 2013; arXiv:1311.2901. [Google Scholar]
  41. Donahue, J.; Jia, Y.; Vinyals, O.; Hoffman, J.; Zhang, N.; Tzeng, E.; Darrell, T. Decaf: A deep convolutional activation feature for generic visual recognition. arXiv, 2013; arXiv:1310.1531. [Google Scholar]
  42. Arthur, D.; Vassilvitskii, S. k-means++: The Advantages of Careful Seeding. In Proceedings of the Eighteenth Annual ACM-SIAM Symposium on Discrete Algorithms, New Orleans, LA, USA, 7–9 January 2007; Volume 11, pp. 1027–1035. [Google Scholar]
  43. Bahmani, B.; Moseley, B.; Vattani, A.; Kumar, R.; Vassilvitskii, S. Scalable k-means++. Proc. Vldb Endow. 2012, 5, 622–633. [Google Scholar] [CrossRef]
  44. Simonyan, K.; Zisserman, A. Very Deep Convolutional Networks for Large-Scale Image Recognition, Computer Science—Computer Vision and Pattern Recognition. arXiv, 2014; arXiv:1409.1556. [Google Scholar]
  45. Zeiler, M.D.; Fergus, R. Stochastic Pooling for Regularization of Deep Convolutional Neural Networks. arXiv, 2013; arXiv:1301.3557. [Google Scholar]
  46. Sujana, S.R.; Abisheck, S.S.; Ahmed, A.T.; Chandran, K.S. Real Time Object Identification Using Deep Convolutional Neural Networks. In Proceedings of the 2017 International Conference on Communication and Signal Processing (ICCSP), Chennai, India, 6–8 April 2017; pp. 1801–1805. [Google Scholar]
  47. Chen, L.C.; Papandreou, G.; Kokkinos, I.; Murphy, K.; Yuille, A.L. DeepLab: Semantic Image Segmentation with Deep Convolutional Nets, Atrous Convolution, and Fully Connected CRFs. IEEE Trans. Pattern Anal. Mach. Intell. 2018, 40, 834–848. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  48. Erhan, D.; Szegedy, C.; Toshev, A.; Anguelov, D. Scalable Object Detection Using Deep Neural Networks. Proc. IEEE Conf. Comput. Vis. Pattern Recognit. 2013, 3, 2155–2162. [Google Scholar]
  49. Everingham, M.; Van Gool, L.; Williams, C.K.; Winn, J.; Zisserman, A. The PASCAL Visual Object Classes (VOC) Challenge. Int. J. Comput. Vis. 2010, 88, 303–338. [Google Scholar] [CrossRef]
  50. Salton, G.; McGill, M.J. Introduction to Modern Information Retrieval; McGraw-Hill: New York, NY, USA, 1986. [Google Scholar]
Figure 1. (a) Original image; (b) Anchor box defining; (c) Object recognition.
Figure 1. (a) Original image; (b) Anchor box defining; (c) Object recognition.
Sensors 18 03041 g001
Figure 2. K-means++ clustering outcome for box number determination. IOU: intersection over union.
Figure 2. K-means++ clustering outcome for box number determination. IOU: intersection over union.
Sensors 18 03041 g002
Figure 3. Clustering box dimensions on PASCLE VOC and COCO. We run k-means++ clustering on the dimensions of bounding boxes to get good priors for EAO model. Both sets of priors favor thinner, taller boxes while COCO has greater variation in size than PASCLE VOC.
Figure 3. Clustering box dimensions on PASCLE VOC and COCO. We run k-means++ clustering on the dimensions of bounding boxes to get good priors for EAO model. Both sets of priors favor thinner, taller boxes while COCO has greater variation in size than PASCLE VOC.
Sensors 18 03041 g003
Figure 4. (a) Top prior boxes acquired from SSD; (b) Bottom prior boxes acquired from EAO.
Figure 4. (a) Top prior boxes acquired from SSD; (b) Bottom prior boxes acquired from EAO.
Sensors 18 03041 g004
Figure 5. EAO architecture.
Figure 5. EAO architecture.
Sensors 18 03041 g005
Figure 6. Working performance for EAO 512 and SSD512.
Figure 6. Working performance for EAO 512 and SSD512.
Sensors 18 03041 g006
Figure 7. Working performance for EAO 512 and SSD512.
Figure 7. Working performance for EAO 512 and SSD512.
Sensors 18 03041 g007
Figure 8. Some examples of EAO recognition results.
Figure 8. Some examples of EAO recognition results.
Sensors 18 03041 g008
Figure 9. Convergence process of object recognition.
Figure 9. Convergence process of object recognition.
Sensors 18 03041 g009
Table 1. Network configuration parameters.
Table 1. Network configuration parameters.
Box Generation123456Sum of BoxesAverage IOU
Anchor box 256----- 256 × 9 230633.7%
Anchor box 512----- 512 × 9 460836.4%
Default box 38 × 38 × 6 19 × 19 × 6 10 × 10 × 6 5 × 5 × 6 3 × 3 × 6 1 × 1 × 6 11,64054.7%
YOLO prior box----- 14 × 14 × 9 176466.5%
EAO SSE box † 38 × 38 × 3 19 × 19 × 3 10 × 10 × 6 5 × 5 × 6 3 × 3 × 9 1 × 1 × 9 595567.4%
EAO IoU box ‡ 38 × 38 × 3 19 × 19 × 3 10 × 10 × 6 5 × 5 × 6 3 × 3 × 9 1 × 1 × 9 595569.3%
† The standard Euclidean distance as a distance calculation method for K-means++. ‡ The intersection over union (IoU) as a distance calculation method for K-means++.
Table 2. Working parameters for different datasets.
Table 2. Working parameters for different datasets.
Working Parameters β 1 β 2 ε Learning Rate
PASCAL VOC 20070.750.777 1 × 10 8 0.0007
PASCAL VOC 20120.90.999 1 × 10 8 0.008
MS COCO0.90.999 1 × 10 8 0.0009
Table 3. PASCAL VOC2007 test detection results.
Table 3. PASCAL VOC2007 test detection results.
Method mAP ( % ) AeroBikeBirdBoatBottleBusCarCatChairCowTableDogHorseMbikePersonPlantSheepSofaTrainTV
Faster R-CNN73.276.27970.965.552.183.184.786.452.081.965.784.884.677.576.738.873.673.983.072.6
YOLO v273.486.38274.859.251.879.876.590.652.178.258.589.382.583.481.349.177.262.483.868.7
SSD30074.375.580.272.366.347.683.084.286.154.778.373.984.585.382.676.248.673.976.083.474.0
SSD51276.882.484.778.473.853.286.287.586.057.883.170.284.985.283.979.750.377.973.982.575.3
EAO30074.975.780.174.366.653.682.083.685.758.678.275.983.783.382.777.249.973.975.382.674.6
EAO51278.185.785.478.871.355.484.987.386.959.282.874.385.987.185.781.954.578.774.184.976.3
Table 4. PASCAL VOC2012 test detection results.
Table 4. PASCAL VOC2012 test detection results.
MethodData mAP % AeroBikeBirdBoatBottleBusCarCatChairCowTableDogHorseMbikePersonPlantSheepSofaTrainTV
Faster R-CNNVOC07+1270.484.979.874.353.949.877.575.988.545.677.155.386.981.780.979.640.172.660.981.261.5
YOLO v2VOC07+1257.97767.257.738.322.768.355.981.436.260.848.577.272.371.363.528.952.254.873.950.8
SSD300VOC07+1272.485.680.170.557.646.279.476.189.253.077.060.887.083.182.379.445.975.969.581.967.5
SSD512VOC07+1274.987.482.375.85952.681.781.590.055.479.159.888.484.384.783.350.278.066.386.372.0
EAO300VOC07+1273.686.883.57458.651.882.083.685.753.475.958.282.183.382.777.948.977.468.482.468.2
EAO512VOC07+1275.876.584.375.466.553.683.887.285.658.378.377.983.585.982.478.355.377.570.280.974.3
EAO300VOC07+12+iLOC76.687.585.477.664.754.684.788.286.958.479.471.584.884.682.579.753.877.673.481.175.4
EAO512VOC07+12+iLOC79.488.786.479.871.957.484.987.488.959.784.877.387.987.784.186.956.579.774.585.977.3
Table 5. MS COCO test detection results.
Table 5. MS COCO test detection results.
ModesBoxes mAP @ [ 0.5 : 0.95 ] mAP @ 0.5 mAP @ 0.75
Faster R-CNNRPN 30021.641.520.9
ION-21.442.119.7
YOLO9620.443.519.4
SSD512873223.543.923.7
EAO300595521.744.123.6
EAO512595526.947.628.2
Table 6. Testing results on inference time comparison.
Table 6. Testing results on inference time comparison.
MethodmAPFPSNumber of Boxes
Faster R-CNN73.27300
YOLO63.44598
Fast YOLO52.715598
SSD30072.1587308
SSD51275.121200,097
EAO30076.6615955
EAO51279.424160,785

Share and Cite

MDPI and ACS Style

Zhu, X.; Ding, B.; Meng, Q.; Gu, L.; Yang, Y. Statistic Experience Based Adaptive One-Shot Detector (EAO) for Camera Sensing System. Sensors 2018, 18, 3041. https://doi.org/10.3390/s18093041

AMA Style

Zhu X, Ding B, Meng Q, Gu L, Yang Y. Statistic Experience Based Adaptive One-Shot Detector (EAO) for Camera Sensing System. Sensors. 2018; 18(9):3041. https://doi.org/10.3390/s18093041

Chicago/Turabian Style

Zhu, Xiaoning, Bojian Ding, Qingyue Meng, Lize Gu, and Yixian Yang. 2018. "Statistic Experience Based Adaptive One-Shot Detector (EAO) for Camera Sensing System" Sensors 18, no. 9: 3041. https://doi.org/10.3390/s18093041

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop