Next Article in Journal
Learning Deep Hierarchical Spatial–Spectral Features for Hyperspectral Image Classification Based on Residual 3D-2D CNN
Previous Article in Journal
Experimental Evaluation of a UWB-Based Cooperative Positioning System for Pedestrians in GNSS-Denied Environment
Open AccessArticle

Deep Ego-Motion Classifiers for Compound Eye Cameras

Department of Eletrical and Computer Engineering and ASRI, Seoul National University, 1 Gwanak-ro, Gwanak-gu, Seoul 08826, Korea
Author to whom correspondence should be addressed.
Sensors 2019, 19(23), 5275;
Received: 17 October 2019 / Revised: 22 November 2019 / Accepted: 24 November 2019 / Published: 29 November 2019
(This article belongs to the Section Intelligent Sensors)
Compound eyes, also known as insect eyes, have a unique structure. They have a hemispheric surface, and a lot of single eyes are deployed regularly on the surface. Thanks to this unique form, using the compound images has several advantages, such as a large field of view (FOV) with low aberrations. We can exploit these benefits in high-level vision applications, such as object recognition, or semantic segmentation for a moving robot, by emulating the compound images that describe the captured scenes from compound eye cameras. In this paper, to the best of our knowledge, we propose the first convolutional neural network (CNN)-based ego-motion classification algorithm designed for the compound eye structure. To achieve this, we introduce a voting-based approach that fully utilizes one of the unique features of compound images, specifically, the compound images consist of a lot of single eye images. The proposed method classifies a number of local motions by CNN, and these local classifications which represent the motions of each single eye image, are aggregated to the final classification by a voting procedure. For the experiments, we collected a new dataset for compound eye camera ego-motion classification which contains scenes of the inside and outside of a certain building. The samples of the proposed dataset consist of two consequent emulated compound images and the corresponding ego-motion class. The experimental results show that the proposed method has achieved the classification accuracy of 85.0%, which is superior compared to the baselines on the proposed dataset. Also, the proposed model is light-weight compared to the conventional CNN-based image recognition algorithms such as AlexNet, ResNet50, and MobileNetV2. View Full-Text
Keywords: Bio-inspired structure; compound eye camera; compound image; ego-motion classification Bio-inspired structure; compound eye camera; compound image; ego-motion classification
Show Figures

Figure 1

MDPI and ACS Style

Yoo, H.; Cha, G.; Oh, S. Deep Ego-Motion Classifiers for Compound Eye Cameras. Sensors 2019, 19, 5275.

AMA Style

Yoo H, Cha G, Oh S. Deep Ego-Motion Classifiers for Compound Eye Cameras. Sensors. 2019; 19(23):5275.

Chicago/Turabian Style

Yoo, Hwiyeon; Cha, Geonho; Oh, Songhwai. 2019. "Deep Ego-Motion Classifiers for Compound Eye Cameras" Sensors 19, no. 23: 5275.

Find Other Styles
Note that from the first issue of 2016, MDPI journals use article numbers instead of page numbers. See further details here.

Article Access Map by Country/Region

Search more from Scilit
Back to TopTop