Next Article in Journal
Geometric Accuracy of Sentinel-1A and 1B Derived from SAR Raw Data with GPS Surveyed Corner Reflector Positions
Previous Article in Journal
Individual and Interactive Influences of Anthropogenic and Ecological Factors on Forest PM2.5 Concentrations at an Urban Scale
Article Menu
Issue 4 (April) cover image

Export Article

Open AccessArticle
Remote Sens. 2018, 10(4), 522; https://doi.org/10.3390/rs10040522

Modeling Environments Hierarchically with Omnidirectional Imaging and Global-Appearance Descriptors

Department of Systems Engineering and Automation, Miguel Hernández University, 03202 Elx Alicante, Spain
Current address: Avda. de la Universidad, s/n. Ed. Innova, 03202 Elx Alicante, Spain.
These authors contributed equally to this work.
*
Author to whom correspondence should be addressed.
Received: 19 January 2018 / Revised: 20 March 2018 / Accepted: 23 March 2018 / Published: 26 March 2018
(This article belongs to the Section Remote Sensing Image Processing)

Abstract

In this work, a framework is proposed to build topological models in mobile robotics, using an omnidirectional vision sensor as the only source of information. The model is structured hierarchically into three layers, from one high-level layer which permits a coarse estimation of the robot position to one low-level layer to refine this estimation efficiently. The algorithm is based on the use of clustering approaches to obtain compact topological models in the high-level layers, combined with global appearance techniques to represent robustly the omnidirectional scenes. Compared to the classical approaches based on the extraction and description of local features, global-appearance descriptors lead to models that can be interpreted and handled more intuitively. However, while local-feature techniques have been extensively studied in the literature, global-appearance ones require to be evaluated in detail to test their efficacy in map-building tasks. The proposed algorithms are tested with a set of publicly available panoramic images captured in realistic environments. The results show that global-appearance descriptors along with some specific clustering algorithms constitute a robust alternative to create a hierarchical representation of the environment. View Full-Text
Keywords: computer vision; omnidirectional imaging; global-appearance descriptors; environment modeling; clustering computer vision; omnidirectional imaging; global-appearance descriptors; environment modeling; clustering
Figures

Graphical abstract

This is an open access article distributed under the Creative Commons Attribution License which permits unrestricted use, distribution, and reproduction in any medium, provided the original work is properly cited. (CC BY 4.0).
SciFeed

Share & Cite This Article

MDPI and ACS Style

Payá, L.; Peidró, A.; Amorós, F.; Valiente, D.; Reinoso, O. Modeling Environments Hierarchically with Omnidirectional Imaging and Global-Appearance Descriptors. Remote Sens. 2018, 10, 522.

Show more citation formats Show less citations formats

Note that from the first issue of 2016, MDPI journals use article numbers instead of page numbers. See further details here.

Related Articles

Article Metrics

Article Access Statistics

1

Comments

[Return to top]
Remote Sens. EISSN 2072-4292 Published by MDPI AG, Basel, Switzerland RSS E-Mail Table of Contents Alert
Back to Top