Next Article in Journal
On-Chip Sensing of Thermoelectric Thin Film’s Merit
Next Article in Special Issue
A Geospatial Semantic Enrichment and Query Service for Geotagged Photographs
Previous Article in Journal
Robust Indoor Human Activity Recognition Using Wireless Signals
Previous Article in Special Issue
MagicFinger: 3D Magnetic Fingerprints for Indoor Location
Article Menu

Export Article

Open AccessArticle
Sensors 2015, 15(7), 17209-17231; doi:10.3390/s150717209

Marginalised Stacked Denoising Autoencoders for Robust Representation of Real-Time Multi-View Action Recognition

1
School of Computing and Information Systems, Kingston University, Penrhyn Road, Kingston upon Thames KT1 2EE, UK
2
Faculty of Science, Engineering and Computing, Kingston University, Penrhyn Road, Kingston upon Thames KT1 2EE, UK
*
Author to whom correspondence should be addressed.
Academic Editors: Jesús Fontecha and Paul Mccullagh
Received: 1 May 2015 / Revised: 9 July 2015 / Accepted: 9 July 2015 / Published: 16 July 2015
View Full-Text   |   Download PDF [724 KB, uploaded 16 July 2015]   |  

Abstract

Multi-view action recognition has gained a great interest in video surveillance, human computer interaction, and multimedia retrieval, where multiple cameras of different types are deployed to provide a complementary field of views. Fusion of multiple camera views evidently leads to more robust decisions on both tracking multiple targets and analysing complex human activities, especially where there are occlusions. In this paper, we incorporate the marginalised stacked denoising autoencoders (mSDA) algorithm to further improve the bag of words (BoWs) representation in terms of robustness and usefulness for multi-view action recognition. The resulting representations are fed into three simple fusion strategies as well as a multiple kernel learning algorithm at the classification stage. Based on the internal evaluation, the codebook size of BoWs and the number of layers of mSDA may not significantly affect recognition performance. According to results on three multi-view benchmark datasets, the proposed framework improves recognition performance across all three datasets and outputs record recognition performance, beating the state-of-art algorithms in the literature. It is also capable of performing real-time action recognition at a frame rate ranging from 33 to 45, which could be further improved by using more powerful machines in future applications. View Full-Text
Keywords: deep learning; marginalised stacked denoising autoencoders; bag of words; multiple kernel learning; multi-view action recognition deep learning; marginalised stacked denoising autoencoders; bag of words; multiple kernel learning; multi-view action recognition
This is an open access article distributed under the Creative Commons Attribution License which permits unrestricted use, distribution, and reproduction in any medium, provided the original work is properly cited. (CC BY 4.0).

Scifeed alert for new publications

Never miss any articles matching your research from any publisher
  • Get alerts for new papers matching your research
  • Find out the new papers from selected authors
  • Updated daily for 49'000+ journals and 6000+ publishers
  • Define your Scifeed now

SciFeed Share & Cite This Article

MDPI and ACS Style

Gu, F.; Flórez-Revuelta, F.; Monekosso, D.; Remagnino, P. Marginalised Stacked Denoising Autoencoders for Robust Representation of Real-Time Multi-View Action Recognition. Sensors 2015, 15, 17209-17231.

Show more citation formats Show less citations formats

Related Articles

Article Metrics

Article Access Statistics

1

Comments

[Return to top]
Sensors EISSN 1424-8220 Published by MDPI AG, Basel, Switzerland RSS E-Mail Table of Contents Alert
Back to Top