Sign in to use this feature.

Years

Between: -

Subjects

remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline

Journals

Article Types

Countries / Regions

Search Results (17)

Search Parameters:
Keywords = hand-crafted features survey

Order results
Result details
Results per page
Select all
Export citation of selected articles as:
73 pages, 2833 KiB  
Article
A Comprehensive Methodological Survey of Human Activity Recognition Across Diverse Data Modalities
by Jungpil Shin, Najmul Hassan, Abu Saleh Musa Miah and Satoshi Nishimura
Sensors 2025, 25(13), 4028; https://doi.org/10.3390/s25134028 - 27 Jun 2025
Cited by 1 | Viewed by 1499
Abstract
Human Activity Recognition (HAR) systems aim to understand human behavior and assign a label to each action, attracting significant attention in computer vision due to their wide range of applications. HAR can leverage various data modalities, such as RGB images and video, skeleton, [...] Read more.
Human Activity Recognition (HAR) systems aim to understand human behavior and assign a label to each action, attracting significant attention in computer vision due to their wide range of applications. HAR can leverage various data modalities, such as RGB images and video, skeleton, depth, infrared, point cloud, event stream, audio, acceleration, and radar signals. Each modality provides unique and complementary information suited to different application scenarios. Consequently, numerous studies have investigated diverse approaches for HAR using these modalities. This survey includes only peer-reviewed research papers published in English to ensure linguistic consistency and academic integrity. This paper presents a comprehensive survey of the latest advancements in HAR from 2014 to 2025, focusing on Machine Learning (ML) and Deep Learning (DL) approaches categorized by input data modalities. We review both single-modality and multi-modality techniques, highlighting fusion-based and co-learning frameworks. Additionally, we cover advancements in hand-crafted action features, methods for recognizing human–object interactions, and activity detection. Our survey includes a detailed dataset description for each modality, as well as a summary of the latest HAR systems, accompanied by a mathematical derivation for evaluating the deep learning model for each modality, and it also provides comparative results on benchmark datasets. Finally, we provide insightful observations and propose effective future research directions in HAR. Full article
(This article belongs to the Special Issue Computer Vision and Sensors-Based Application for Intelligent Systems)
Show Figures

Figure 1

24 pages, 822 KiB  
Article
Survey on Image-Based Vehicle Detection Methods
by Mortda A. A. Adam and Jules R. Tapamo
World Electr. Veh. J. 2025, 16(6), 303; https://doi.org/10.3390/wevj16060303 - 29 May 2025
Viewed by 857
Abstract
Vehicle detection is essential for real-world applications such as road surveillance, intelligent transportation systems, and autonomous driving, where high accuracy and real-time performance are critical. However, achieving robust detection remains challenging due to scene complexity, occlusion, scale variation, and varying lighting conditions. Over [...] Read more.
Vehicle detection is essential for real-world applications such as road surveillance, intelligent transportation systems, and autonomous driving, where high accuracy and real-time performance are critical. However, achieving robust detection remains challenging due to scene complexity, occlusion, scale variation, and varying lighting conditions. Over the past two decades, numerous studies have been proposed to address these issues. This study presents a comprehensive and structured survey of image-based vehicle detection methods, systematically comparing classical machine learning techniques based on handcrafted features with modern deep learning approaches. Deep learning methods are categorized into one-stage detectors (e.g., YOLO, SSD, FCOS, CenterNet), two-stage detectors (e.g., Faster R-CNN, Mask R-CNN), transformer-based detectors (e.g., DETR, Swin Transformer), and GAN-based methods, highlighting architectural trade-offs concerning speed, accuracy, and practical deployment. We analyze widely adopted performance metrics from recent studies, evaluate characteristics and limitations of popular vehicle detection datasets, and explicitly discuss technical challenges, including domain generalization, environmental variability, computational constraints, and annotation quality. The survey concludes by clearly identifying open research challenges and promising future directions, such as efficient edge deployment strategies, multimodal data fusion, transformer-based enhancements, and integration with Vehicle-to-Everything (V2X) communication systems. Full article
(This article belongs to the Special Issue Vehicle Safe Motion in Mixed Vehicle Technologies Environment)
Show Figures

Figure 1

52 pages, 29859 KiB  
Review
2D Object Detection: A Survey
by Emanuele Malagoli and Luca Di Persio
Mathematics 2025, 13(6), 893; https://doi.org/10.3390/math13060893 - 7 Mar 2025
Cited by 1 | Viewed by 3554
Abstract
Object detection is a fundamental task in computer vision, aiming to identify and localize objects of interest within an image. Over the past two decades, the domain has changed profoundly, evolving into an active and fast-moving field while simultaneously becoming the foundation for [...] Read more.
Object detection is a fundamental task in computer vision, aiming to identify and localize objects of interest within an image. Over the past two decades, the domain has changed profoundly, evolving into an active and fast-moving field while simultaneously becoming the foundation for a wide range of modern applications. This survey provides a comprehensive review of the evolution of 2D generic object detection, tracing its development from traditional methods relying on handcrafted features to modern approaches driven by deep learning. The review systematically categorizes contemporary object detection methods into three key paradigms: one-stage, two-stage, and transformer-based, highlighting their development milestones and core contributions. The paper provides an in-depth analysis of each paradigm, detailing landmark methods and their impact on the progression of the field. Additionally, the survey examines some fundamental components of 2D object detection such as loss functions, datasets, evaluation metrics, and future trends. Full article
(This article belongs to the Special Issue Advanced Research in Image Processing and Optimization Methods)
Show Figures

Graphical abstract

51 pages, 21553 KiB  
Review
Deep Feature-Based Hyperspectral Object Tracking: An Experimental Survey and Outlook
by Yuchao Wang, Xu Li, Xinyan Yang, Fuyuan Ge, Baoguo Wei, Lixin Li and Shigang Yue
Remote Sens. 2025, 17(4), 645; https://doi.org/10.3390/rs17040645 - 13 Feb 2025
Cited by 1 | Viewed by 1820
Abstract
With the rapid advancement of hyperspectral imaging technology, hyperspectral object tracking (HOT) has become a research hotspot in the field of remote sensing. Advanced HOT methods have been continuously proposed and validated on scarce datasets in recent years, which can be roughly divided [...] Read more.
With the rapid advancement of hyperspectral imaging technology, hyperspectral object tracking (HOT) has become a research hotspot in the field of remote sensing. Advanced HOT methods have been continuously proposed and validated on scarce datasets in recent years, which can be roughly divided into handcrafted feature-based methods and deep feature-based methods. Compared with methods via handcrafted features, deep feature-based methods can extract highly discriminative semantic features from hyperspectral images (HSIs) and achieve excellent tracking performance, making them more favored by the hyperspectral tracking community. However, deep feature-based HOT still faces challenges such as data-hungry, band gap, low tracking efficiency, etc. Therefore, it is necessary to conduct a thorough review of current trackers and unresolved problems in the HOT field. In this survey, we systematically classify and conduct a comprehensive analysis of 13 state-of-the-art deep feature-based hyperspectral trackers. First, we classify and analyze the trackers based on the framework and tracking process. Second, the trackers are compared and analyzed in terms of tracking accuracy and speed on two datasets for cross-validation. Finally, we design a specialized experiment for small object tracking (SOT) to further validate the tracking performance. Through in-depth investigation, the advantages and weaknesses of current HOT technology based on deep features are clearly demonstrated, which also points out the directions for future development. Full article
(This article belongs to the Special Issue Remote Sensing Image Thorough Analysis by Advanced Machine Learning)
Show Figures

Figure 1

45 pages, 5749 KiB  
Article
Evaluation of Hand-Crafted Feature Extraction for Fault Diagnosis in Rotating Machinery: A Survey
by René-Vinicio Sánchez, Jean Carlo Macancela, Luis-Renato Ortega, Diego Cabrera, Fausto Pedro García Márquez and Mariela Cerrada
Sensors 2024, 24(16), 5400; https://doi.org/10.3390/s24165400 - 21 Aug 2024
Cited by 3 | Viewed by 2518
Abstract
This article presents a comprehensive collection of formulas and calculations for hand-crafted feature extraction of condition monitoring signals. The documented features include 123 for the time domain and 46 for the frequency domain. Furthermore, a machine learning-based methodology is presented to evaluate the [...] Read more.
This article presents a comprehensive collection of formulas and calculations for hand-crafted feature extraction of condition monitoring signals. The documented features include 123 for the time domain and 46 for the frequency domain. Furthermore, a machine learning-based methodology is presented to evaluate the performance of features in fault classification tasks using seven data sets of different rotating machines. The evaluation methodology involves using seven ranking methods to select the best ten hand-crafted features per method for each database, to be subsequently evaluated by three types of classifiers. This process is applied exhaustively by evaluation groups, combining our databases with an external benchmark. A summary table of the performance results of the classifiers is also presented, including the percentage of classification and the number of features required to achieve that value. Through graphic resources, it has been possible to show the prevalence of certain features over others, how they are associated with the database, and the order of importance assigned by the ranking methods. In the same way, finding which features have the highest appearance percentages for each database in all experiments has been possible. The results suggest that hand-crafted feature extraction is an effective technique with low computational cost and high interpretability for fault identification and diagnosis. Full article
Show Figures

Figure 1

25 pages, 930 KiB  
Systematic Review
Reviewing Material-Sensitive Computed Tomography: From Handcrafted Algorithms to Modern Deep Learning
by Moritz Weiss and Tobias Meisen
NDT 2024, 2(3), 286-310; https://doi.org/10.3390/ndt2030018 - 30 Jul 2024
Cited by 2 | Viewed by 1471
Abstract
Computed tomography (CT) is a widely utilised imaging technique in both clinical and industrial applications. CT scan results, presented as a volume revealing linear attenuation coefficients, are intricately influenced by scan parameters and the sample’s geometry and material composition. Accurately mapping these coefficients [...] Read more.
Computed tomography (CT) is a widely utilised imaging technique in both clinical and industrial applications. CT scan results, presented as a volume revealing linear attenuation coefficients, are intricately influenced by scan parameters and the sample’s geometry and material composition. Accurately mapping these coefficients to specific materials is a complex task. Traditionally, material decomposition in CT relied on classical algorithms using handcrafted features based on X-ray physics. However, there is a rising trend towards data-driven approaches, particularly deep learning, which offer promising improvements in accuracy and efficiency. This survey explores the transition from classical to data-driven approaches in material-sensitive CT, examining a comprehensive corpus of literature identified through a detailed and reproducible search using Scopus. Our analysis addresses several key research questions: the origin and generation of training datasets, the models and architectures employed, the extent to which deep learning methods reduce the need for domain-specific expertise, and the hardware requirements for training these models. We explore the implications of these findings on the integration of deep learning into CT practices and the potential reduction in the necessity for extensive domain knowledge. In conclusion, this survey highlights a significant shift towards deep learning in material-resolving CT and discusses the challenges and opportunities this presents. The transition suggests a future where data-driven approaches may dominate, offering enhanced precision and robustness in material-resolving CT while potentially transforming the role of domain experts in the field. Full article
Show Figures

Figure 1

61 pages, 7868 KiB  
Article
Advances in Facial Expression Recognition: A Survey of Methods, Benchmarks, Models, and Datasets
by Thomas Kopalidis, Vassilios Solachidis, Nicholas Vretos and Petros Daras
Information 2024, 15(3), 135; https://doi.org/10.3390/info15030135 - 28 Feb 2024
Cited by 25 | Viewed by 27502
Abstract
Recent technological developments have enabled computers to identify and categorize facial expressions to determine a person’s emotional state in an image or a video. This process, called “Facial Expression Recognition (FER)”, has become one of the most popular research areas in computer vision. [...] Read more.
Recent technological developments have enabled computers to identify and categorize facial expressions to determine a person’s emotional state in an image or a video. This process, called “Facial Expression Recognition (FER)”, has become one of the most popular research areas in computer vision. In recent times, deep FER systems have primarily concentrated on addressing two significant challenges: the problem of overfitting due to limited training data availability, and the presence of expression-unrelated variations, including illumination, head pose, image resolution, and identity bias. In this paper, a comprehensive survey is provided on deep FER, encompassing algorithms and datasets that offer insights into these intrinsic problems. Initially, this paper presents a detailed timeline showcasing the evolution of methods and datasets in deep facial expression recognition (FER). This timeline illustrates the progression and development of the techniques and data resources used in FER. Then, a comprehensive review of FER methods is introduced, including the basic principles of FER (components such as preprocessing, feature extraction and classification, and methods, etc.) from the pro-deep learning era (traditional methods using handcrafted features, i.e., SVM and HOG, etc.) to the deep learning era. Moreover, a brief introduction is provided related to the benchmark datasets (there are two categories: controlled environments (lab) and uncontrolled environments (in the wild)) used to evaluate different FER methods and a comparison of different FER models. Existing deep neural networks and related training strategies designed for FER, based on static images and dynamic image sequences, are discussed. The remaining challenges and corresponding opportunities in FER and the future directions for designing robust deep FER systems are also pinpointed. Full article
(This article belongs to the Special Issue Deep Learning for Image, Video and Signal Processing)
Show Figures

Figure 1

29 pages, 1635 KiB  
Article
A Survey on Different Plant Diseases Detection Using Machine Learning Techniques
by Sk Mahmudul Hassan, Khwairakpam Amitab, Michal Jasinski, Zbigniew Leonowicz, Elzbieta Jasinska, Tomas Novak and Arnab Kumar Maji
Electronics 2022, 11(17), 2641; https://doi.org/10.3390/electronics11172641 - 24 Aug 2022
Cited by 19 | Viewed by 10554
Abstract
Early detection and identification of plant diseases from leaf images using machine learning is an important and challenging research area in the field of agriculture. There is a need for such kinds of research studies in India because agriculture is one of the [...] Read more.
Early detection and identification of plant diseases from leaf images using machine learning is an important and challenging research area in the field of agriculture. There is a need for such kinds of research studies in India because agriculture is one of the main sources of income which contributes seventeen percent of the total gross domestic product (GDP). Effective and improved crop products can increase the farmer’s profit as well as the economy of the country. In this paper, a comprehensive review of the different research works carried out in the field of plant disease detection using both state-of-art, handcrafted-features- and deep-learning-based techniques are presented. We address the challenges faced in the identification of plant diseases using handcrafted-features-based approaches. The application of deep-learning-based approaches overcomes the challenges faced in handcrafted-features-based approaches. This survey provides the research improvement in the identification of plant diseases from handcrafted-features-based to deep-learning-based models. We report that deep-learning-based approaches achieve significant accuracy rates on a particular dataset, but the performance of the model may be decreased significantly when the system is tested on field image condition or on different datasets. Among the deep learning models, deep learning with an inception layer such as GoogleNet and InceptionV3 have better ability to extract the features and produce higher performance results. We also address some of the challenges that are needed to be solved to identify the plant diseases effectively. Full article
(This article belongs to the Special Issue Machine Learning: System and Application Perspective)
Show Figures

Figure 1

20 pages, 2369 KiB  
Article
Recent Iris and Ocular Recognition Methods in High- and Low-Resolution Images: A Survey
by Young Won Lee and Kang Ryoung Park
Mathematics 2022, 10(12), 2063; https://doi.org/10.3390/math10122063 - 15 Jun 2022
Cited by 14 | Viewed by 4248
Abstract
Among biometrics, iris and ocular recognition systems are the methods that recognize eye features in an image. Such iris and ocular regions must have a certain image resolution to achieve a high recognition performance; otherwise, the risk of performance degradation arises. This is [...] Read more.
Among biometrics, iris and ocular recognition systems are the methods that recognize eye features in an image. Such iris and ocular regions must have a certain image resolution to achieve a high recognition performance; otherwise, the risk of performance degradation arises. This is even more critical in the case of iris recognition where detailed patterns are used. In cases where such low-resolution images are acquired and the acquisition apparatus and environment cannot be improved, recognition performance can be enhanced by obtaining high-resolution images with methods such as super-resolution reconstruction. However, previous survey papers have mainly summarized studies on high-resolution iris and ocular recognition, but do not provide detailed summaries of studies on low-resolution iris and ocular recognition. Therefore, we investigated high-resolution iris and ocular recognition methods and introduced in detail the low-resolution iris and ocular recognition methods and methods of solving the low-resolution problem. Furthermore, since existing survey papers have focused on and summarized studies on traditional handcrafted feature-based iris and ocular recognition, this survey paper also introduced the latest deep learning-based methods in detail. Full article
(This article belongs to the Special Issue Computational Intelligent and Image Processing)
Show Figures

Figure 1

54 pages, 10378 KiB  
Review
Iris Liveness Detection for Biometric Authentication: A Systematic Literature Review and Future Directions
by Smita Khade, Swati Ahirrao, Shraddha Phansalkar, Ketan Kotecha, Shilpa Gite and Sudeep D. Thepade
Inventions 2021, 6(4), 65; https://doi.org/10.3390/inventions6040065 - 6 Oct 2021
Cited by 25 | Viewed by 8484
Abstract
Biometrics is progressively becoming vital due to vulnerabilities of traditional security systems leading to frequent security breaches. Biometrics is an automated device that studies human beings’ physiological and behavioral features for their unique classification. Iris-based authentication offers stronger, unique, and contactless identification of [...] Read more.
Biometrics is progressively becoming vital due to vulnerabilities of traditional security systems leading to frequent security breaches. Biometrics is an automated device that studies human beings’ physiological and behavioral features for their unique classification. Iris-based authentication offers stronger, unique, and contactless identification of the user. Iris liveness detection (ILD) confronts challenges such as spoofing attacks with contact lenses, replayed video, and print attacks, etc. Many researchers focus on ILD to guard the biometric system from attack. Hence, it is vital to study the prevailing research explicitly associated with the ILD to address how developing technologies can offer resolutions to lessen the evolving threats. An exhaustive survey of papers on the biometric ILD was performed by searching the most applicable digital libraries. Papers were filtered based on the predefined inclusion and exclusion criteria. Thematic analysis was performed for scrutinizing the data extracted from the selected papers. The exhaustive review now outlines the different feature extraction techniques, classifiers, datasets and presents their critical evaluation. Importantly, the study also discusses the projects, research works for detecting the iris spoofing attacks. The work then realizes in the discovery of the research gaps and challenges in the field of ILD. Many works were restricted to handcrafted methods of feature extraction, which are confronted with bigger feature sizes. The study discloses that dep learning based automated ILD techniques shows higher potential than machine learning techniques. Acquiring an ILD dataset that addresses all the common Iris spoofing attacks is also a need of the time. The survey, thus, opens practical challenges in the field of ILD from data collection to liveness detection and encourage future research. Full article
(This article belongs to the Collection Feature Innovation Papers)
Show Figures

Figure 1

26 pages, 4889 KiB  
Article
A Fusion-Based Hybrid-Feature Approach for Recognition of Unconstrained Offline Handwritten Hindi Characters
by Danveer Rajpal, Akhil Ranjan Garg, Om Prakash Mahela, Hassan Haes Alhelou and Pierluigi Siano
Future Internet 2021, 13(9), 239; https://doi.org/10.3390/fi13090239 - 18 Sep 2021
Cited by 5 | Viewed by 3092
Abstract
Hindi is the official language of India and used by a large population for several public services like postal, bank, judiciary, and public surveys. Efficient management of these services needs language-based automation. The proposed model addresses the problem of handwritten Hindi character recognition [...] Read more.
Hindi is the official language of India and used by a large population for several public services like postal, bank, judiciary, and public surveys. Efficient management of these services needs language-based automation. The proposed model addresses the problem of handwritten Hindi character recognition using a machine learning approach. The pre-trained DCNN models namely; InceptionV3-Net, VGG19-Net, and ResNet50 were used for the extraction of salient features from the characters’ images. A novel approach of fusion is adopted in the proposed work; the DCNN-based features are fused with the handcrafted features received from Bi-orthogonal discrete wavelet transform. The feature size was reduced by the Principal Component Analysis method. The hybrid features were examined with popular classifiers namely; Multi-Layer Perceptron (MLP) and Support Vector Machine (SVM). The recognition cost was reduced by 84.37%. The model achieved significant scores of precision, recall, and F1-measure—98.78%, 98.67%, and 98.69%—with overall recognition accuracy of 98.73%. Full article
(This article belongs to the Special Issue Service-Oriented Systems and Applications)
Show Figures

Figure 1

16 pages, 24528 KiB  
Article
Human Behavior Analysis: A Survey on Action Recognition
by Bruno Degardin and Hugo Proença
Appl. Sci. 2021, 11(18), 8324; https://doi.org/10.3390/app11188324 - 8 Sep 2021
Cited by 17 | Viewed by 4858
Abstract
The visual recognition and understanding of human actions remain an active research domain of computer vision, being the scope of various research works over the last two decades. The problem is challenging due to its many interpersonal variations in appearance and motion dynamics [...] Read more.
The visual recognition and understanding of human actions remain an active research domain of computer vision, being the scope of various research works over the last two decades. The problem is challenging due to its many interpersonal variations in appearance and motion dynamics between humans, without forgetting the environmental heterogeneity between different video images. This complexity splits the problem into two major categories: action classification, recognising the action being performed in the scene, and spatiotemporal action localisation, concerning recognising multiple localised human actions present in the scene. Previous surveys mainly focus on the evolution of this field, from handcrafted features to deep learning architectures. However, this survey presents an overview of both categories and respective evolution within each one, the guidelines that should be followed and the current benchmarks employed for performance comparison between the state-of-the-art methods. Full article
(This article belongs to the Collection Machine Learning in Computer Engineering Applications)
Show Figures

Figure 1

17 pages, 3141 KiB  
Article
PointNet++ Network Architecture with Individual Point Level and Global Features on Centroid for ALS Point Cloud Classification
by Yang Chen, Guanlan Liu, Yaming Xu, Pai Pan and Yin Xing
Remote Sens. 2021, 13(3), 472; https://doi.org/10.3390/rs13030472 - 29 Jan 2021
Cited by 57 | Viewed by 9870
Abstract
Airborne laser scanning (ALS) point cloud has been widely used in the fields of ground powerline surveying, forest monitoring, urban modeling, and so on because of the great convenience it brings to people’s daily life. However, the sparsity and uneven distribution of point [...] Read more.
Airborne laser scanning (ALS) point cloud has been widely used in the fields of ground powerline surveying, forest monitoring, urban modeling, and so on because of the great convenience it brings to people’s daily life. However, the sparsity and uneven distribution of point clouds increases the difficulty of setting uniform parameters for semantic classification. The PointNet++ network is an end-to-end learning network for irregular point data and highly robust to small perturbations of input points along with corruption. It eliminates the need to calculate costly handcrafted features and provides a new paradigm for 3D understanding. However, each local region in the output is abstracted by its centroid and local feature that encodes the centroid’s neighborhood. The feature learned on the centroid point may not contain relevant information of itself for random sampling, especially in large-scale neighborhood balls. Moreover, the centroid point’s global-level information in each sample layer is also not marked. Therefore, this study proposed a modified PointNet++ network architecture which concentrates the point-level and global features on the centroid point towards the local features to facilitate classification. The proposed approach also utilizes a modified Focal Loss function to solve the extremely uneven category distribution on ALS point clouds. An elevation- and distance-based interpolation method is also proposed for the objects in ALS point clouds which exhibit discrepancies in elevation distributions. The experiments on the Vaihingen dataset of the International Society for Photogrammetry and Remote Sensing and the GML(B) 3D dataset demonstrate that the proposed method which provides additional contextual information to support classification achieves high accuracy with simple discriminative models and new state-of-the-art performance in power line categories. Full article
(This article belongs to the Special Issue Urban Multi-Category Object Detection Using Aerial Images)
Show Figures

Graphical abstract

20 pages, 1846 KiB  
Article
Automatic Hierarchical Classification of Kelps Using Deep Residual Features
by Ammar Mahmood, Ana Giraldo Ospina, Mohammed Bennamoun, Senjian An, Ferdous Sohel, Farid Boussaid, Renae Hovey, Robert B. Fisher and Gary A. Kendrick
Sensors 2020, 20(2), 447; https://doi.org/10.3390/s20020447 - 13 Jan 2020
Cited by 48 | Viewed by 5237
Abstract
Across the globe, remote image data is rapidly being collected for the assessment of benthic communities from shallow to extremely deep waters on continental slopes to the abyssal seas. Exploiting this data is presently limited by the time it takes for experts to [...] Read more.
Across the globe, remote image data is rapidly being collected for the assessment of benthic communities from shallow to extremely deep waters on continental slopes to the abyssal seas. Exploiting this data is presently limited by the time it takes for experts to identify organisms found in these images. With this limitation in mind, a large effort has been made globally to introduce automation and machine learning algorithms to accelerate both classification and assessment of marine benthic biota. One major issue lies with organisms that move with swell and currents, such as kelps. This paper presents an automatic hierarchical classification method local binary classification as opposed to the conventional flat classification to classify kelps in images collected by autonomous underwater vehicles. The proposed kelp classification approach exploits learned feature representations extracted from deep residual networks. We show that these generic features outperform the traditional off-the-shelf CNN features and the conventional hand-crafted features. Experiments also demonstrate that the hierarchical classification method outperforms the traditional parallel multi-class classifications by a significant margin (90.0% vs. 57.6% and 77.2% vs. 59.0%) on Benthoz15 and Rottnest datasets respectively. Furthermore, we compare different hierarchical classification approaches and experimentally show that the sibling hierarchical training approach outperforms the inclusive hierarchical approach by a significant margin. We also report an application of our proposed method to study the change in kelp cover over time for annually repeated AUV surveys. Full article
(This article belongs to the Special Issue Imaging Sensor Systems for Analyzing Subsea Environment and Life)
Show Figures

Figure 1

27 pages, 1189 KiB  
Review
A Survey of Vision-Based Human Action Evaluation Methods
by Qing Lei, Ji-Xiang Du, Hong-Bo Zhang, Shuang Ye and Duan-Sheng Chen
Sensors 2019, 19(19), 4129; https://doi.org/10.3390/s19194129 - 24 Sep 2019
Cited by 89 | Viewed by 7913
Abstract
The fields of human activity analysis have recently begun to diversify. Many researchers have taken much interest in developing action recognition or action prediction methods. The research on human action evaluation differs by aiming to design computation models and evaluation approaches for automatically [...] Read more.
The fields of human activity analysis have recently begun to diversify. Many researchers have taken much interest in developing action recognition or action prediction methods. The research on human action evaluation differs by aiming to design computation models and evaluation approaches for automatically assessing the quality of human actions. This line of study has become popular because of its explosively emerging real-world applications, such as physical rehabilitation, assistive living for elderly people, skill training on self-learning platforms, and sports activity scoring. This paper presents a comprehensive survey of approaches and techniques in action evaluation research, including motion detection and preprocessing using skeleton data, handcrafted feature representation methods, and deep learning-based feature representation methods. The benchmark datasets from this research field and some evaluation criteria employed to validate the algorithms’ performance are introduced. Finally, the authors present several promising future directions for further studies. Full article
(This article belongs to the Special Issue From Sensors to Ambient Intelligence for Health and Social Care)
Show Figures

Figure 1

Back to TopTop