Next Article in Journal
A Visual Saliency-Based Neural Network Architecture for No-Reference Image Quality Assessment
Next Article in Special Issue
Modelling, Simulation and Performance Validation of the Pneumatic Actuation System of a Rehabilitation Device of the Human Hand Joints
Previous Article in Journal
Numerical Simulation and Field Test of the Interaction between Existing Station and Enclosure in Open Excavation and Adding Stories Construction
Previous Article in Special Issue
A Wheels-on-Knees Quadruped Assistive Robot to Carry Loads
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Assistive Robot with an AI-Based Application for the Reinforcement of Activities of Daily Living: Technical Validation with Users Affected by Neurodevelopmental Disorders

by
Nadia Nasri
1,2,*,
Roberto J. López-Sastre
1,
Soraya Pacheco-da-Costa
3,
Iván Fernández-Munilla
1,
Carlos Gutiérrez-Álvarez
1,
Thais Pousada-García
4,
Francisco Javier Acevedo-Rodríguez
1 and
Saturnino Maldonado-Bascón
1
1
Multisensory Analysis and Recognition Group (GRAM), Department of Signal Theory and Communications, University of Alcalá, 28805 Alcala de Henares, Spain
2
Institute for Computer Research, University of Alicante, P.O. Box 99, 03080 Alicante, Spain
3
FINEMEV, Department of Nursing and Physiotherapy, University of Alcalá, 28805 Alcala de Henares, Spain
4
CITIC (Centre for Information and Communications Technology Research), TALIONIS Research Group, Universidade da Coruña, 15071 A Coruña, Spain
*
Author to whom correspondence should be addressed.
Appl. Sci. 2022, 12(19), 9566; https://doi.org/10.3390/app12199566
Submission received: 18 July 2022 / Revised: 12 September 2022 / Accepted: 16 September 2022 / Published: 23 September 2022

Abstract

:
In this work, we propose the first study of a technical validation of an assistive robotic platform, which has been designed to assist people with neurodevelopmental disorders. The platform is called LOLA2 and it is equipped with an artificial intelligence-based application to reinforce the learning of daily life activities in people with neurodevelopmental problems. LOLA2 has been integrated with an ROS-based navigation system and a user interface for healthcare professionals and their patients to interact with it. Technically, we have been able to embed all these modules into an NVIDIA Jetson Xavier board, as well as an artificial intelligence agent for online action detection (OAD). This OAD approach provides a detailed report on the degree of performance of a set of daily life activities that are being learned or reinforced by users. All the human–robot interaction process to work with users with neurodevelopmental disorders has been designed by a multidisciplinary team. Among its main features are the ability to control the robot with a joystick, a graphical user interface application that shows video tutorials with the activities to reinforce or learn, and the ability to monitor the progress of the users as they complete tasks. The main objective of the assistive robotic platform LOLA2 is to provide a system that allows therapists to track how well the users understand and perform daily tasks. This paper focuses on the technical validation of the proposed platform and its application. To do so, we have carried out a set of tests with four users with neurodevelopmental problems and special physical conditions under the supervision of the corresponding therapeutic personnel. We present detailed results of all interventions with end users, analyzing the usability, effectiveness, and limitations of the proposed technology. During its initial technical validation with real users, LOLA2 was able to detect the actions of users with disabilities with high precision. It was able to distinguish four assigned daily actions with high accuracy, but some actions were more challenging due to the physical limitations of the users. Generally, the presence of the robot in the therapy sessions received excellent feedback from medical professionals as well as patients. Overall, this study demonstrates that our developed robot is capable of assisting and monitoring people with neurodevelopmental disorders in performing their daily living tasks.

1. Introduction

Our lives have become more accommodating and improved as technology advances rapidly, and these changes affect people from all walks of life. Besides improving general aspects of human life, some research teams also focus on improving the quality of life for people with functional diversity. These advances are intended to benefit these individuals. In recent years, advances in robotics and artificial intelligence (AI) have allowed social assistive robots (SARs) to become increasingly capable of helping and assisting humans in a variety of different environments and conditions. As outlined in [1], SARs are robots that provide assistance through social mechanisms rather than physical ones. A number of studies have been conducted on SARs and the interaction between humans and robots in relation to cognitive and social skills [2,3,4,5].
In 1988, the first documented robotic-assisted surgical procedure was performed with a robotic arm and a computerized tomography (CT) scanner for a CT-guided brain tumor biopsy [6]. Since then, technological advancements have greatly increased trust in robot capabilities and they have helped people with both their mental and physical needs [7,8,9,10].
Generally, SARs are utilized for the following purposes: caring for humans, performing domestic duties, carrying out human tasks or assisting in the performance of human tasks, enhancing education systems, and assisting in the provision of medical care. Based on Dawe et al.’s [11] analysis, social robots are primarily companions, providing pleasure, entertainment, role models, buddies for playing and learning, and coaches for providing information and demonstrating exercises.
As part of the development of SARs, one of the focus areas is the provision of assistance to people with neurodevelopmental disorders (NDDs) with their activities of daily living (ADLs). It is estimated that 74% of Spanish citizens, older than 6, with disabilities have difficulty performing ADLs [12], and prevalence rates of NDDs range from 4 to 13% in school-age children depending on the country [13,14,15]. These difficulties often continue into late adolescence and adulthood [16], so it is important to assist them in overcoming their limitations and learning or reinforcing their skills in performing ADLs. According to Pivetti et al. [17], children with NDDs can benefit from interacting with and learning from robots in a more engaging and social way. It was found that SARs assisted kids in attaining learning objectives, engaging in learning activities, and interacting with others as a result of this study.
In this work, we propose the first study of a technical validation for an assistive robotic platform, which has been designed to assist people with neurodevelopmental disorders in enhancing ADL learning. This platform is referred to as LOLA2 and it is based on an artificial intelligence application that is intended to support the learning and reinforcement of daily living activities by people with disabilities. Figure 1 shows an overview of the action monitoring process with LOLA2 and users with NDDs. Our robot is equipped with an online action detection module, which allows real-time monitoring of the actions that users are performing during interventions with the platform. Those with NDDs of all ages can benefit from this work in learning or reinforcing their daily living skills. Our robot is equipped with an intuitive graphical user interface that enables users or their responsible professionals to choose the ADLs they wish to learn or practice. Upon activating the AI-based action detection application, they can begin performing the requested action in front of the robot. A report containing the details of the action monitoring process can be reviewed by the responsible professional following each session to assess the user’s progress. The main contributions of this paper are:
  • We introduce the novel robotic platform called LOLA2. It is an improved version of our previous assistive platform LOLA [18]. Section 3 details all the technical improvements made.
  • For the robot, a user-friendly graphical interface is designed and developed to enable efficient human–robot interaction (see Section 3.2.2). Its objective is to interact with individuals with NDDs and to assist in their therapy sessions to reinforce activities of daily living. Based on artificial intelligence techniques, we have embedded in LOLA2 an online action detection module designed for monitoring ADLs, which is detailed in Section 3.2.3.
  • This work presents in Section 4 the first technical validation of the technology proposed with a set of four real final users with NDDs. The results confirm that our developed robot is capable of assisting and monitoring people with NDDs in performing their daily living tasks.
This paper is organized as follows—Section 2 reviews the literature on technological solutions for assistive robots and action detection applications. We present our assistive robot in Section 3 as well as the system for monitoring action applications based on artificial intelligence. In Section 4, experimental validation and results are presented and discussed. Finally, in Section 5, we conclude with some conclusions and suggestions for future lines of work.

2. Related Work

SARs have gained importance in the last decade and have demonstrated a variety of benefits and various spheres of life have been influenced by this. A variety of studies have been carried out in SARs, such as those investigating reductions of pain and anxiety as a result of hospitalization [19,20], consumption of meals [21], housework [22], monitoring the user’s health [23,24], teaching language skills [9], and reducing social isolation and improving well-being through social interaction with the user [1,25,26].
For children with NDDs, it has been proven that they experience difficulties taking care of themselves, playing, and moving, as well as thinking socially [27,28,29]. Gelosmini et al. [30] have determined that the use of robots will allow them to improve their quality of life, encourage human–human interaction, and develop social and cognitive skills.
In light of our robot’s ability to assist socially as well as monitoring actions, we have divided the related work into two parts.

2.1. Social Assistive Robots

A number of assistive robots have been shown to be effective in treating children with NDDs, according to [31], such as Troy [32,33], FACE [34], IROMEC [35], KASPAR [36,37,38], Lego Mindstorm [39,40], and NAO [41]. Designed to evoke one or more specific behaviors in children, each has a unique look, design element, and interaction technique. In addition, children with NDDs are particularly drawn to technological devices [42], and the presence of SARs in therapy can help them become more interested in getting treated and gain confidence and independence in their life [17].
The presence of SARs and their benefits in rehabilitation have also been the subject of several studies focusing on the robotic–human interaction [1,43]. Rehabilitation can be exhausting, painful, and difficult as its exercises must be performed according to specific guidelines. Previous works [44,45], which studied interaction with a SAR over short-term periods of time, suggest that including SARs into a practice that requires repetition may increase patients’ motivation. In some works [46,47], healthcare professionals have been surveyed to determine their perceptions of SARs and their use in therapy and rehabilitation, and their results suggest that SARs can be used to increase engagement, motivation, and compliance. Another study investigating the effectiveness of SAR feedback for rehabilitation showed that sessions with users using SARs enhanced task performance [48].
We have developed our robot based on all previous studies to assist in the rehabilitation and therapy process for people of all ages and with varying levels of NDDs so that healthcare professionals can track their improvement during the process.
While most SARs are expensive commercial platforms adapted for use with assistive objects, they are not accessible to a wide range of people of varying income levels [41,49,50,51]. However, our assistive robot is a low-cost model specially designed to assist individuals with NDDs in learning and reinforcing ADLs, and is integrated with an AI module for monitoring daily living activities [18].

2.2. AI for Monitoring ADLs

In essence, AI-based monitoring activities systems rely on action detection modules, which detect and recognize a specific sequence of sub-actions associated with a particular activity. This can be accomplished in a number of ways, including extraction of descriptive information from the environment accompanied by estimation of the human posture [52]. Sensor-based recognition is another popular method of detecting actions, which uses an algorithm to detect defined actions based on sensor information [53,54,55]. Although these methods have high accuracy, they are limited in terms of their ability to adapt to new actions with different objects as a result of the complexity of sensor placement and the possible requirement for additional sensors.
Alternatively, a vision-based method can be used for action detection, where a camera is used to capture action information. The modules integrated into PHAROS [7], RoboPhilo [56], NAO [57], and Gymmy [26] robots can be examples of vision-based methods to monitor activities. The systems mentioned above, however, are designed for the purpose of exercising and they do not include the daily activities upon which our work is based.
A number of studies have proposed a system of action monitoring that includes activities that are more similar to those that are performed on a daily basis [58,59,60]. As most of the proposed systems detect actions offline, the entire video must be fed into the action recognition model to determine when and where the specific actions have taken place [61,62,63]. The concept of early action detection has also been proposed, by which the action label of an action video is predicted prior to the end of the current action execution [64,65]. Based on the assumption that there is only one action instance in the video stream, the start and end frames are determined once the action has ended. Therefore, they are not suitable for monitoring actions according to our objectives, and we have chosen online action detection systems that are capable of detecting actions as soon as they occur. In the field of online action detection, few studies have been conducted [66,67,68,69], and the majority of those studies do not explicitly distinguish action from background [70]. This is why we have returned to our previous approach that can explicitly distinguish action from background [18].

3. Human-Robot Interaction Application for the Reinforcement of ADLs

3.1. The Assistive Robotic Platform: LOLA2

LOLA [18] is a low-cost mobile assistive robotic platform, which has been equipped with several sensors that allow it to navigate autonomously through the environment and interact with the users.
As we described in our previous work [18], the internal structure of LOLA is constructed from wood and metal, and the outer shell is entirely constructed from 3D printing. The entire mechanical and electrical design of the platform has been conceived by our research team. For more specific details about the main technical features of this differential wheeled robot, we refer the reader to [18].
For this research work, we have made some improvements with respect to the previous version. LOLA2 now integrates a new NVIDIA Jetson Xavier processing card. Contrary to the previous model, where we integrated a NIVIDA Jetson TX2. The Jetson board acts as an embedded Linux computer, capable of providing robotic platforms with the same level of functionality as a desktop computer. The Jetson Xavier, which we use for this project, is equipped with a Volta GPU with 384 CUDA Cores and 48 Tensor Cores, 8GB of RAM, and a 64-bit CPU Carmel ARM. This change in the Jetson processing board has been motivated by the need to increase the computing capacity of the platform. This has been necessary due to (1) the resources consumed by having a Robotic Operating System (ROS) [71] embedded in the system for all robot control; (2) the need to implement faster AI models for action recognition and navigation; and (3) the integration of the novel human–robot interaction applications specifically designed for the interventions with users with neurodevelopmental disorders. Other design improvements have been applied to the platform to obtain this new version: changes in the position of the sensors, elimination of ultrasonic sensors, and design changes in the casing. LOLA2 integrates a new vision sensor too. It is an Intel RealSense camera to receive high-quality RGB images, which are fed to our online action detection module. Figure 2 shows both versions of LOLA.

3.2. Human–Robot Interaction Application: Software Description

This paper describes a process of applied science in which our robotic platform interacts with both the healthcare personnel who must operate it and the end users of the platform. For this research, therefore, we have had to develop a complex human–machine interaction system, which is detailed in this section.
As is shown in Figure 1, the main objective with LOLA2 consists of the development of an AI application for the reinforcement of ADLs for users with neurodevelopmental disorders. It has therefore been necessary to provide LOLA2 with an interface to efficiently control its movement, as well as to develop a graphical user interface that allows efficient interaction with patients and healthcare professionals.

3.2.1. ROS Integration and Navigation Interface

Robot LOLA2 has been fully integrated with ROS [71] (Melodic version). The whole robotic platform, including all its sensors, is controlled via ROS. In addition, to provide the platform with advanced navigation capabilities, it was decided to use the ROS navigation libraries. Figure 3 shows the complete ROS nodes architecture we have embedded in LOLA, mainly using Python. Blue nodes represent distributed software ROS packages, and yellow nodes are those ROS nodes entirely developed by us to integrate ROS into our platform.
Recall that the ROS system is fully embedded in the NVIDIA Jetson board of LOLA2, communicating with an Arduino board that is in charge of the control of the motors and wheels. The ROS Navigation Stack, which is the backbone of the navigation system, receives input from the following sources: a map from the map server, odometry from the wheels, LIDAR data, and an estimated position, in this case provided by the Adaptive Monte Carlo Localization (AMCL) package. RVIZ can be used as the user interface. The visual environment provided by this ROS package shows the map and the estimated robot position. The user can direct the robot to move to a specific place while defining yaw rotation, as well as give it an initial position to assist it localize itself. The move_base core package (see Figure 3), which creates the linear and angular velocity commands required to move the robot to the appropriate place, will receive ROS messages from RVIZ. The information about any new potential obstacles, the odometry, and the projected localization are all continuously updated for all these commands and their parameters. We created the Hardware Interface package to fully integrate ROS into our robotic platform. It translates ROS commands to the Arduino board and provides signals back to the ROS system in the precise format required to complete the communication loop with the Arduino. The developed libraries for establishing serial port-based communication between the platform’s engines and ROS are included in the Arduino package in Figure 3. We have specifically created a communication protocol with commands that enable us to read the data from the encoders and to produce a series of speed directives to be sent to each wheel.
Finally, for enhancing the robot’s movement, as well as the usability of the platform, we have incorporated a joystick (Logitech gamepad F710) to enable professionals and users to more precisely control the robot’s displacement. With a Directional Pad or Thumbstick, the joystick can be used to navigate the robot in four different directions, depending upon the user’s position.

3.2.2. Graphical User Interfaces

Our assistive robot application was developed with a user-friendly graphical interface that includes distinct sections; see Figure 4. It is the responsibility of healthcare professionals to manage the interface shown, but it can also be managed by users who are cognitively capable of making their own decisions and taking control over the application. Any process of reinforcement of an activity of daily living requires the following steps:
  • Configuration step: The healthcare professional selects a user (identified by an ID) and particular action for the monitoring.
  • Activity observation: The user watches a video of the chosen action.
  • Activity monitoring: The user is encouraged to replicate the observed action, and the AI online action detection module starts the automatic monitoring of the performance. The software also provides the option to generate a report of the performance of the user during therapy sessions.
The user ID, as well as the action that we want the user to perform, can be selected from the main menu of the application (see Figure 4a). Once the configuration step has been completed, the final users must watch an example video of the chosen action. This phase is started with the button in Figure 4b. For this research work, our team had to record for each of the actions of interest, up to four videos, along with step-by-step sub-action instructions for achieving the final objective of that action. Figure 5 shows some examples of two actors performing some of the actions of interest. All this dataset of videos is integrated into LOLA2 along with the application. When the professional selects an action to work with, one of these sample videos is randomly presented to the user. This option illustrates and instructs the user on the steps involved in taking a daily action. Then, the user may begin performing the chosen action in accordance with the observed video. Once the user is ready to carry out the requested action, the activity monitoring step will be activated by professionals. An AI-based online action detection (OAD) system is integrated to monitor the user’s actions, as detailed in the following section. The developed application allows the generation of a report containing a detailed log of the entire work session with the patient. In particular, the log file can be used to obtain the following statistics: How much time has the user spent on certain tasks or actions? What actions are carried out in certain time slots? Have been the actions correctly performed?

3.2.3. AI for Online Action Monitoring of ADLs

Considering the need to monitor the activity of users, the inclusion of an action detection system in our robot was a no-brainer. As an standard action detection system would require that an entire video of the action be fed into the neural network to recognize the action correctly, we believe that these models are ineligible for monitoring actions in this project. They work offline, in the sense that the action of interest must be completely covered by the video. Instead, for our robotic application, we need an action recognition system that is able to work with partial observations of the actions, or, in other words: to address the problem of localizing actions in untrimmed videos as soon as they happen, which was coined as Online Action Detection (OAD). Therefore, action detections in an OAD technique must be made over video streams, requiring the use of incomplete observations in which the action segments may be more likely to be the exception than the rule relative to the background. Furthermore, this online setting includes a crucial element: the anticipation of the action. For an OAD model, the objective is to anticipate the action even before the action is fully completed, an aspect that is fundamental to attaining an efficient human–robot interaction in the monitoring application designed for this project: LOLA must be able to recognize the actions the users are performing in an online fashion, so as to provide a real-time response.
For our robotic platform, we propose a Pytorch implementation of the 3D CNN model described in [72]. See Figure 6 for details of the deep learning architecture deployed on the NVIDIA Jetson Xavier board. Our approach uses a 3D Convolutional Neural Network (CNN) that is trained on the UCF-101 dataset containing 101 distinct categories of actions [73]. As input, our OAD model is given a 16 frame clip. The model is able to cast an action prediction for every clip, classifying the actions as soon as they appear. Using the Jetson board and a live video stream received from the robot’s camera, this method produces action estimations at a rate of more than 5 frames per second.
Considering that the robot’s camera is pointing at users, a recommended distance between users and the robot, in this case, is 1.5–2 m. The actions have been split into two groups according to whether the face needs to be viewed or whether the body needs to be viewed, as it is shown in Table 1. The option of detecting the face and zooming on it is activated or not based on the requested action by the professionals. Whenever the face is required, we have implemented a traditional face detection model [74] with python and the OpenCV library [75]. Technically, our solution performs tracking of the user’s face over the captured 16 frames, creates a specific clip with the face only, and feeds it to the OAD model for the recognition of the action category. Figure 7 graphically shows how the implemented pipeline works. In the case of actions that are performed via body movements, the original video frames are prepared for feeding the OAD model.
Our OAD-3D-CNN is a heavy model containing a large number of parameters. The choice to implement this complex OAD model on the Jetson Xavier board slows down the GUI-integrated displays. The users were unable to see themselves with a mirror-like effect on the robot’s touch screen and were viewing the output of the camera with a delay. This aspect might provide a poor human–machine interaction experience. We, therefore, decided to implement a multi-threading application to solve this issue. Using different threads for the AI model and the visualizations in the GUI, we were able to provide a mirror-like effect to the users when they are performing the actions and seeing themselves on the screen of LOLA. All these improvements to the application were developed to make the human–robot interaction satisfactory.
Finally, we have recorded a video in which we show how the complete application described for the LOLA robot works https://youtu.be/JJUhtwISZaU (accessed on 19 April 2022).

4. Experiments

4.1. Experimental Validation

4.1.1. Sample of Final Users

The main objective of this work is to perform a technical validation of the robotic platform LOLA and of the application described. For doing so, a fundamental step consists in the selection of a sample of final users so that LOLA and the developed application can be experimentally validated.
This technical validation we propose includes four different patients. They were selected following the principles described in the protocol detailed in [76]. According to this study, all participants should be classified as having disabilities resulting from diseases or permanent health conditions (with the appropriate certificate, according to Spanish law [77]). Their functional independence level should be moderate-low, their functional skills-mobility level should be moderate-low, and they must be registered at one of the collaboration centers. Each of the patients we selected met these requisites. They ranged in age from 7 to 23 and all had physical limitations and cognitive deficits.
The first patient is a 7-year-old female diagnosed with idiopathic ataxia. She is level III in the Gross Motor Function Classification System Extended and Revised (GMFCS E&R), level III on Manual Ability Classification System (MACS), and level III in the Communication Function Classification System (CFCS). The nature of her condition required her to use a walker for short distances and a wheelchair for longer distances. She manipulates objects with difficulty and needs assistance in preparing or modifying her activities. She offers a challenging scenario where the AI models will be technically validated when the users need a walker and direct support from the healthcare professionals.
The second patient is a female 14-year-old who is diagnosed with Cerebral Palsy—Spastic Diparesis. She is level II on GMFCS E&R, level II on MACS, and level III on CFCS. Consequently, she can walk under most conditions and manipulate most objects, but with some reduction in quality or speed.
The third patient is a 15-year-old female who has Cerebral Palsy—Spastic Triparesis with the predominance on the right side. She is level II on GMFCS E&R, level III on MACS in the right upper limb, and level II on CFCS. She is capable of walking in most conditions, manipulating objects with difficulty, and needs assistance preparing and/or modifying activities.
A female 23-year-old, the fourth participant in this technical validation study, is diagnosed with metabolic syndrome due to GLUT 1 deficiency. She is level II in GMFCS E&R, level II in MACS, and level II in CFCS. Thus, she is capable of walking under most conditions and manipulating most objects, though with a corresponding reduction in quality or speed.
In light of the fact that the level of autonomy regarding movement varies from patient 1 to patient 4, our proposed SAR will be validated in all these different situations with a certain granularity.

4.1.2. Design of the Interventions

Before starting interventions with end users, health professionals in the therapy center were given a manual on how to operate and use the LOLA platform. A briefing session was held with the personnel to answer any questions they may have. In therapy, the patient’s physiotherapist is responsible for all evaluation processes involving the assistive robot. We aim that they can autonomously work with the platform and the applications designed. In the context of this technical validation, this is an important aspect, as it is crucial for us to identify what users actually think about the implemented application: is the proposed graphical user interface easy to use? Does the LOLA platform move and interact with the users in an effective way?
Once the health personnel were trained, we were able to start the interventions with the 4 users that made up the study sample.
Each patient is expected to take part in the therapy sessions with the robot for a maximum period of one month (four weeks) to study the results of the system. Only patients 1 and 2 could complete the 4 sessions. Patients 3 and 4 participated for two weeks and one week, respectively.
In each week of the trial, the patients had a 30 min session with LOLA during which they interacted with it and were monitored while performing ADLs. Almost five minutes were required to complete each action, divided into two parts that lasted around two minutes each. For this technical validation, the healthcare professionals selected 4 particular ADLs to be used during the interventions. They are the following: Blow Dry Hair; Brushing Teeth; Writing On Board; and Mopping Floor. All these 4 actions can be recognized by the AI OAD model integrated in LOLA.
Before each session of action monitoring with the robot, all the processes were explained to the patients. After selecting the patient user’s ID and the requested action by the professionals, one of the sample videos is shown to the patients with all sub-actions included. After that, the patients were asked to wait some moments for the AI-based model to be loaded and then start performing the action that they had watched in the videos. When the patients performed certain actions, such as blowing dry hair and brushing their teeth, it was not necessary for them to stand to accomplish the task. Instead, they sat in front of the robot, watching the action and trying to mimic what had been demonstrated. These actions were implemented with the aid of face detection, so the patients were instructed to remove their masks, which were mandatory due to the COVID-19 restrictions. In the remainder of the actions, it was necessary for the patients to stand up and perform the actions so that they maintained an appropriate distance from the robot. Figure 8 shows some patients developing some of the ADLs during the intervention sessions.
Since the main purpose of LOLA is to assist and monitor users, activating the OAD model will generate a report to inform professionals of the progress. The report contains information about the details provided by the AI model for action recognition. For every action recognition by the AI model, the report includes its starting and ending times, as well as an image of the user performing the action. The system also specifies whether the users performed the correct actions or not. All these details included in the report are a fundamental part that will allow the healthcare professionals to follow the protocol in [76]. Specifically, all the variables included in the reports will allow one to validate the level of matching between user and technology, the psychosocial impact of the assistive device on the user’s life, the level of satisfaction in different domains, and the level of independence in the performance for activities of daily living. After each session, these variables can be evaluated alongside the robot’s report to follow the patient’s progress.

4.2. Results

An overview of the progress and results for each of the four patients during this technical validation is presented in Table 2. The average accuracy of each action in each session for all patients are reported in the table. As we previously mentioned, each action was carried out in two separate attempts to better understand the precision of the OAD model with respect to patients with NDDs. Due to the fact that the OAD model is trained with videos with healthy subjects, depending on the physical limitations of the patients, certain actions were harder for the model to recognize.
With respect to the first patient, here are our conclusions. Due to her condition, she can not walk without a walker and is unable to control her movements. For the first two weeks, she assisted the sessions with her walker. Since she was using the walker, the system was not able to correctly recognize those actions which required an image of her entire body. This is a clear limitation of our OAD model, which would need an update to deal with such situations to attend to all possible patients. Overall, the accuracy of most of the actions she performed was 0% in the first two weeks. In terms of the actions which were implemented with face detection and zooming, the system was able to detect that she was brushing her teeth in one session despite all of her movement limitations. Another two sessions were conducted without her walker, but with a direct assistance of her physiotherapist in holding her. Under these circumstances, Table 2 shows how the accuracy of the AI model increases for all the actions. When the patient brushed her teeth and mopped the floor, LOLA recognized the activity with 100% accuracy. Due to her inability to write or hold the hair dryer in her hand, the system had difficulty detecting these two actions, which resulted in mainly 50% error rates. Figure 9 shows the average accuracy for all detected actions over all sessions for the first patient.
The second case of study is a patient that is capable of walking, although she experiences some difficulty with her balance and has mostly cognitive impairment. Her first encounter with LOLA was pleasant. The experience of seeing herself on the screen while performing an action in front of the camera of the robot appeared enjoyable to her. Table 2 shows the following numbers. There were no errors while the patient was brushing her teeth, and the Al-based module was 100% accurate in detecting this activity. During blow drying hair, there were sometimes conflicts between brushing teeth and blow drying hair because the patient was moving the mouth while performing other activities, and the results were only 50% accurate. When writing on the board, the patient, due to her balance difficulties, relied on the board and covered what she was writing. Several times she was advised to write something visible for the robot’s camera, and LOLA was able to detect her action with an accuracy of 50% over the course of four sessions. A perfect performance was achieved by the patient and the system detected the mopping of the floor with a high accuracy. An average accuracy for each action is shown in Figure 10 based on the detection of actions in all sessions.
The third patient in this evaluation trial had physical and cognitive disabilities, and as a consequence, she had difficulty maintaining an object in her hand. She could only participate in the study for two weeks, covering two complete sessions for the technical validation. She was able to brush her teeth despite all of her limitations, and the system detected this action category with a 100% of accuracy. It was difficult for her to hold a blow dryer and perform the corresponding action, but nevertheless, she was able to perform the action in less quality, and the system was able to recognize it with an average accuracy of 75%. Her difficulty in manipulating the whiteboard marker also made writing on the board a challenging process. Despite her best efforts, she was unable to hold the marker and write on the board at the same time, resulting in the system not recognizing that she had performed the action. Mopping the floor was not a particularly difficult task for her and the system consistently detected her actions correctly with 100% accuracy. Based on the detection of actions in all sessions, Figure 11 shows the average accuracy for each action.
Our last patient was able to grasp and hold objects in her hand but she primarily suffered from cognitive impairments. In addition to only participating in one session, she was not able to focus on her actions and did not show any interest in participating in other sessions. In Table 2, it can be seen that blow drying hair was a challenging task for her as she had short hair and the system was only capable of recognizing the action with 50% accuracy. In terms of brushing teeth and mopping the floor, she was able to perform them perfectly, and the system was able to correctly detect them 100% of the time (see Figure 12 for an example of user 4 performing an action). As she was tall, writing on the whiteboard proved to be the most complicated action, as we had to put a greater distance between the robot and the whiteboard to ensure that she was in the field of view of the camera. The patient was asked to write on the whiteboard in a larger size so that the robot could see the action, but she ignored the instructions. The system, therefore, could not correctly detect this action.
All of the results discussed above were extracted from detailed reports generated by the platform following each session. These reports are primarily intended to be distributed to health professionals at the conclusion of each session to allow them to monitor the progress of their patients during their treatment. With these reports, they will have access to all relevant information such as the action requested, the action detected by the system, the duration of performing the action, and an image of the patient while the action is being performed. From these reports, the health professionals can determine whether the patient understood the process of action correctly, whether he or she performed the action independently in a manner that the robot could detect, and how long the patient remained focused on the action. Additionally, it is also an opportunity for them to observe the progress of the patient’s independence and their ability to interact with new technologies as time progresses [76].
Our general observation is that all patients were interested in interacting with the robot and enjoyed seeing themselves on the screen. In most cases, LOLA is able to correctly detect all of the actions, but we note that there are some actions that need to be refined for the system to detect them with greater accuracy. These are Blow dry hair and Writing on Board. It is important to note that all these trials were performed on individuals with disabilities and impairments. Despite all these issues, LOLA received positive feedback from healthcare professionals and patients and demonstrated promising results during its first technical validation with real users. We can therefore conclude that LOLA and the developed application have successfully passed the technical validation proposed.

5. Conclusions

A study of the first technical validation of a low-cost robotic platform designed to monitor and assist people with neurodevelopmental disorders in their learning of daily living activities is presented in this paper. Our developed robot includes a user-friendly graphical interface, coupled with an online action detection module that helps users learn or reinforce daily living activities and lets health professionals observe their progress through reports of each session.
The evaluation trial included four users performing four daily routine activities in each session. As these users had varying degrees of physical and cognitive impairment, we were able to assess the functionality of our proposed platform in the context of real users with disabilities. In this trial period, we achieved acceptable results which demonstrated that our developed robot is capable of assisting and monitoring people with neurodevelopmental disorders in performing their daily living tasks.
Our future work consists of two different lines. The first line will be to deliver the assistive robot to a care center for people with NDDs for a long-term clinical study with a greater number of participants. In the second line, improvements will be made to the robot’s technical aspects. This will entail enriching the robot’s ability to recognize actions to accommodate a broader range of daily living actions appropriate for people with NDDs. A second improvement will be the incorporation of vocal commands into the graphical user interface for users with high cognitive skills and physical limitations to facilitate a more independent use of the platform.

Author Contributions

Conceptualization, all; methodology, R.J.L.-S., N.N., S.P.-d.-C. and F.J.A.-R.; software, N.N., I.F.-M., C.G.-Á. and R.J.L.-S.; validation, S.P.-d.-C., T.P.-G. and R.J.L.-S.; investigation, R.J.L.-S., F.J.A.-R., S.M.-B. and S.P.-d.-C.; writing—original draft preparation, all; writing—review and editing, all; supervision, R.J.L.-S.; project administration, R.J.L.-S.; funding acquisition, R.J.L.-S. and S.M.-B. All authors have read and agreed to the published version of the manuscript.

Funding

This research was funded by project AIRPLANE, with reference PID2019-104323RB-C31, of Spain’s Ministry of Science and Innovation.

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

All participants signed an informed consent form to take part in this study.

Data Availability Statement

In order to protect patients’ privacy, data cannot be published.

Acknowledgments

Authors wish to express their gratitude to all patients and their families who generously and patiently cooperated during the technical validation sessions with the robotic platform. Moreover, the authors would like to extend their sincere gratitude to the group of students who graciously volunteered to be part of the recording sessions for the videos.

Conflicts of Interest

The authors declare no conflict of interest.

Abbreviations

AIArtificial Intelligence
SARSocial Assistive Robot
CTComputerized Tomography
NDDNeurodevelopmental Disorder
ADLsActivities of Daily Living
ROSRobotic Operating System
OADOnline Action Detection
CNNConvolutional Neural Network
AMCLAdaptive Monte Carlo Localization
CFCSCommunication Function Classification System
MACSManual Ability Classification System
GMFCS E&RGross Motor Function Classification System Extended and Revised

References

  1. Feil-Seifer, D.; Matarić, M.J. Defining socially assistive robotics. In Proceedings of the 9th International Conference on Rehabilitation Robotics, ICORR 2005, Chicago, IL, USA, 28 June–1 July 2005; pp. 465–468. [Google Scholar]
  2. Clodic, A.; Alami, R.; Montreuil, V.; Li, S.; Wrede, B.; Swadzba, A. A study of interaction between dialog and decision for human-robot collaborative task achievement. In Proceedings of the RO-MAN 2007—The 16th IEEE International Symposium on Robot and Human Interactive Communication, Jeju, Korea, 26–29 August 2007; pp. 913–918. [Google Scholar]
  3. Santhanaraj, K.K.; Ramya, M.M.; Dinakaran, D. A survey of assistive robots and systems for elderly care. J. Enabling Technol. 2021, 15, 66–72. [Google Scholar] [CrossRef]
  4. Trost, M.J.; Chrysilla, G.; Gold, J.I.; Matarić, M.J. Socially-Assistive Robots Using Empathy to Reduce Pain and Distress during Peripheral IV Placement in Children. Pain Res. Manag. 2020, 2020, 7935215. [Google Scholar] [CrossRef] [PubMed]
  5. Bajones, M.; Fischinger, D.; Weiss, A.; de la Puente, P.; Wolf, D.; Vincze, M.; Körtner, T.; Weninger, M.; Papoutsakis, K.E.; Michel, D.; et al. Results of Field Trials with a Mobile Service Robot for Older Adults in 16 Private Households. ACM Trans. Hum.-Robot. Interact. 2020, 9, 1–27. [Google Scholar] [CrossRef]
  6. Kwoh, Y.S.; Hou, J.; Jonckheere, E.; Hayati, S. A robot with improved absolute positioning accuracy for CT guided stereotactic brain surgery. IEEE Trans. Biomed. Eng. 1988, 35, 153–160. [Google Scholar] [CrossRef] [PubMed]
  7. Costa, Â.; Martínez-Martín, E.; Cazorla, M.; Julián, V. PHAROS—PHysical Assistant RObot System. Sensors 2018, 18, 2633. [Google Scholar] [CrossRef]
  8. Pineau, J.; Montemerlo, M.; Pollack, M.E.; Roy, N.; Thrun, S. Towards robotic assistants in nursing homes: Challenges and results. Robot. Auton. Syst. 2003, 42, 271–281. [Google Scholar] [CrossRef]
  9. Lee, S.; Noh, H.; Lee, J.; Lee, K.; Lee, G.G.; Sagong, S.; Kim, M. On the effectiveness of Robot-Assisted Language Learning. ReCALL 2011, 23, 25–58. [Google Scholar] [CrossRef]
  10. Nasri, N.; Gomez-Donoso, F.; Orts, S.; Cazorla, M. Using Inferred Gestures from sEMG Signal to Teleoperate a Domestic Robot for the Disabled. In Proceedings of the International Work-Conference on Artificial Neural Networks, IWANN, Gran Canaria, Spain, 12–14 June 2019. [Google Scholar]
  11. Dawe, J.; Sutherland, C.J.; Barco, A.; Broadbent, E. Can social robots help children in healthcare contexts? A scoping review. BMJ Paediatr. Open 2019, 3, e000371. [Google Scholar] [CrossRef]
  12. Murfet, D. Estrategia Española Sobre Discapacidad. Available online: https://sid-inico.usal.es/idocs/F8/FDO26112/Estrategia2012_2020.pdf (accessed on 8 December 2018).
  13. Lemer, C. Annual Report of the Chief Medical Officer 2012: Our Children Deserve Better: Prevention Pays. 2013. Available online: https://assets.publishing.service.gov.uk/government/uploads/system/uploads/attachment_data/file/255237/2901304_CMO_complete_low_res_accessible.pdf (accessed on 1 October 2013).
  14. Arora, N.K.; Nair, M.K.C.; Gulati, S.; Deshmukh, V.; Mohapatra, A.; Mishra, D.; Patel, V.; Pandey, R.M.; Das, B.C.; Divan, G.; et al. Neurodevelopmental disorders in children aged 2–9 years: Population-based burden estimates across five regions in India. PLoS Med. 2018, 15, e1002615. [Google Scholar] [CrossRef]
  15. Dietrich, K.N.; Eskenazi, B.; Schantz, S.L.; Yolton, K.; Rauh, V.A.; Johnson, C.B.; Alkon, A.; Canfield, R.L.; Pessah, I.N.; Berman, R.F. Principles and Practices of Neurodevelopmental Assessment in Children: Lessons Learned from the Centers for Children’s Environmental Health and Disease Prevention Research. Environ. Health Perspect. 2005, 113, 1437–1446. [Google Scholar] [CrossRef] [Green Version]
  16. Cleaton, M.A.M.; Kirby, A. Why Do We Find it so Hard to Calculate the Burden of Neurodevelopmental Disorders. Arch. Med. 2018, 4, 10. [Google Scholar] [CrossRef]
  17. Pivetti, M.; Battista, S.D.; Agatolio, F.; Simaku, B.; Moro, M.; Menegatti, E. Educational Robotics for children with neurodevelopmental disorders: A systematic review. Heliyon 2020, 6, e05160. [Google Scholar] [CrossRef] [PubMed]
  18. López-Sastre, R.J.; Baptista-Ríos, M.; Acevedo-Rodríguez, F.J.; da Costa, S.P.; Maldonado-Bascón, S.; Lafuente-Arroyo, S. A Low-Cost Assistive Robot for Children with Neurodevelopmental Disorders to Aid in Daily Living Activities. Int. J. Environ. Res. Public Health 2021, 18, 3974. [Google Scholar] [CrossRef] [PubMed]
  19. Okita, S.Y. Self-Other’s Perspective Taking: The Use of Therapeutic Robot Companions as Social Agents for Reducing Pain and Anxiety in Pediatric Patients. Cyberpsychol. Behav. Soc. Netw. 2013, 16, 436–441. [Google Scholar] [CrossRef]
  20. Trost, M.J.; Ford, A.R.; Kysh, L.; Gold, J.I.; Matarić, M.J. Socially Assistive Robots for Helping Pediatric Distress and Pain: A Review of Current Evidence and Recommendations for Future Research and Practice. Clin. J. Pain 2019, 35, 451–458. [Google Scholar] [CrossRef]
  21. Soyama, R.; IShii, S.; Fukase, A. Selectable operating interfaces of the meal-assistance device My Spoon. In Lecture Notes in Control and Information Sciences; Springer: Berlin/Heidelberg, Germany, 2004; pp. 155–163. [Google Scholar]
  22. Dario, P.; Guglielmelli, E.; Laschi, C.; Teti, G. MOVAID: A personal robot in everyday life of disabled and elderly people. Technol. Disabil. 1999, 10, 77–93. [Google Scholar] [CrossRef]
  23. Esposito, R.; Fiorini, L.; Limosani, R.; Bonaccorsi, M.; Manzi, A.; Cavallo, F.; Dario, P. Supporting active and healthy aging with advanced robotics integrated in smart environment. In Optimizing Assistive Technologies for Aging Populations; IGI Global: Hershey, PA, USA, 2016; pp. 46–77. [Google Scholar] [CrossRef]
  24. Ribeiro, T.; Gonçalves, F.; Garcia, I.S.; Lopes, G.; Ribeiro, A.F. CHARMIE: A Collaborative Healthcare and Home Service and Assistant Robot for Elderly Care. Appl. Sci. 2021, 11, 7248. [Google Scholar] [CrossRef]
  25. Scoglio, A.A.J.; Reilly, E.D.; Gorman, J.A.; Drebing, C.E. Use of Social Robots in Mental Health and Well-Being Research: Systematic Review. J. Med. Internet Res. 2019, 21, e13322. [Google Scholar] [CrossRef]
  26. Krakovski, M.; Kumar, S.; Givati, S.; Bardea, M.; Zafrani, O.; Nimrod, G.; Bar-Haim, S.; Edan, Y. “Gymmy”: Designing and Testing a Robot for Physical and Cognitive Training of Older Adults. Appl. Sci. 2021, 11, 6431. [Google Scholar] [CrossRef]
  27. Bumin, G.; Günal, A. The effects of motor and cognitive impairments on daily living activities and quality of life in autistic children. Eur. J. Paediatr. Neurol. 2008, 12, 444–454. [Google Scholar] [CrossRef]
  28. Elbasan, B.; Kayihan, H.; Duzgun, I. Sensory integration and activities of daily living in children with developmental coordination disorder. Ital. J. Pediatr. 2012, 38, 1–7. [Google Scholar] [CrossRef] [PubMed]
  29. Blanco-Martínez, N.; Delgado-Lobete, L.; Montes-Montes, R.; Ruiz-Pérez, N.; Ruiz-Pérez, M.; Santos-Del-Riego, S. Participation in Everyday Activities of Children with and without Neurodevelopmental Disorders: A Cross-Sectional Study in Spain. Children 2020, 7, 157. [Google Scholar] [CrossRef] [PubMed]
  30. Gelsomini, M.; Degiorgi, M.; Garzotto, F.; Leonardi, G.; Penati, S.; Ramuzat, N.; Silvestri, J.; Clasadonte, F. Designing a Robot Companion for Children with Neuro-Developmental Disorders. In Proceedings of the 2017 Conference on Interaction Design and Children, Stanford, CA, USA, 27–30 June 2017. [Google Scholar]
  31. Cabibihan, J.J.; Javed, H.; Ang, M.H.; Aljunied, S.M. Why Robots? A Survey on the Roles and Benefits of Social Robots in the Therapy of Children with Autism. Int. J. Soc. Robot. 2013, 5, 593–618. [Google Scholar] [CrossRef]
  32. Giullian, N.; Ricks, D.J.; Atherton, J.A.; Colton, M.B.; Goodrich, M.A.; Brinton, B. Detailed requirements for robots in autism therapy. In Proceedings of the 2010 IEEE International Conference on Systems, Man and Cybernetics, Istanbul, Turkey, 10–13 October 2010; pp. 2595–2602. [Google Scholar]
  33. Goodrich, M.A.; Colton, M.B.; Brinton, B.; Fujiki, M.; Atherton, J.A.; Robinson, L.A.; Ricks, D.J.; Maxfield, M.H.; Acerson, A. Incorporating a robot into an autism therapy team. IEEE Intell. Syst. 2012, 27, 52–59. [Google Scholar] [CrossRef]
  34. Mazzei, D.; Billeci, L.; Armato, A.; Lazzeri, N.; Cisternino, A.; Pioggia, G.; Igliozzi, R.; Muratori, F.; Ahluwalia, A.; Rossi, D.D. The FACE of autism. In Proceedings of the 19th International Symposium in Robot and Human Interactive Communication, Viareggio, Italy, 13–15 September 2010; pp. 791–796. [Google Scholar]
  35. Ferrari, E.; Robins, B.; Dautenhahn, K. Therapeutic and educational objectives in robot assisted play for children with autism. In Proceedings of the RO-MAN 2009—The 18th IEEE International Symposium on Robot and Human Interactive Communication, Toyama, Japan, 27 September–2 October 2009; pp. 108–114. [Google Scholar]
  36. Robins, B.; Dautenhahn, K.; Dickerson, P. From Isolation to Communication: A Case Study Evaluation of Robot Assisted Play for Children with Autism with a Minimally Expressive Humanoid Robot. In Proceedings of the Second International Conferences on Advances in Computer-Human Interactions, Cancun, Mexico, 1–7 February 2009; pp. 205–211. [Google Scholar]
  37. Wood, L.J.; Dautenhahn, K.; Rainer, A.W.; Robins, B.; Lehmann, H.; Syrdal, D.S. Robot-Mediated Interviews—How Effective Is a Humanoid Robot as a Tool for Interviewing Young Children? PLoS ONE 2013, 8, e59448. [Google Scholar] [CrossRef]
  38. Wood, L.J.; Dautenhahn, K.; Robins, B.; Zaraki, A. Developing child-robot interaction scenarios with a humanoid robot to assist children with autism in developing visual perspective taking skills. In Proceedings of the 2017 26th IEEE International Symposium on Robot and Human Interactive Communication (RO-MAN), Lisbon, Portugal, 28 August–1 September 2017; pp. 1055–1060. [Google Scholar]
  39. Costa, S.; Resende, J.; Soares, F.O.; Ferreira, M.J.; dos Santos, C.P.; Moreira, F. Applications of simple robots to encourage social receptiveness of adolescents with autism. In Proceedings of the 2009 Annual International Conference of the IEEE Engineering in Medicine and Biology Society, Minneapolis, MN, USA, 3–6 September 2009; pp. 5072–5075. [Google Scholar]
  40. Costa, S.; Soares, F.O.; dos Santos, C.P.; Ferreira, M.J.; Moreira, M.F.; Vieira, A.P.P.; Cunha, F. An approach to promote social and communication behaviors in children with autism spectrum disorders: Robot based intervention. In Proceedings of the 2011 RO-MAN, Atlanta, GA, USA, 31 July–3 August 2011; pp. 101–106. [Google Scholar]
  41. Gillesen, J.; Barakova, E.I.; Huskens, B.; Feijs, L.M.G. From training to robot behavior: Towards custom scenarios for robotics in training programs for ASD. In Proceedings of the 2011 IEEE International Conference on Rehabilitation Robotics, Zurich, Switzerland, 29 June–1 July 2011; pp. 1–7. [Google Scholar]
  42. Bonarini, A.; Garzotto, F.; Gelsomini, M.; Romero, M.; Clasadonte, F.; Yilmaz, A.N.C. A huggable, mobile robot for developmental disorder interventions in a multi-modal interaction space. In Proceedings of the 2016 25th IEEE International Symposium on Robot and Human Interactive Communication (RO-MAN), New York, NY, USA, 26–31 August 2016; pp. 823–830. [Google Scholar]
  43. Céspedes, N.; Irfan, B.; Senft, E.; Cifuentes, C.A.; Gutiérrez, L.F.; Rincon-Roncancio, M.; Belpaeme, T.; Múnera, M.C. A Socially Assistive Robot for Long-Term Cardiac Rehabilitation in the Real World. Front. Neurorobot. 2021, 15, 633248. [Google Scholar] [CrossRef] [PubMed]
  44. Matarić, M.J.; Tapus, A.; Winstein, C.J.; Eriksson, J. Socially assistive robotics for stroke and mild TBI rehabilitation. Stud. Health Technol. Inform. 2009, 145, 249–262. [Google Scholar]
  45. Tapus, A.; Tapus, C.; Matarić, M.J. User—Robot personality matching and assistive robot behavior adaptation for post-stroke rehabilitation therapy. Intell. Serv. Robot. 2008, 1, 169–183. [Google Scholar] [CrossRef]
  46. Winkle, K.; Caleb-Solly, P.; Turton, A.J.; Bremner, P.A. Social Robots for Engagement in Rehabilitative Therapies: Design Implications from a Study with Therapists. In Proceedings of the 2018 13th ACM/IEEE International Conference on Human-Robot Interaction (HRI), Chicago, IL, USA, 5–8 March 2018; pp. 289–297. [Google Scholar]
  47. Bos, L. Personals attitudes towards robot assisted health care—A pilot study in 111 respondents. Med. Care Compunetics 2008, 5, 56. [Google Scholar]
  48. Swift-Spong, K.; Short, E.S.; Wade, E.; Matarić, M.J. Effects of comparative feedback from a Socially Assistive Robot on self-efficacy in post-stroke rehabilitation. In Proceedings of the 2015 IEEE International Conference on Rehabilitation Robotics (ICORR), Singapore, 11–14 August 2015; pp. 764–769. [Google Scholar]
  49. Martínez-Martín, E.; Costa, Â.; Cazorla, M. PHAROS 2.0—A PHysical Assistant RObot System Improved. Sensors 2019, 19, 4531. [Google Scholar] [CrossRef] [PubMed]
  50. Lamas, C.M.; Bellas, F.; Guijarro-Berdiñas, B. SARDAM: Service Assistant Robot for Daily Activity Monitoring. Multidiscip. Digit. Publ. Inst. Proc. 2020, 54, 3. [Google Scholar]
  51. Sunny, M.S.H.; Zarif, M.I.I.; Rulik, I.; Sanjuan, J.; Rahman, M.H.; Ahamed, S.I.; Wang, I.; Schultz, K.; Brahmi, B. Eye-gaze control of a wheelchair mounted 6DOF assistive robot for activities of daily living. J. Neuroeng. Rehabil. 2021, 18, 1–12. [Google Scholar] [CrossRef] [PubMed]
  52. Poppe, R. A survey on vision-based human action recognition. Image Vis. Comput. 2010, 28, 976–990. [Google Scholar] [CrossRef]
  53. Xu, Y.; Shen, Z.; Zhang, X.; Gao, Y.; Deng, S.; Wang, Y.; Fan, Y.; Chang, E.I.C. Learning multi-level features for sensor-based human action recognition. arXiv 2017, arXiv:1611.07143. [Google Scholar] [CrossRef]
  54. Chen, L.L.; Hoey, J.; Nugent, C.D.; Cook, D.J.; Yu, Z. Sensor-Based Activity Recognition. IEEE Trans. Syst. Man Cybern. Part C Appl. Rev. 2012, 42, 790–808. [Google Scholar] [CrossRef]
  55. Wang, J.; Chen, Y.; Hao, S.; Peng, X.; Hu, L. Deep Learning for Sensor-based Activity Recognition: A Survey. Pattern Recognit. Lett. 2019, 119, 3–11. [Google Scholar] [CrossRef]
  56. Gadde, P.; Kharrazi, H.; Patel, H.; MacDorman, K.F. Toward Monitoring and Increasing Exercise Adherence in Older Adults by Robotic Intervention: A Proof of Concept Study. J. Robot. 2011, 2011, 438514:1–438514:11. [Google Scholar] [CrossRef]
  57. Görer, B.; Salah, A.A.; Akin, H.L. An autonomous robotic exercise tutor for elderly people. Auton. Robot. 2017, 41, 657–678. [Google Scholar] [CrossRef]
  58. Carreira, J.; Zisserman, A. Quo Vadis, Action Recognition? A New Model and the Kinetics Dataset. In Proceedings of the 2017 IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Honolulu, UI, USA, 21–26 July 2017; pp. 4724–4733. [Google Scholar]
  59. Kay, W.; Carreira, J.; Simonyan, K.; Zhang, B.; Hillier, C.; Vijayanarasimhan, S.; Viola, F.; Green, T.; Back, T.; Natsev, A.; et al. The Kinetics Human Action Video Dataset. arXiv 2017, arXiv:1705.06950. [Google Scholar]
  60. Reddy, K.K.; Shah, M. Recognizing 50 human action categories of web videos. Mach. Vis. Appl. 2012, 24, 971–981. [Google Scholar] [CrossRef]
  61. Shou, Z.; Wang, D.; Chang, S.F. Temporal Action Localization in Untrimmed Videos via Multi-stage CNNs. In Proceedings of the 2016 IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Las Vegas, NV, USA, 27–30 June 2016; pp. 1049–1058. [Google Scholar]
  62. Shou, Z.; Chan, J.; Zareian, A.; Miyazawa, K.; Chang, S.F. CDC: Convolutional-De-Convolutional Networks for Precise Temporal Action Localization in Untrimmed Videos. In Proceedings of the 2017 IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Honolulu, HI, USA, 21–26 July 2017; pp. 1417–1426. [Google Scholar]
  63. Buch, S.; Escorcia, V.; Shen, C.; Ghanem, B.; Niebles, J.C. SST: Single-Stream Temporal Action Proposals. In Proceedings of the 2017 IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Honolulu, HI, USA, 21–26 July 2017; pp. 6373–6382. [Google Scholar]
  64. Wang, D.; Yuan, Y.; Wang, Q. Early Action Prediction With Generative Adversarial Networks. IEEE Access 2019, 7, 35795–35804. [Google Scholar] [CrossRef]
  65. Nguyen, M.H.; la Torre, F.D. Max-Margin Early Event Detectors. Int. J. Comput. Vis. 2012, 107, 191–202. [Google Scholar]
  66. Lin, J.; Gan, C.; Han, S. TSM: Temporal Shift Module for Efficient Video Understanding. In Proceedings of the 2019 IEEE/CVF International Conference on Computer Vision (ICCV), Seoul, Korea, 27 October–2 November 2019; pp. 7082–7092. [Google Scholar]
  67. Gao, J.; Yang, Z.; Nevatia, R. RED: Reinforced Encoder-Decoder Networks for Action Anticipation. arXiv 2017, arXiv:1707.04818. [Google Scholar]
  68. Feichtenhofer, C.; Fan, H.; Malik, J.; He, K. SlowFast Networks for Video Recognition. In Proceedings of the 2019 IEEE/CVF International Conference on Computer Vision (ICCV), Seoul, Korea, 27 October–2 November 2019; pp. 6201–6210. [Google Scholar]
  69. Wang, X.; Zhang, S.; Qing, Z.; Shao, Y.; Zuo, Z.; Gao, C.; Sang, N. OadTR: Online Action Detection with Transformers. In Proceedings of the 2021 IEEE/CVF International Conference on Computer Vision (ICCV), Montreal, BC, Canada, 11–17 October 2021; pp. 7545–7555. [Google Scholar]
  70. Geest, R.D.; Gavves, E.; Ghodrati, A.; Li, Z.; Snoek, C.G.M.; Tuytelaars, T. Online Action Detection. In ECCV; Springer: Berlin/Heidelberg, Germany, 2016. [Google Scholar]
  71. Quigley, M.; Gerkey, B.; Conley, K.; Faust, J.; Foote, T.; Leibs, J.; Berger, E.; Wheeler, R.; Ng, A. ROS: An open-source Robot Operating System. In Proceedings of the IEEE International Conference on Robotics and Automation (ICRA) Workshop on Open Source Robotics, Kobe, Japan, 12–17 May 2009. [Google Scholar]
  72. Baptista-Ríos, M.; López-Sastre, R.J.; Heilbron, F.C.; Van Gemert, J.C.; Acevedo-Rodríguez, F.J.; Maldonado-Bascón, S. Rethinking online action detection in untrimmed videos: A novel online evaluation protocol. IEEE Access 2019, 8, 5139–5146. [Google Scholar] [CrossRef]
  73. Soomro, K.; Zamir, A.R.; Shah, M. UCF101: A Dataset of 101 Human Actions Classes From Videos in The Wild. arXiv 2012, arXiv:1212.0402. [Google Scholar]
  74. Viola, P.; Jones, M.J. Robust Real-Time Face Detection. Int. J. Comput. Vision 2004, 57, 137–154. [Google Scholar] [CrossRef]
  75. Bradski, G. The OpenCV Library. Dr. Dobb’s J. Softw. Tools 2000, 25, 120–123. [Google Scholar]
  76. García, T.P.; González, B.G.; Nieto-Riveiro, L.; Domínguez, N.C.; Maldonado-Bascón, S.; López-Sastre, R.J.; DaCosta, S.P.; González-Gómez, I.; Molina-Cantero, A.J.; Loureiro, J.P. Assessment and counseling to get the best efficiency and effectiveness of the assistive technology (MATCH): Study protocol. PLoS ONE 2022, 17, e0265466. [Google Scholar]
  77. del Estado, B.O. Real Decreto 1971/1999, de 23 de Diciembre, de Procedimiento Para el Reconocimiento, Declaración y Calificación del Grado de Discapacidad. Available online: https://bit.ly/3AY2ZAH (accessed on 3 July 2020).
Figure 1. An overview of the action monitoring process with LOLA and users with NDDs. The patient should be placed in front of LOLA, the AI-based monitoring action model should be activated, and a report with information regarding the patient’s performance during the session will be generated and sent to the health professional in charge.
Figure 1. An overview of the action monitoring process with LOLA and users with NDDs. The patient should be placed in front of LOLA, the AI-based monitoring action model should be activated, and a report with information regarding the patient’s performance during the session will be generated and sent to the health professional in charge.
Applsci 12 09566 g001
Figure 2. Graphical comparison between version 1 (a) and 2 (b) of LOLA. Novel sensors, design improvements, and a novel NVIDIA Jetson Xavier board have been integrated into LOLA2 (b).
Figure 2. Graphical comparison between version 1 (a) and 2 (b) of LOLA. Novel sensors, design improvements, and a novel NVIDIA Jetson Xavier board have been integrated into LOLA2 (b).
Applsci 12 09566 g002
Figure 3. ROS Melodic architecture embedded in LOLA2 platform. Blue nodes represent distributed software ROS packages and yellow nodes are those ROS nodes entirely developed by us to integrate ROS in our platform.
Figure 3. ROS Melodic architecture embedded in LOLA2 platform. Blue nodes represent distributed software ROS packages and yellow nodes are those ROS nodes entirely developed by us to integrate ROS in our platform.
Applsci 12 09566 g003
Figure 4. Different options in the developed graphical user interface for action monitoring. (a) gives options to specify the user ID and the actions to be monitored. There are options to choose from for observing and monitoring the action in (b). (a) Main menu. (b) Action monitoring options.
Figure 4. Different options in the developed graphical user interface for action monitoring. (a) gives options to specify the user ID and the actions to be monitored. There are options to choose from for observing and monitoring the action in (b). (a) Main menu. (b) Action monitoring options.
Applsci 12 09566 g004
Figure 5. For this work, we have built a database with videos containing actors that execute, step by step, a set of actions of daily life that we want to work with patients. These images contain some captures where we can observe two actors performing two different actions. (a) The actor is showing the action of writing on the board. (b) The actor is showing the action of mopping the floor.
Figure 5. For this work, we have built a database with videos containing actors that execute, step by step, a set of actions of daily life that we want to work with patients. These images contain some captures where we can observe two actors performing two different actions. (a) The actor is showing the action of writing on the board. (b) The actor is showing the action of mopping the floor.
Applsci 12 09566 g005
Figure 6. This figure illustrates the OAD-3D-CNN model that is used to monitor patients’ daily living activities. The AI model is fed with the video stream captured by the camera installed on the LOLA platform. In particular, clips of 16 contiguous video frames are passed to the model. Eight 3D convolutions are included in the deep learning architecture, followed by five max-pooling layers and three fully connected layers. A final softmax layer is responsible for producing the classification of the action for every video clip.
Figure 6. This figure illustrates the OAD-3D-CNN model that is used to monitor patients’ daily living activities. The AI model is fed with the video stream captured by the camera installed on the LOLA platform. In particular, clips of 16 contiguous video frames are passed to the model. Eight 3D convolutions are included in the deep learning architecture, followed by five max-pooling layers and three fully connected layers. A final softmax layer is responsible for producing the classification of the action for every video clip.
Applsci 12 09566 g006
Figure 7. Upon detecting the face of the user, the model zooms in on it to create the video clip that is passed to the OAD model.
Figure 7. Upon detecting the face of the user, the model zooms in on it to create the video clip that is passed to the OAD model.
Applsci 12 09566 g007
Figure 8. This figure includes images of several participants during the interventions performing the four study actions: blow-drying hair (first row); brushing teeth (second row); writing on a blackboard (third row); and mopping (fourth row).
Figure 8. This figure includes images of several participants during the interventions performing the four study actions: blow-drying hair (first row); brushing teeth (second row); writing on a blackboard (third row); and mopping (fourth row).
Applsci 12 09566 g008
Figure 9. Results for patient UAH-1 based on LOLA’s actions monitoring system.
Figure 9. Results for patient UAH-1 based on LOLA’s actions monitoring system.
Applsci 12 09566 g009
Figure 10. Results for patient UAH-2 based on LOLA’s actions monitoring system.
Figure 10. Results for patient UAH-2 based on LOLA’s actions monitoring system.
Applsci 12 09566 g010
Figure 11. Results for patient UAH-3 based on LOLA’s actions monitoring system.
Figure 11. Results for patient UAH-3 based on LOLA’s actions monitoring system.
Applsci 12 09566 g011
Figure 12. Mopping floor action performed by the patient 4.
Figure 12. Mopping floor action performed by the patient 4.
Applsci 12 09566 g012
Table 1. The four categories of actions and their status regarding zooming.
Table 1. The four categories of actions and their status regarding zooming.
Action CategoryWith Zoom/ Without Zoom
Blow Dry HairWith zoom
Brushing TeethWith zoom
Writing On BoardWithout zoom
Mopping FloorWithout zoom
Table 2. This table details the results of the entire process of technical validation of the platform. We report the average accuracy of the AI model when recognizing the actions during the different sessions. Our OAD model has detected the actions of brushing teeth and mopping the floor with the highest accuracy during the sessions. Considering the patients’ physical limitations, blow drying hair, and writing on board were challenging activities for them to undertake.
Table 2. This table details the results of the entire process of technical validation of the platform. We report the average accuracy of the AI model when recognizing the actions during the different sessions. Our OAD model has detected the actions of brushing teeth and mopping the floor with the highest accuracy during the sessions. Considering the patients’ physical limitations, blow drying hair, and writing on board were challenging activities for them to undertake.
User IDActionsSession 1Session 2Session 3Session 4
UAH-1Blow Dry Hair0%0%50%50%
Brushing Teeth100%0%100%100%
Writing On Board0%0%0%50%
Mopping Floor0%0%100%100%
UAH-2Blow Dry Hair50%50%50%50%
Brushing Teeth100%100%100%100%
Writing On Board100%0%0%100%
Mopping Floor100%100%100%100%
UAH-3Blow Dry Hair100%50%
Brushing Teeth100%100%
Writing On Board0%0%
Mopping Floor100%100%
UAH-4Blow Dry Hair50%
Brushing Teeth100%
Writing On Board0%
Mopping Floor100%
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Share and Cite

MDPI and ACS Style

Nasri, N.; López-Sastre, R.J.; Pacheco-da-Costa, S.; Fernández-Munilla, I.; Gutiérrez-Álvarez, C.; Pousada-García, T.; Acevedo-Rodríguez, F.J.; Maldonado-Bascón, S. Assistive Robot with an AI-Based Application for the Reinforcement of Activities of Daily Living: Technical Validation with Users Affected by Neurodevelopmental Disorders. Appl. Sci. 2022, 12, 9566. https://doi.org/10.3390/app12199566

AMA Style

Nasri N, López-Sastre RJ, Pacheco-da-Costa S, Fernández-Munilla I, Gutiérrez-Álvarez C, Pousada-García T, Acevedo-Rodríguez FJ, Maldonado-Bascón S. Assistive Robot with an AI-Based Application for the Reinforcement of Activities of Daily Living: Technical Validation with Users Affected by Neurodevelopmental Disorders. Applied Sciences. 2022; 12(19):9566. https://doi.org/10.3390/app12199566

Chicago/Turabian Style

Nasri, Nadia, Roberto J. López-Sastre, Soraya Pacheco-da-Costa, Iván Fernández-Munilla, Carlos Gutiérrez-Álvarez, Thais Pousada-García, Francisco Javier Acevedo-Rodríguez, and Saturnino Maldonado-Bascón. 2022. "Assistive Robot with an AI-Based Application for the Reinforcement of Activities of Daily Living: Technical Validation with Users Affected by Neurodevelopmental Disorders" Applied Sciences 12, no. 19: 9566. https://doi.org/10.3390/app12199566

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop