Next Article in Journal
Swarm Crawler Robots Using Lévy Flight for Targets Exploration in Large Environments
Previous Article in Journal
Toward Avoiding Misalignment: Dimensional Synthesis of Task-Oriented Upper-Limb Hybrid Exoskeleton
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Review

A Survey on Recent Advances in Social Robotics

College of Engineering and Technology, American University of the Middle East, Egaila 54200, Kuwait
*
Author to whom correspondence should be addressed.
Robotics 2022, 11(4), 75; https://doi.org/10.3390/robotics11040075
Submission received: 13 June 2022 / Revised: 8 July 2022 / Accepted: 8 July 2022 / Published: 18 July 2022
(This article belongs to the Section Humanoid and Human Robotics)

Abstract

:
Over decades, social robotics has evolved as a concept that presently covers different areas of application, and interacts with different domains in technology, education, medicine and others. Today, it is possible to envision social robots in tasks that were not expected years ago, and that is not only due to the evolution of social robots, but also to the evolution of the vision humans have for them. This survey addresses recent advances in social robotics from different perspectives. Different contexts and areas of application of social robots are addressed, as well as modalities of interaction with humans. Different robotic platforms used in social contexts are shown and discussed. Relationships of social robotics with advances in other technological areas are surveyed, and methods and metrics used for the human evaluation of the interaction with robots are presented. The future of social robotics is also envisioned based on surveyed works and from different points of view.

1. Introduction

A social robot can be defined as a robot designed to operate and interact with humans, in contexts similar to human–human interaction contexts. Knowing that several types of machines and robots can interact with humans in different ways and modalities, a question emerges on what differentiates a social robot from a robot that is not social. In the recent literature, several definitions of social robots can be found, showing differences in visions that authors have of what can be called a social robot. This also arises because there is no agreement on the definition of robots, regardless of their social qualification. The question of defining social robots has been addressed in [1] where scientific and popular definitions of social robots have been examined. The work presented in [1] shows the following points among other findings:
  • Not all articles using the “social robot” label defined what the authors meant by it.
  • In highly pertinent articles, the majority acknowledges the lack of a generally accepted and accurate definition and the difficulty of defining a social robot.
  • There has been an increase in the number of times the label “social robot” has been mentioned since the late 1990s.
  • Among the definitions made in scientific articles, the following properties are associated with social robots: autonomy, ability to act in a socially appropriate manner, and in meaningful social interactions, communication, intelligence, operation according to established social and cultural norms, the ability to sense the presence of humans and to engage in physical acknowledgment, the ability to use gestures, to express and/or perceive emotions and to engage in a conversation [2,3,4,5,6].
  • Different alternative future-oriented definitions have been made, mentioning a merge between biological and technological elements in social robots, and social robots representing applications of technology meant to solve social problems that are non-technical in nature, for pressing issues in contemporary society [7,8].
Based on the previous definitions and findings obtained from [1], and on other work done in the field of social robotics, it can be stated that social robots can do services and use different means of communication while interacting with humans. It is possible to envision social robots in different types of areas, such as industrial, educational, commercial, and homes, being part of the daily life of humans [9,10]. Social robotics is becoming a major area of research, with the development of new techniques allowing computer systems to process signals and extract information efficiently, such as sound, images, and natural language. Furthermore, social robots can convey information to humans via different means such as speech, gestures, and facial expressions [11]. Advances in machine learning and pattern recognition are allowing robots to acquire realistic social behaviors involving behaviors in speech and gestures [12,13,14,15,16,17]. This enabled robots to be active conversation parties either in text-based or voice-based conversations. It is possible presently to envision robots in applications that would not have been expected two decades ago, such as voice messaging [18], ecology awareness [19], health data acquisition in hospitals [20] and vocabulary enhancement games [21].
This paper addresses recent advances in social robotics from different points of view. This domain has witnessed a rise in popularity allowing it to expand to various areas of application. Research work has been done with robotic platforms that are either commercially available, programmable, and re-configurable, or designed and manufactured for specific persons and specific usages. Such platforms are equipped with sensing and actuation capacities allowing perceptions of the environment, including humans and other robots, and analysis of the perceptions and actions used in the interaction such as emotion display, gestures, and speech. All this would not have been possible without the advances taking place in several fields, ranging from engineering fields as mentioned previously to psychological fields where the assessment of the perception that humans have of robots plays an important role in their designs and usage. Different works related to all the previously mentioned points are reviewed and cited. The contributions of the proposed survey can be summarized by the following points, each supported by different previous works and studies:
  • Demonstrating the extent to which social robotics can be involved in humans’ life.
  • Proving different capacities allowing a robot to interact in a social manner.
  • Emphasizing the impacts of technology and research on social robots.
  • Addressing the human side of the human-robot interaction.
  • Providing a multi-viewpoint image of the future of social robots.
This survey does not claim to be exhaustive but covers enough references corresponding to each point, allowing a reader who is not specialized in all the mentioned fields, to have a view of the recent work being done in each one. As the aim is to show trends in social robots across different disciplines, works have been included in the survey based on their relevance and date. Robotics journals and robotics conference proceedings addressing social robotics have been relied on to access publications. Mainly, publications from the International Journal of Social Robotics from the last 4 years were accessed if relevant to the survey alongside other sources. A part of the publications was obtained by searching for specific keywords such as “social robot” and “assistive robot” while others were obtained by navigating through the sources. The authors conducting the survey and selecting the articles are researchers in robotics, PhD holders working in a robotics research center with activities and previous publications in the field of social robotics. Figure 1 shows the different points addressed in the survey.
This paper is organized as follows. Section 2.1 addresses recent advances in social robotics, emphasizing the usages of social robots in different applications. Section 3 addresses modalities of interaction between humans and social robots, mainly showing vision, conversations and gestures. Section 4 shows characteristics of different platforms used in social robotics contexts. Section 5 tackles the advances in different domains, impacting social robotics, namely artificial intelligence, manufacturing technologies, sensing and actuation technologies, and operating systems. In Section 6, different metrics and methods of the measurement of the acceptability of social robot interactions are presented. Later, a link between the past, present, and future applications of social robotics is made in Section 7. Finally, Section 8 concludes the paper.

2. Domains of Application

Research works done in social robotics reveal the wide array of domains and applications they are having. Social robots are being proposed in telepresence, medicine, education, entertainment, assistance, and other domains. Benefiting from their information acquisition and processing, and actuation capacities, social robots are conceived to either replace or assist humans in daily social interaction contexts. The following subsections show examples of work where social robots have been used in several domains. Table 1 summarizes several the surveyed works shown in the following subsections. It shows the different applications, the end users, and the robots used in each study. These studies, along with others will be explained further in the next subsections and the robots used will be addressed in more detail in Section 4.

2.1. Telepresence

In telepresence applications, a user can rely on a robotic platform to ensure a certain extent of social interaction with other persons while being at a distant location from them. Different technologies have been implemented on social robots to ensure a realistic interaction from both user and interaction partner sides. Telepresence robots require features such as autonomy, controllability, maneuverability, and stability to ensure safe interaction with humans [46]. For instance, in [15], a deep-learning approach has been proposed for a telepresence robot to learn by demonstrating how to maintain an appropriate position and orientation within a group of people, and how to follow moving interaction targets. In this context, the robot has been qualified as semi-autonomous as its pilot still had control over certain high-level tasks. A similar platform has been used in [22] for the interaction between users in and outside an elderly day center. In [46], a robotic telepresence system design and a control approach for social interaction have been presented. The robot has been equipped with capabilities of vision, hearing, speaking, and moving, all controlled remotely by a user. In [23], a study has been shown where a Double telepresence robot was installed in rooms of care homes, for the purpose of allowing old persons to communicate with their family members. Despite some technical difficulties, the experience of using this system was positively evaluated by persons involved in the study. Figure 2 shows a Double telepresence robot [47]. A telepresence robotic system for people with motor disabilities has been proposed in [24]. Eye gaze was used as an input to the system as an eye-tracking mechanism involving a virtual reality head-mounted display was intended to provide driving commands to the robotic platform.

2.2. Education

Robots have also been involved in education where they had applications in language teaching, teaching assistance, writing, and vocabulary enhancement for example [48,49,50]. Indeed, they can facilitate learning and improve the educational performance of students, adding social interaction to the learning context in certain cases [51]. In this context, the attitudes of teachers and students towards robots are important to evaluate and the study shown in [52] tackled the negative attitude toward robot scale (NARS) that was developed in [53] to measure general human attitudes towards robots. The study pointed to the importance of knowing the attitudes of teachers towards robots used in classes. Additionally, the study made in [27] studied the acceptability of robots by teaching second language words to adults with the robot and human tutors. A negative attitude toward robots was shown to have a possible negative impact on the ability of individuals to learn vocabulary. Second language tutoring was addressed in [30,54] with children and a social robot with a tablet. Children were not shown to be able to learn more words when learning from a robot and a tablet than from a tablet without a robot. Additionally, iconic gestures from the robot were not shown to help children learn more words. A closely related topic in second language learning was addressed in [26] where different interaction styles of human moderators in language café style conversations were developed for a robot in the role of a host in spoken conversation practice. These styles were rated differently by human participants, due to different factors, not only due to the robot interaction but also due to the participants themselves. This allowed us to state that interaction can be improved, taking these different factors into account. In [21], the humanoid robot Pepper has been used in vocabulary enhancement in children in a game scenario intended to improve their abilities to manipulate and learn words. The capabilities of the Pepper robot such as connecting to the Internet, and acquiring and emitting sound signals have been exploited to accomplish a proper interaction with children in this context. A robotic platform was used to support professors in mathematics classes in [29]. The Nao robot was used for giving theoretical explanations, as well as the instructions and evaluation of the activities made during class. It was programmed before each session to move the most humanly possible to generate accurate visual gestures. Moreover, the vision and sound generation capacities of Nao were exploited for the interaction. In [28], a robotic platform was used for transferring students’ opinions or questions to the lecturer. The desktop humanoid robot was collaboratively controlled and provided a messaging interface where messages consisted of questions or opinions of students to be uttered by the robot. This approach was shown to improve the participation of students during lectures. Another paradigm of learning in children is learning by teaching , which engages a student in the act of teaching another. In [25], a humanoid robot was used as a handwriting partner with simulated handwriting mistakes intentionally made, being typical of children learning to handwrite. Children taught the robot how to write, learning through their teaching.

2.3. Care and Assistance

Another domain of application where social robots have emerged is assistance in health and daily care services. In [55,56], the increasing demand for elderly care and the Socially assistive robot (SAR) technology roles in this field are highlighted. It was proposed in [57] that socially assistive robots could support people with health conditions in social interactions, with the aim of improving their health and well-being. For elderly people, social robots can be embedded into their homes or care facilities and play different roles. In [58], the field of in-house assistance for older adults was reviewed. It is suggested that the deployments that have been made for robots for in-house assistance are mostly prototypes and that robots have yet to succeed as personal assistance. It is reported that in healthcare, robots have a variety of applications, and can be classified into three categories: surgical, rehabilitation, and social robots. Furthermore, social robots in this context were divided into service and companion categories, dedicated for assistance in one or more tasks, or for user companionship, respectively. In the last context, the appearance of the robot and the extent to which it resembles a human being were reported to affect its acceptability by end users [59,60]. In this context, the study made in [61] focused on the perception of care robots among end users. It addressed the end users’ understandings, assumptions, and expectations of care robots. The study covered different groups of stakeholders such as line managers, frontline care staff, older people, and students training to become careers. Congruent frames between the groups regarding the understanding of the nature of care robots were found. But incongruent frames of the nature of care robots were found between the different groups when addressing the question of sketching the ideal robot. The study identified adequate training, usability, and finances among potential criteria for the successful use of care robots. Perceptions of stakeholders of socially assistive robots were also addressed in [62] where a study on a robot known as Stevie was conducted with older adults and professional care workers in a retirement community. Focus groups were formed where the robot was teleoperated and observations and discussions were made. Staff and residents had different views and concerns about the robot but they both saw its potential utility and suggested many possible use-cases. Older people care has also been addressed in [31] where a user-centered soft and bedside communication robot was developed and evaluated. A collaborative and iterative development process was adopted, involving different stakeholders. The resulting system helped in the improvement of the mood and behavior of participants, as well as in obtaining a positive engagement from their side. Another aspect of assistance was addressed in [63] where an approach to develop a robot for the assistance of workers with intellectual and developmental disabilities was presented.
In the context of care and assistance for older persons, Ambient Assistive Living (AAL) has been defined, as reported in [64] from [65] as “the use of information and communication technologies (ICT) in a person’s daily living and working environment to enable them to stay active longer, remain socially connected and live independently into old age”. AAL covers tasks such as observation and the detection of events such as falls but goes beyond that to interact with users [64]. In this context, the integration of socially assistive robots into AAL has been shown in [32] to succeed in long-term support to older adults. The authors reported that the robot incentivized the usage of the system but slightly lowered its overall acceptability. This work used the Giraff-X (a version of the Giraff teleoperated robot [66,67]) as an embodiment for a virtual caregiver at the elder’s house.

2.4. Medicine

Another field of application of social robots in health care. Different usages can be found for social robots in this context, ranging from the assistance of nurses to rehabilitation [10]. The interventions of socially assistive robots in supporting mental health in children have been reviewed in [68]. It was found that the contexts of the interventions affect their impacts. Indeed, the place, the culture of the user, and the robot used are all factors contributing to the outcomes of the intervention. The study showed different robotic platforms used in this context and reported consistent positive outcomes such as relief of distress and increase of positive effects regardless of the robot used. However, disparities have been seen between outcome measures, robots used, and study quality. The usage of a social robot was shown to have possible benefits in attention improvement for children with cognitive impairement [33]. In this context, a child-robot interaction was designed and implemented, consisting of several modules during which the children played short games with the robot, taking the capacities of the robot into account. Additionally, dementia was addressed in [34,35]. The research work in [35] focused on designing robots for dementia caregiving, addressing the needs of both the caregiver and the caregiver. This covered the intended purpose and functions of the robots as “robots for joy”, “robots for repetition” and “robots for wellness” were designed. Additionally, different morphologies and modalities for interacting with the robots such as voice interaction where the voices of people that whom caregivers were familiar were discussed. Moreover, different roles were assigned to robots, such as “the bad guy”, “the facilitator” and “the counselor”. The Softbank robot, Pepper, was used in [34] for encouraging exercise in dementia patients. Specifically, the study used simple dance moves as the exercise modality due to the engagement and repetitiveness of dancing. A heart-rate monitor was used for sending feedback to the robot to adjust the intensity of the exercise. Preliminary results were reported to be promising. Pepper was also used in [43] in a system developed for the audiometry tests and rehabilitation of children with hearing disabilities. Positive and negative emotions of children were shown to be better distinguished when they interact with the robot than in setups without the robot. Social anxiety disorder, a condition pushing people to fear social situations, was addressed in [69] where an overview of certain usages of social robots in clinical interventions was made. This work proposed that social robots can be used to complement the work of clinicians. Additionally, Pepper’s usage in health data acquisition was explored in [20] to act as a nurse assistant and reduce data registration workloads on nurses.A multimodal dialogue involving verbal, gesture, and screen display aspects were designed to facilitate the interaction between the robot and the patient. Evaluations made by patients and nurses showed the possible acceptability of the robot. Another usage of robots was shown in [56], not directly as a health assistant, but as an assistant in enhancing the skills of nursing students, specifically in patient transfer, where a patient is moved from a bed to a wheelchair, and vice-versa. The robot in this work simulated a patient to be transferred while measuring different motion parameters during the transfer and assessing whether the transfer was made accurately by the nursing student or not. Indeed, proper body mechanics need to be used in this task, indispensable to the patient’s daily life, especially with elderly patients affected by weaknesses in their limbs. Results showed that the robot can be a good substitute for an actual patient when performing this task. Rehabilitation was addressed in [70] and specifically, the trust of users interacting with a rehabilitation robot. Exercises were performed at different velocities of robot motion and data on participants’ heart rates and perception of safety were collected. Notably, the perception of safety was negatively affected by increasing velocity and exercise extent. Another application of socially assistive robots has been shown in [71] where robot prototypes that assist persons in sorting their medications have been developed and tested, to organize the days and times pills should be taken.

2.5. Autism Spectrum Disorders

Autism Spectrum Disorders (ASD) cause abnormalities of impaired development in social communication and interaction, and in restricted and repetitive patterns of behavior, interests, or activities. Individuals with ASD have difficulties interacting and communicating with others [72,73]. This is due to their inability to understand social cues and the behaviors and feelings of others. Research on information communication technology (ICT) has been active in the domain of the education of people with autism [74]. In the same context. different works in socially assistive robotics have tackled the treatment of individuals with ASD, increasingly since 2000, with different directions of research [75]. Such works target the improvement of the social functioning of children with ASD [39]. One of these directions is imitation, as a deficit in imitation is a symptom of ASD. The study made in [38] compared the body gesture imitation performance of participants with ASD and typically developing subjects. It also compared this performance in robot-child and adult-child imitation tasks. In the presented experimental setup, both participants with typical development and with ASD performed better in adult-child mode than in robot-child mode. Additionally, participants with typical development showed better performance than participants with ASD. Additionally, in [36], children with typical development and with ASD performed better with a therapist in a joint attention elicitation task. Joint attention is related to social relationships and has been defined in [76] as a triadic interaction between two agents focusing on a single object. Additionally, in [36], children with typical development performed better than children with ASD with the robot. Another direction of research, where the usage of robots proved efficient in the enhancement of social skills of children with ASD was shown in [39]. In this study, a robot was used for an intervention in a social skill training program that consisted of different phases with and without the robot. Results showed that the intervention of the robot improved the social motivation and skills of children with ASD, measured by eye contact frequency and duration, and verbal initiation frequency. This improvement lasted even after the robot was withdrawn from the program. A similar result was obtained in [37] where children with ASD participated in sessions divided into sessions with the Kaspar robot and sessions with a human teacher. The usage of the robot increased interactions among children such as non-verbal imitation, touching, and attention duration. The Kaspar robot’s evolution is shown in [77]. Initially, the Kaspar robot was developed in 2005 for research in human-robot interaction, then it was adopted for investigation as a therapeutic device for children with ASD which has been its primary application and target of improvement since then. The evolution of this robot benefited from the knowledge improvement in the therapy of children with ASD and shows hardware modifications and improvement of sensory devices aiming to improve its usability and autonomy for child-robot interaction.

2.6. Other Applications of Children Companionship

As reported, social robots have been used with children in applications such as education and ASD. Nevertheless, other applications of social robots used with children can be reported from other domains such as entertainment, awareness-raising and cognition, perception, and behavioral studies. In [19], a social robot was used in a game intended to make children more aware of the importance of waste recycling. The game had 2 players: the child and the Softbank robot Pepper, and a human judge. The study reported promising results in changing children’s attitudes toward recycling and showed a positive evaluation of Pepper by children. Children’s gaze aversion was addressed in [41] where gaze aversion was reported from other sources to refer to human being reflexive redirection of the gaze away from a potentially distracting visual stimulus while solving a mentally demanding task, facilitating thinking. The study evaluated the influence of the interaction with a humanoid robot on children’s gaze aversion. Results showed that gaze aversion rates increased when children interacted with other humans, or with robots that were told to be human-controlled, in contrast with their interactions with robots controlled by computers. These findings were linked to the perception children make of minds in their interaction agents. Child-robot interaction was also explored in [42] to develop a method for emotion recognition relying on functional infrared imaging. It allowed for the assessment of the level of child engagement while interacting with an artificial agent and the presented work was said to constitute a step toward a more natural interaction between a child and an artificial agent, based on physiological signals. In [40], the focus was on implementing the behavior of a robot storyteller using an analysis of human storytellers. The effects of implementing emotions in the storytelling, contextual storyteller head movements, and voice fitting to characters were evaluated. Positive impacts on listeners were found in emotional robot and voice-acting robot storytellers. Contextual head movements, on the other hand, did not have effects on the perception users make of the robot storyteller. The usage of robots in childcare was addressed in [78] where the requirements, needs, and attitudes of working parents toward childcare social robots were identified. The study suggested socialization, education, entertainment, and expert counseling as childcare functions of social robots and created questionnaire items to explore different aspects of the parents’ views of these different functions. The results suggested positive impacts of social robots in childcare through aspects such as social interactions and entertainment. Different parenting conditions such as parenting styles (work-oriented, dominant, etc.) and children’s ages were reported to change parents’ needs for specific childcare functions. This implies that robots can be strategically designed and introduced to customer groups in line with their characteristics. The study shown in [79] focused on games involving humans and robots in physical and demanding activities. In this context, robots need to be perceived as rational agents aiming to win the game, and the study focused on deciding and communicating deceptive behaviors in robots. This strategy improved human-robot interaction by helping robots match the expectation of interacting with people to attribute rationality to the robot companion. Another field of research where social robotics had applications in affective computing, aiming to understand the effect of a person using specific signals and modalities, and applied in education for example [80]. In [81], a children companion robot was equipped with the capacity of real-time affective computing, allowing the robot to adapt its behavior to the effect of the child it is interacting with, improving the interaction.

2.7. Other Domains of Research and Application

As stated in [82], human-robot interaction has been explored in children, adults, and seniors, but it was less explored in teens. The authors in [82] state that designing robots for interaction with teens is different from other types of human-robot interaction. Additionally, aside from the different domains of research and application that have already been shown, social robots have been used and explored in different contexts, for different objectives. For instance, “edutainment”, where robots participate in people’s education and entertainment can be mentioned [83]. Additionally, several studies have been made to improve human-robot interaction by embedding human social skills in robots. For example, instead of using pre-programmed manually crafted gestures, a humanoid robot learned, using neural networks and a database of TED talks, to generate gestures by the uttered speech as humans would, in [12]. Storytelling is also a field of human-robot interaction where different aspects of robot behavior can be explored [84] a service robot’s ability to adapt its behavior was addressed by implementing a human-like thought process. Behavior in this context can be defined as a combination of a facial expression, a gesture, and a movement. Social intelligence and familiarity with robots have also been the objective in [14,85]. In [86], A robot was equipped with the ability to assess whether the human interaction partner is lying or not, in the purpose of assessing his trustworthiness and improving the interaction. In [14], a robot used deep neural networks to learn human behavior based on data it gathered during its interactions. The purpose was to use the most appropriate action among waving, looking toward human, waving, and handshaking Additionally, in the context of social intelligence, a vision-based framework for allowing robots to recognize and respond to hand waving gestures were presented in [16], increasing its social believability. Furthermore, a humanoid robot was endowed with human-like welcoming behaviors with enthusiasm for the purpose of drawing the attention of persons entering a building in [17]. In a related application in terms of constraints, a flyer distributing robot for pedestrians in a shopping mall has been presented in [45]. Indeed, the robot needed to draw the attention of pedestrians, plan its motions and behave in a manner helping to make the pedestrians accept the flyers. Additionally, a guide robot was developed in [44] for a science museum. The robot had abilities to build relationships with humans through friendly attitudes and was positively evaluated by visitors. Finally, a design and framework were shown in [87] for a robot intended to be used as a receptionist in a university. The platform consisted of an animatronic head with several degrees of freedom and a capacity in engaging in conversations without ‘a priori information about questions it may have to answer. Such an application is an example of how a social robot can combine aspects from design, hardware, software, artificial intelligence, and communication to play roles that are usually attributed to humans.

3. Modalities of Human-Robot Interaction

As in human–human interaction, several modalities can be used at once in human-robot interaction in social contexts. Vision, eye gaze, verbal dialogue, touch, and gestures are examples of modalities that can be used in this context. In a social context, the intelligence that a robot display depends on the modalities it uses and each modality can have specific importance and effect on the human side of the interaction [88] which translates into the degree of trust that the robot has [89]. Moreover, the acceptance of robots in social interaction depends on their ability to express emotions and they require a proper design of emotional expressions to improve their likability and believability [90] as multimodal interaction can enhance the engagement [91,92,93].

3.1. Vision Systems in Robots

Visual perception provides what was suggested to be the most important information to robots, allowing them to achieve successful interaction with human partners [94]. This information can be used in a variety of tasks, such as navigation, obstacle avoidance, detection, understanding, and manipulation of objects, and assigning meanings to a visual configuration of a scene [95,96]. More specifically, the vision has been used for the estimation of the 3D position and orientation of a user in an environment [97], the estimation of distances between a robot and users [98], tracking human targets and obtaining their poses [83], understanding human behavior aiming to contribute to the cohabitation between assistive robots and humans [99]. Similarly, the vision has been used in a variety of other applications, such as recognizing patterns and figures in exercises in a teaching assistance context in a high school [29], detecting and classifying waste material as a child would do [19], and detecting people entering a building for a possible interaction [17]. Moreover, the vision has been used in [71] for medication sorting, taking into account pill types and numbers, in [100] for sign recognition in a sign tutoring task with deaf or hard of hearing children, and in [101] as part of a platform used for cognitive stimulation in elderly users with mild cognitive impairments.

3.2. Conversational Systems in Robots

Although some applications of social robotics involve robots taking vocal commands without generating a vocal reply [102], interactions can be made richer when the robot can engage in conversations. A typical social robot with autonomous conversation ability must have the capacity to acquire sound signals, process them to recognize the speech, recognize the whole sequence of words pronounced by the human interlocutor, formulate an appropriate reply, and synthesize the sound signal corresponding to the reply, then emit this signal using a loudspeaker. The core component of this ability is the recognition of word sequences and the generation of reply sequences [103]. This can rely on a learning stage where the system acquires the experience of answering word sequences by observing a certain number of conversations that are mainly between humans. Techniques used in this area involve word and character embeddings, and learning through recurrent neural network (RNN) architectures, long short-term memory networks (LSTM), and gated recurrent units (GRU) [103,104]. It is to note that not all social robotic systems with conversational capacities have the same levels of complexity as some use limited vocabularies in their verbal dialogues. In this context, Conversation scenarios were seen in [31], verbal dialogue in [20], dialogues between children and a robot in [33], and some word utterances in [19]. A review of conversational systems usages in psychiatry was made in [105]. It covered different aspects such as therapy bots, avatars, and intelligent animal-like robots. Additionally, an algorithm for dialogue management has been proposed in [106] for social robots and conversational agents. It is aimed at ensuring a rich and interesting conversation with users. Furthermore, robot rejection of human commands has been addressed in [107] with aspects such as how rejections can be phrased by the robot. GPT-3 [108] has emerged as a language model with potential applications in conversational systems and social robotics [109]. However, in several conversational systems, problems have been reported, such as hallucinations [110,111], response blandness, and incoherence [103]. The research work presented in [87] aimed at improving the conversational capabilities of a social robot by reducing the possibility of problems as described above, and improving the human-robot interaction with an expressive face. It intended to have a 3-D printed animatronic robotics head with an eye mechanism, a jaw mechanism, and a head mechanism. The three mechanisms are designed to be driven by servo motors to actuate the head synchronously with the audio output. The robotics head design is optimized to fit microphones, cameras, and speakers. The robotics head is envisioned to meet students and visitors in a university. To ensure the appropriateness of the interactions, several stages will be included in the control framework of the robot and a database of human–human conversations will be built upon for the machine learning of the system. This database will be built in the aim of training the conversational system in contexts similar to its contexts of usage. This will increase the adequacy of the conversational system’s parameters with respect to the tasks it is required to do, and increase the coherence and consistency of the utterances it produces. For that, the recorded data will comply with the following specifications:
  • Context of the interactions: Visitors approach the receptionist and engage in conversations in English. Both questions and answers will be included in the database.
  • Audio recordings of the conversations: a text by speech recognition modules is used to transcript the conversation.
  • Video recordings of the interaction, showing the face and upper body of the receptionist, with a quality of images usable by body posture recognition systems.
  • The collected data will be used to progressively train the system. Each conversation will be labeled with the corresponding date, time and interaction parties.
  • Participants will be asked to be free to ask questions they may have to inquire about the center in English, without having any other constraint or any specific text to pronounce.

3.3. Expressions and Gestures

Aside from the ability to process and generate sequences of words, a social robot requires more capacities to increase engagement and realism in the interaction with a human. This can be done through speech-accompanying gestures and facial expressions. Indeed, facial expression has an important role in communication between humans because it is rich in information, together with gestures and sound [112,113,114]. This issue has been studied in psychology, and research indicates that there are six main emotions associated with distinctive facial expressions [115]. At Columbia University [116], scientists and engineers developed a robot that can raise eyebrows, smile, and have forehead wrinkles similar to humans. This robot can express the face more accurately compared to the rest of the robots. This robot, called Eva, can mimic head movements and facial expressions. In this robot, 25 muscles are used, and 12 of them are dedicated specifically to the face. These muscles can produce facial skin excitations of up to 15 mm. In other works, different examples can be found for applications of gestures and expressions in social robotics. For instance, gestures have been combined with verbal dialogue and screen display in [20] for health data acquisition in hospitals with Pepper. In [40], a robot with the ability to display facial expressions was used in studies related to storytelling robots. These studies focused on the roles of the emotional facial display, contextual head movements, and voice acting. In [113], a framework for generating robot behaviors using speech, gestures, and facial expressions was proposed, to improve the expressiveness of a robot in interaction with humans. of interaction with human users

4. Robotic Platforms Used in Social Robotics

Over the different research works covered in this paper, different robotic platforms have been used, with different possible modalities of interaction with humans. It can be easily stated that the Softbank Robotics [117] Nao robot [118] has had a great share of the previous work and studies [12,25,27,29,30,36,38,39,41,48,50,54,119,120] for the different aspects of its usability as it will be shown later. Another humanoid robot of Softbank Robotics, Pepper [121], was also widely used [16,19,20,21,43]. Other robotic platforms that can be mentioned are ASIMO [44,84,122,123], iCub [124], TERESA [15,22], Double [23,47], Furhat [26,125], Stevie [62,126], LUCA [33] which was inspired by the OPSORO platform [127,128], The Universal Robot Trainer (URT) [70], Kaspar [37,77], Mio Amico [42], Reeti [40,129], BabeBay [81], Kiddo [49], Paro [130,131], Tega [132], and Huggable [133,134]. Additionally, Sophia has emerged as a framework for robotics and artificial intelligence research that made appearances in television shows and conferences, also becoming the world’s first robot citizen and the first robot Innovation Ambassador for the United Nations Development Program [135]. It is a human-like robot with the capacity to dialogue, show expressions, walk, shake hands and perform gestures [136]. Another platform for social robotics, enabling research and development applications and providing a wide range of capabilities such as appearance and voice customization, human-mimicking expressions, user tracking and conversations is Furhat [125,137]. Additionally, Sony Aibo has been presented in 1999 as a dog-like robot and is still being produced in its fourth generation currently [138,139]. The research was conducted on the social behaviors toward Aibo and the bonding humans can have with it [140].
Naturally, a common ability of these different types of robotic platforms is interaction with humans. However, the interaction modalities are not always the same, which affects their usability according to the purposes of each study. Figure 3 shows images of the Nao and Pepper robots, exhibiting some of their capabilities such as changing body posture and displaying information.
Table 2 shows several platforms with some of their characteristics and features. It allows us to see the different shapes, sizes, shapes, and capabilities of robots used in social contexts. Many robots had different versions since their first appearance, with features and degrees of freedom being added or omitted. In this context, criteria for the development of robots for interaction with humans have been becoming defined and improved over the years. For instance, as stated in [77], a humanoid robot developed for children with ASD should be user-focused rather than technology-focused, usable, reliable in terms of the usage made by its users, and safe. Another parameter to take into consideration when characterizing social robots is the degree of autonomy a robot has, with the modalities of interaction with humans that its capacities allow it to use.

5. Relationships with Recent Advances in Different Domains

The rise of popularity of social robots has, with no doubt, happen not only due to advances in different areas of science and technology but also to the rising acceptability of humans for machines and robots in a variety of environments and contexts. To make a robot, several components are required, including sensors, actuators, processors, communication devices, and chassis materials. Software and data processing techniques are then required for the control of the different parts of the robot, according to the goals and tasks defined. Advances in different domains, positively impact the usage of social robots, are reviewed in this section.

5.1. Artificial Intelligence

Artificial intelligence advances have been benefiting technology on different levels, whether robotic or not [147,148]. In particular, the different tasks required in a social robotic platform, involving aspects of artificial intelligence, have taken benefited from the development and usage of databases, toolboxes and programming environments that have been shown to be reliable and used for that in several works targeting different applications. The increasing availability of machine learning and data processing toolboxes, which are open source in a variety of cases, is allowing fast advances in different areas with applications in social robotics. For instance, OpenPose allows the estimation of body and face characteristics of humans in interaction with robots [12,17,149]. Also, GloVe is a reliable word embedding model used in conversational systems [149,150]. Another example is Handtrack, an end-to-end trainable CNN model that learns to track the human hand [16,151]. Also, studies have been made in the aim of modeling and digitizing human emotions, which can have projections on social robot behavior and intelligence perception [152,153].

5.2. Manufacturing Technologies

Robotics in general, and the social robotics domain in particular, have been positively impacted by the advances in technologies related to manufacturing. Indeed, it has been made easier to design, prototype, and manufacture a robotic platform chassis and exterior due to the rising availability of laser cutting, 3D printers, and design software [128,154,155,156,157]. Moreover, the advances in semiconductor technologies have made it possible to equip robots with processors and sensors of higher performance, smaller dimensions, and lower costs.

5.2.1. Additive Manufacturing

In [158], the flexibility and customization enabled by 3D printing are highlighted and it is suggested that this technology will continue to be adopted by robotics companies. The same paper reports that even though 3D printing was primarily used for prototyping, it was becoming ready to be used in production. In [159], a literature review of 3D printing in humanoid and social robotics has been made. It leads to a set of design principles contributing to a home-built 3D printed humanoid robot. InMoov, an open-source 3D-printed humanoid robot, has been conceived as a platform for different uses, which is easily reproducible. It was relied on to reach other platforms with the aim of reproducibility, affordability, adding more features, and improvement [160,161,162]. The work shown in [163] addresses specific polymeric materials used in 3D printing according to their suitability for robotic applications and constraints. Presently, 3D printing-based robots can be seen in robotics education [155], allowing the reduction of costs and demonstration and application concepts, and in entertainment and assistance for children and elderly people [156]. They can also be seen in conversational humanoid robotics [157] allowing the configuration of a human-like face with emotions displayed due to servomotors. The flexibility allowed by 3D printing can also be exploited to easily change designs and reconfigure robots, as shown in [164] where an approach for socially interactive robots reconfiguration for end users has been presented, offering different functionalities. In the same context, 3D printing was relied on in [165] to reach a robot head in an iterative process taking into account the opinions of novices and expert robot users. Additionally, the design shown in [87] was conceived for 3D printing, it consisted of a head shape with several mechanisms inside to actuate the head, the eyes, the jaw, and the neck. This flexibility allows implementing designs with the ability to perform rich interactions with humans through software and hardware-based expression generation. Figure 4 shows this design.

5.2.2. Semiconductor Devices

Being used in several parts of a robot, semiconductor-based devices have benefited robotic technologies, allowing the expansion of the capabilities and usefulness of robots. At the same time, the usage of robotic technologies in semiconductor manufacturing has allowed remarkable advances in efficiency, productivity and quality [166,167]. The work shown in [168] presents a robotic system architecture, where the role of semiconductors in some of its main components is crucial as follows:
  • Processors: to carry out and coordinate different tasks of the robotic system.
  • Human machine interface: in screen displays, LEDs, microphones, speakers and their drivers.
  • Sensors: for position, speed, current, distance and orientation for example.
  • Drivers: for the different types of actuators used in robots.

5.3. Processing Technologies

As expectations from social robots evolve, tasks assigned to them increase in computational complexity and require processors, storage spaces, and interfaces capable of handling them. This can be said about various signal and image processing algorithms, machine learning techniques, kinematics, motor command and control, and access to a variety of sensors that the robots can be equipped with. Social robotics has thus taken benefit of the development of processing technologies that are embeddable and reliable in such contexts with real-time constraints. Table 3 lists several robotic platforms and shows their respective processors and other resources. As it can be seen, several platforms are equipped with computational capabilities of average to high-performance computers, with advanced central processing units, spacious memory units, and graphical processing units (GPUs) in some cases. In 2011, an article published on the Robotics Business Review website reported evidence suggesting that GPU processing was already impacting robotics [169]. The article showed the usages of GPUs in autonomous driving, creating depth maps for robots allowing them to navigate and climb stairs for example. In social robotics, and in the same context, GPUs have been used in social robotic experiments, even if not embedded on the robotic platforms [170,171]. For instance, the musical robot companion Shimi, was run in [172] using a GPU for the purpose of enabling it to perform gestures and musical responses to human speech. In this context, NVIDIA developed the Isaac SDK to enable GPU-accelerated algorithms and deep neural networks, and machine learning workflows [173] to accelerate robot developments.

5.4. Operating Systems and Other Software

Aside from the increasing availability of open-source artificial intelligence tools used in social robotics, other types of software and operating systems have been relied on, in allowing social robots operation and tasks. For instance, the Robot Operating System (ROS) [180] is used to build applications in several areas of robotics, including social robots [181]. Additionally, among the various types and usages of robots displayed on the ROS website, social robots such as QTrobot and ARI can be found [182,183] ARI uses ROS to implement capabilities such as perception, cognition, navigation, and expression through gestures and behaviors, and QTrobot uses ROS for emotion, speech, and gestures. ROS has also been used in ROSBOT [184], a low-cost robotic platform with the capacities of obstacle avoiding navigation, face detection, and other social behaviors. Moreover, ROS was used in the implementation of cognitive mechanisms to allow robots to socially interact with humans [185,186,187]. Figure 5 shows another platform, the UXA-90 humanoid robot, developed to serve several purposes with 23 degrees of freedom and supporting ROS [188].
The Softbank Robotics NAO and Pepper robots use the NaoQI framework [189], relying on OpenNao which is a GNU/Linux distribution based on Gentoo [190]. However, the ROS driver naoqi_driver can be used for these robots. A review of the most famous robotic frameworks and middleware has been made in [191]. Not all these frameworks and middleware have been witnessed to be used in social contexts, but the paper provides a comparison between them. This comparison took into account metrics such as the operating systems supported, the programming languages that can be used, open-source aspects, distributed execution, and real-time orientation. ROS has been said to be the robotic framework with the larger momentum and appeal to the community.
Not only have social robotic platforms made use of the available software tools and operating systems but also some tools have been proposed specifically for the social robotics field. In [192], a set of conventions and standard interfaces for HRI scenarios have been presented, designed to be used with ROS, under the ROS4HRI framework. They cover functionalities required for human-robot interaction such as skeleton tracking, face recognition, and natural language processing. The authors of this paper reported the packages people and cob_people_perception as two ROS projects in the human-robot interaction context, which were not multimodal [193,194]. The Human-Robot Interaction Operating System has been proposed in [195] as an infrastructure providing a structured software framework allowing humans and robots to communicate and work as partners, coordinating their actions through dialogue. Additionally, Adaptive Character of Thought-Rational/Embodied (ACT-R/E) has been proposed for human-robot interaction in building models of people to understand how they think [196]. These models were proposed to be used in improving the abilities of robots in interacting with humans.

6. Metrics of Human Perception and Acceptability

The usage of social robots in the different environments and contexts presented above is subjected to their acceptability by humans as partners in the interaction. Indeed, to be accepted in social contexts, robots need to show degrees of intelligence, morphology, or usefulness that can be judged positively by users, not to mention cultural influences on expectations towards and responses to social robots [197]. The study published in 2021 in [198] focused on the perception that humans have of the cognitive and affective abilities of robots and began with the hypothesis that this perception varied in accordance with the degree of human-likeness that robots have. However, the results obtained with students on four robots used in the study did not prove this hypothesis. A study made in 2005 in [199] showed the acceptability of persons for robots as companions in the home, more as assistants, machines, or servants than as a friend. More recently, the literature review and study made in [200] mentions anthropomorphism, animacy, likeability, perceived intelligence, and perceived safety as five key concepts in human-robot interaction. The study also emphasized the importance of being aware of human perception and cognition measures developed by psychologists for engineers developing robots. Additionally, according to the tasks expected from the robots, different measures of performance can be made, such as true recognition measures in speech recognition tasks. But a robot can have a high performance in a specific task, without having a positive impact on its social context. Therefore, the performances of robots in social usage are in many cases measured through evaluations made by humans using questionnaires and metrics calculated based on them. The outcomes of such evaluations are affected by the subjectivity of the persons participating in them and their numbers. In this context, certain metrics/measures can be mentioned as follows:
  • in [15], a robotic platform was equipped with the capacity to perform the two tasks of group interaction, where it had to maintain an appropriate position and orientation in a group, and the person following. The human evaluation began with a briefing of 15 subjects about the purpose of each task, followed by a calibration step where the subjects were shown human-level performance in each task, followed by interaction with the robotic platform for each task. Then, the subjects were asked to rate the social performance of the platform with a number from 1 to 10 where 10 was human-level performance. The authors suggested increasing the number of subjects and a more detailed questionnaire to be necessary for reaching definitive conclusions.
  • the “Godspeed” series of questionnaires has been proposed in [200] to help creators of robots in the robot development process. Five questionnaires using 5-point scales address the anthropomorphism, animacy, likeability, perceived intelligence, and perceived safety of robots. For example, in the anthropomorphism questionnaire (Godspeed I), participants are asked to rate their impressions of the robot with an integer from fake (1) to natural (5), and from machine-like (1) to human-like (5), and from artificial (1) to lifelike (5). Also in the animacy questionnaire (Godspeed II), participants can rate the robot for example from dead (1) to alive (5), from stagnant (1) to lively (5), and from inert (1) to interactive (5). The authors in [200] report cultural backgrounds, prior experiences with robots, and personality to be among the factors affecting the measurements made in such questionnaires. Furthermore, the perceptions of humans are unstable as their expectations and knowledge change with the increase of their experiences with robots. This means, for the authors in [200], that repeating the same experiment after a long duration of time would yield different results.
  • in the context of elderly care and assistance, the Almere model was proposed in [201] as an adaptation and theoretical extension of the Unified Theory of Acceptance and Use of Technology (UTAUT) questionnaire [202]. Questionnaire items in the Almere model were adapted from the UTAUT questionnaire to fit the context of assistive robot technology and address elderly users in a care home. Different constructs are adopted and defined and questionnaires related to them, respectively. This resulted in constructs such as the users’ attitude towards the technology their intention to use it, their perceived enjoyment, perceived ease of use, perceived sociability and usefulness, social influence and presence, and trust. Experiments made on the model consisted of a data collection instrument with different questionnaire items on a 5-point Likert-type scale ranging from 1 to 5, corresponding to statements ranging from “totally disagree” to “totally agree”, respectively.
Other metrics and approaches for the evaluation of the engagement in the interaction between humans and robots have been proposed. The work presented in [203] proposes metrics that can be easily retrieved from off-the-shelf sensors, by static and dynamic analysis of body posture, head movements and gaze of the human interaction partner.
The work made in [200] revealed two important points related to the assessment of human-robot interaction: the need for a standardized measurement tool and the effects of user background and time on the measurements. The authors also invited psychologists to contribute to the development of the questionnaires. These issues can have implications for social robotics studies that should be addressed to improve assessment quality and results and advance robotic system designs and tasks accurately. More recently, the work shown in [204] proposed a standardized process for choosing and using scales and questionnaires used in human-robot interaction. For instance, the authors in [204] specified that a scale cannot be trusted in a certain study if not already validated in a similar study and that scales can be unfit or have limitations concerning a specific study. In such a case, they should be modified and re-validated.

7. Future of Social Robotics

Across the different areas of research that have been shown in Section 2.1, research work has been made in many cases as propositions and proofs of concept with expectations and propositions for future work. Different works have addressed the future of social robotics from their respective contexts and points of view. Additionally, while some suggestions take into account purely technical aspects, such as adding or improving functionalities of robots, others address the importance of robot acceptability and familiarity, as well as the robot users’ viewpoints as factors to be taken into account. It is possible to envision robots being used in various everyday life scenarios, from entertainment to education and medical care, engaging in active and rich interactions with humans. Humans would have a clear idea of the capabilities of the robots they interact with, they would attribute degrees of trust to them and grow familiar with their shapes. Robots would be capable of processing various types of signals to understand the needs, behaviors, and emotions of people, and they would convey signals in real time, according to their perceptions. Below are summaries of future expectations from papers in different areas.
  • Education: in [28], the authors suggested motivating schoolteachers to introduce a collaborative robot into their lectures. Enabling robot-supported language learning for preschool children was proposed as a long-term goal in [50]. In [51], where a review of robots in education was made, improving the speech understanding capabilities of robots and reproducing human-like behavior were proposed. In [27], the authors specified that second language anxiety and negative attitudes toward robots need to be carefully considered before introducing robots to students as second language tutors. Additionally, in [52], incorporating robots into teacher education or professional development programs was proposed. Teachers, students, and social robots were said to become all key actors in future classrooms, and teachers’ attitudes and beliefs were said to have possible influences on the future deployment of social robots.
  • Care and assistance: in [61], the importance of experience in working with robots and raising awareness about what care robots can do was shown, in the objective of moving away from preconceptions based on fiction and imaginaries of care robots. In [58], an aim to achieve a design that is easily deployed in multiple locations, and contains all the necessary information for repeated deployments was expressed. In [205], an emotion recognition algorithm and an imitation algorithm was said to bring improvements to a robotic system for physical training of older adults. In [55], where a review of the usages of socially assistive robots in elderly care was made, the authors concluded that studies should be clearer about the precise role of any robot, and should use validated measures to assess their effectiveness. In the same context, a sophisticated speech analysis ability and accuracy of understanding language were said to be desired to improve the interaction between a human being and a robot in [59]. Under the topic of the acceptance of healthcare robots for older adults in [60], matching individual needs and preferences to the robot was said to possibly improve the acceptance. An alternatively proposed approach was to alter users’ expectations to match the capabilities of the robot.
  • Children companionship: in [41], it was said that humanoid robots are promising for robot-mediated education with primary school-aged children due to their capacity of making voices and gestures that motivate children in learning activities. In [79], where the work addressed robots playing games with people, it was said that a robot can have a sort of character that would support its perception as a rational agent, by taking measures such as adapting the behavior and strategy of the robot to the real-time perception it has of the humans it interacts with.
  • Autism and medicine: in [77], a list of considerations to be taken into account when developing robots for children with ASD has been made. It shows that such robots should be user-focused, usable, reliable, safe, and affordable. In [74], the robotic scenario was said to be an excellent way to elicit behaviors in children with ASD through interaction, analysis of the child’s behavior, and adaptation to it. According to the authors, Introducing robots into therapy would be of great clinical interest. In [75], works on social signal processing and socially assistive robotics were reported and issues that should be addressed by researchers in these research domains were listed. Among them are the machine understanding of typical and autistic behaviors, and the availability of databases of children with ASD interactions.
  • security of robotic systems: an important aspect to address in social robotics is security and cybersecurity. Indeed, intelligent systems can help protect the security of users but hackers could attack social robot users from different vectors [206,207,208]. This should be taken into account when considering the use of social robots [209]. Work has been done in this domain to improve security, such as the guidelines published by a High-Level Expert Group established by the European Commission on trustworthy artificial intelligence. Additionally, the Trusted-ROS system was proposed in [207] for security improvement in humanoid robots. Additionally, recommendations were presented in [208] such as multi-factor authentication and multi-factor cryptography.

8. Conclusions

This paper showed a survey made on different aspects of social robotics, without a focus on a specific aspect. Applications of social robotics, modalities of the interaction of social robots with humans, platforms used, relationships with advances in different technological areas, and measures of acceptability of humans towards robots have been addressed. It can be seen that humanity can expect to co-exist with robots in the future, in different aspects of everyday life. This is not currently the case, with many of the previous work consisting of research studies rather than actual implementations of robots that are ready to use. This is not only due to the incomplete maturity of technologies that robots require to operate with humans, with enough performances to be trusted and believed, but also to the incomplete comprehension of how humans may be willing to accept robots and in which contexts. In this context, it was argued in [210] that there is a phenomenon becoming common in the social robotics field, named the roboid: a robot that is still at the prototype stage but claims to be fully functioning.
The survey made in this paper is not exhaustive with regards to any of the areas addressed but aims to provide readers with a tour of the world of social robotics. More areas can be addressed in the future, with similar surveys that can be done every few years, due to the rapid advance and expansion of social robotics technologies and usages.

Author Contributions

Conceptualization, K.Y., S.S., S.A. and T.B.; methodology, K.Y., S.S., S.A. and T.B.; software, K.Y. and S.S.; validation, S.S., S.A. and T.B.; formal analysis, K.Y.; investigation, K.Y. and S.S.; resources, K.Y., S.S., S.A. and T.B.; data curation, K.Y. and S.S.; writing—original draft preparation, K.Y. and S.S.; writing—review and editing, K.Y., S.S., S.A. and T.B.; visualization, K.Y. and S.S.; supervision, S.A. and T.B.; project administration, S.A. and T.B. All authors have read and agreed to the published version of the manuscript.

Funding

This research received no external funding.

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Sarrica, M.; Brondi, S.; Fortunati, L. How many facets does a “social robot” have? A review of scientific and popular definitions online. Inf. Technol. People 2019, 33, 1–21. [Google Scholar] [CrossRef]
  2. Duffy, B.R. Anthropomorphism and the social robot. Robot. Auton. Syst. 2003, 42, 177–190. [Google Scholar] [CrossRef]
  3. Fong, T.; Nourbakhsh, I.; Dautenhahn, K. A survey of socially interactive robots. Robot. Auton. Syst. 2003, 42, 143–166. [Google Scholar] [CrossRef] [Green Version]
  4. Breazeal, C. Toward sociable robots. Robot. Auton. Syst. 2003, 42, 167–175. [Google Scholar] [CrossRef]
  5. Li, H.; Cabibihan, J.J.; Tan, Y. Towards an Effective Design of Social Robots. Int. J. Soc. Robot. 2011, 3, 333–335. [Google Scholar] [CrossRef] [Green Version]
  6. Li, J.; Chignell, M. Communication of Emotion in Social Robots through Simple Head and Arm Movements. Int. J. Soc. Robot. 2010, 3, 125–142. [Google Scholar] [CrossRef]
  7. Shaw-Garlock, G. Looking Forward to Sociable Robots. Int. J. Soc. Robot. 2009, 1, 249–260. [Google Scholar] [CrossRef]
  8. Sabanovic, S. Robots in Society, Society in Robots. Int. J. Soc. Robot. 2010, 2, 439–450. [Google Scholar] [CrossRef]
  9. Korn, O. (Ed.) Social Robots: Technological, Societal and Ethical Aspects of Human-Robot Interaction; Springer: Berlin/Heidelberg, Germany, 2019. [Google Scholar]
  10. Karar, A.; Said, S.; Beyrouthy, T. Pepper Humanoid Robot as a Service Robot: A Customer Approach. In Proceedings of the 2019 3rd International Conference on Bio-Engineering for Smart Technologies (BioSMART), Paris, France, 24–26 April 2019; pp. 1–4. [Google Scholar] [CrossRef]
  11. Bennett, C.C.; Sabanociv, S. Deriving Minimal Features for Human-Like Facial Expressions in Robotic Faces. Int. J. Soc. Robot. 2014, 6, 367–381. [Google Scholar] [CrossRef]
  12. Yoon, Y.; Ko, W.R.; Jang, M.; Lee, J.; Kim, J.; Lee, G. Robots Learn Social Skills: End-to-End Learning of Co-Speech Gesture Generation for Humanoid Robots. In Proceedings of the International Conference on Robotics and Automation (ICRA), Montreal, QC, Canada, 20–24 May 2019. [Google Scholar]
  13. Ko, W.R.; Kim, J.H. Behavior SeBehavior Selection of Social Robots Using Developmental Episodic Memory-based Mechanism of Thought. In Proceedings of the IEEE International Conference on Consumer Electronics—Asia (ICCE-Asia), Jeju, Korea, 24–26 June 2018. [Google Scholar]
  14. Qureshi, A.H.; Nakamura, Y.; Yoshikawa, Y.; Ishiguro, H. Robot gains Social Intelligence through Multimodal Deep Reinforcement Learning. In Proceedings of the IEEE-RAS 16th International Conference on Humanoid Robots (Humanoids), Cancun, Mexico, 15–17 November 2016. [Google Scholar]
  15. Shiarlis, K.; Messias, J.; Whiteson, S. Acquiring Social Interaction Behaviours for Telepresence Robots via Deep Learning from Demonstration. In Proceedings of the IEEE/RSJ International Conference on Intelligent Robots and Systems (IROS), Vancouver, BC, Canada, 24–28 September 2017. [Google Scholar]
  16. Castellano, G.; Cervelione, A.; Cianciotta, M.; De Carolis, B.; Vessio, G. Recognizing the Waving Gesture in the Interaction with a Social Robot. In Proceedings of the 29th IEEE International Conference on Robot and Human Interactive Communication (RO-MAN), Naples, Italy, 31 August–4 September 2020. [Google Scholar]
  17. Saad, E.; Broekens, J.; Neerincx, M.A.; Hindriks, K.V. Enthusiastic Robots Make Better Contact. In Proceedings of the IEEE/RSJ International Conference on Intelligent Robots and Systems (IROS), Macau, China, 3–8 November 2019. [Google Scholar]
  18. Barra, P.; Bisogni, C.; Distasi, R.; Rapuano, A. VMPepper: How to Use a Social Humanoid Robot for Interactive Voice Messaging. In Proceedings of the Fourth International Conference on Applications and Systems of Visual Paradigms, VISUAL, Rome, Italy, 30 June–4 July 2019. [Google Scholar]
  19. Castellano, G.; De Carolis, B.; D’Errico, F.; Macchiarulo, N.; Rossano, V. PeppeRecycle: Improving Children’s Attitude Toward Recycling by Playing with a Social Robot. Int. J. Soc. Robot. 2021, 13, 97–111. [Google Scholar] [CrossRef]
  20. Van der Putte, D.; Boumans, R.; Neerincx, M.; Rikkert, M.O.; De Mul, M. A Social Robot for Autonomous Health Data Acquisition among Hospitalized Patients: An Exploratory Field Study. In Proceedings of the 14th ACM/IEEE International Conference on Human-Robot Interaction (HRI), Daegu, Korea, 11–14 March 2019. [Google Scholar]
  21. Schicchi, D.; Pilato, G. A Social Humanoid Robot as a Playfellow for Vocabulary Enhancement. In Proceedings of the Second IEEE International Conference on Robotic Computing, Laguna Hills, CA, USA, 31 January–2 February 2018. [Google Scholar]
  22. Shiarlis, K.; Messias, J.; van Someren, M.; Whiteson, S.; Kim, J.; Vroon, J.; Englebienne, G.; Truong, K.; Evers, V.; Pérez-Higueras, N.; et al. TERESA: A Socially Intelligent SEmi-autonomous Telepresence System. In Proceedings of the International Conference on Robotics and Automation, Seattle, WA, USA, 26–30 May 2015. [Google Scholar]
  23. Niemela, M.; van Aerschot, L.; Tammela, A.; Aaltonen, L.; Lammi, H. Towards Ethical Guidelines of Using Telepresence Robots in Residential Care. Int. J. Soc. Robot. 2019, 13, 431–439. [Google Scholar] [CrossRef] [Green Version]
  24. Zhang, G.; Hansen, J.P.; Minkata, K.; Alapetite, A.; Wang, Z. Eye0Gaze-Controlled Telepresence Robots for People with Motor Disabilities. In Proceedings of the 14th ACM/IEEE International Conference on Human-Robot Interaction (HRI), Daegu, Korea, 11–14 March 2019. [Google Scholar]
  25. Hood, D.; Lemaignan, S.; Dillenbourg, P. When Children Teach a Robot to Write: An Autonomous Teachable Humanoid Which Uses Simulated Handwriting. In Proceedings of the 10th ACM/IEEE International Conference on Human-Robot Interaction, Portland, OR, USA, 2–5 March 2015. [Google Scholar]
  26. Engwall, O.; Lopes, J.; Ahlund, A. Robot Interaction Styles for Conversation Practice in Second Language Learning. Int. J. Soc. Robot. 2020, 13, 251–276. [Google Scholar] [CrossRef] [Green Version]
  27. Kanero, J.; Oranc, C.; Koskulu, S.; Kumkale, G.T.; Goksun, T.; Kuntay, A.C. Are Tutor Robots for Everyone? The Influence of Attitudes, Anxiety, and Personality on Robot-Led Language Learning. Int. J. Soc. Robot. 2022, 14, 297–312. [Google Scholar] [CrossRef]
  28. Shimaya, J.; Yoshikawa, Y.; Palinko, O.; Ogawa, K.; Jinnai, N.; Ishiguro, H. Active Participation in Lectures via a Collaboratively Controlled Robot. Int. J. Soc. Robot. 2021, 13, 587–598. [Google Scholar] [CrossRef]
  29. Reyes, G.E.B.; Lopez, E.; Ponce, P.; Mazon, N. Role Assignment Analysis of an Assistive Robotic Platform in a High School Mathematics Class, Through a Gamification and Usability Evaluation. Int. J. Soc. Robot. 2021, 13, 1063–1078. [Google Scholar] [CrossRef]
  30. Vogt, P.; van den Berghe, R.; de Haas, M.; Hoffman, L.; Kanero, J.; Mamus, E.; Montanier, J.M.; Oranc, C.; Oudgenoeg-Paz, O.; Hernandez Garcia, D.; et al. Second language tutoring using social robots: L2TOR—The movie. In Proceedings of the 14th ACM/IEEE International Conference on Human-Robot Interaction (HRI), Daegu, Korea, 11–14 March 2019. [Google Scholar]
  31. Obayashi, K.; Kodate, N.; Masuyama, S. Assessing the Impact of an Original Soft Communicative Robot in a Nursing Home in Japan: Will Softness or Conversations Bring more Smiles to Older People? Int. J. Soc. Robot. 2022, 14, 645–656. [Google Scholar] [CrossRef] [PubMed]
  32. Luperto, M.; Monroy, J.; Renoux, J.; Lunardini, F.; Basilico, N.; Bulgheroni, M.; Cangelosi, A.; Cesari, M.; Cid, M.; Ianes, A.; et al. Integrating Social Assistive Robots, IoT, Virtual Communities and Smart Objects to Assist at-Home Independently Living Elders: The MoveCare Project. Int. J. Soc. Robot. 2022, 14, 1–31. [Google Scholar] [CrossRef]
  33. Ismail, L.I.; Hanapiah, F.A.; Belpaeme, T.; Dambre, J.; Wyffels, F. Analysis of Attention in Child-Robot Interaction Among Children Diagnosed with Cognitive Impairement. Int. J. Soc. Robot. 2021, 13, 141–152. [Google Scholar] [CrossRef]
  34. Schrum, M.; Park, C.H.; Howard, A. Humanoid Therapy Robot for Encouraging Exercise in Dementia Patients. In Proceedings of the 14th ACM/IEEE International Conference on Human-Robot Interaction (HRI), Daegu, Korea, 11–14 March 2019. [Google Scholar]
  35. Moharana, S.; Panduro, A.E.; Lee, H.R.; Rick, L.D. Robots for Joy, Robots for Sorrow: Community Based Robot Design for Dementia Caregivers. In Proceedings of the 14th ACM/IEEE International Conference on Human-Robot Interaction (HRI), Daegu, Korea, 11–14 March 2019. [Google Scholar]
  36. Anzalone, S.M.; Tilmont, E.; Boucenna, S.; Xavier, J.; Jouen, A.L.; Bodeau, N.; Maharatna, K.; Chetouani, M.; Cohen, D.; the MICHELANGELO Study Group. How children with autism spectrum disorder behave and explore the 4-dimensional (spatial 3D + time) environment during a joint attention induction task with a robot. Res. Autism Spectr. Disord. 2014, 8, 814–826. [Google Scholar] [CrossRef]
  37. Huijnen, C.A.G.J.; Verreussel-Willen, H.A.M.D.; Lexis, M.A.S.; de Witte, L.P. Robot KASPAR as Mediator in Making Contact with Children with Autism: A Pilot Study. Int. J. Soc. Robot. 2021, 13, 237–249. [Google Scholar] [CrossRef]
  38. Taheri, A.; Meghdari, A.; Mahoor, M.H. A Close Look at the Imitation Performance of Children with Autism and Typically Developing Children Using a Robotic System. Int. J. Soc. Robot. 2021, 13, 1125–1147. [Google Scholar] [CrossRef]
  39. Chung, E.Y.H. Robot-Mediated Social Skill Intervention Programme for Children with Autism Spectrum Disorder: An ABA Time-Series Study. Int. J. Soc. Robot. 2021, 13, 1095–1107. [Google Scholar] [CrossRef]
  40. Striepe, H.; Donnermann, M.; Lein, M.; Lugrin, B. Modeling and Evaluating Emotion, Contextual Head Movement and Voices for a Social Robot Storyteller. Int. J. Soc. Robot. 2021, 13, 441–457. [Google Scholar] [CrossRef]
  41. Desideri, L.; Bonifacci, P.; Croati, G.; Dalena, A.; Gesualdo, M.; Molinario, G.; Gherardini, A.; Cesario, L.; Ottaviani, C. The Mind in the Machine: Mind Perception Modulates Gaze Aversion During Child-Robot Interaction. Int. J. Soc. Robot. 2021, 13, 599–614. [Google Scholar] [CrossRef]
  42. Filippini, C.; Spadolini, E.; Cardone, D.; Bianchi, D.; Preziuso, M.; Sciarretta, C.; del Cimmuto, V.; Lisciani, D.; Merla, A. Facilitating the Child-Robot Interaction by Endowing the Robot with the Capability of Understanding the Child Engagement: The Case of Mio Amico Robot. Int. J. Soc. Robot. 2021, 13, 677–689. [Google Scholar] [CrossRef]
  43. Uluer, P.; Kose, H.; Gumuslu, E.; Erol Barkana, D. Experience with an Affective Robot Assistant for Children with Hearing Disabilities. Int. J. Soc. Robot. 2021, 16, 1–8. [Google Scholar] [CrossRef]
  44. Iio, T.; Satake, S.; Kanda, T.; Hayashi, K.; Ferreri, F.; Hagita, N. Human-Like Guide Robot that Proactively Explains Exhibits. Int. J. Soc. Robot. 2020, 12, 549–566. [Google Scholar] [CrossRef] [Green Version]
  45. Shi, C.; Satake, S.; Kanda, T.; Ishiguro, H. A Robot that Distributes Flyers to Pedestrians in a Shopping Mall. Int. J. Soc. Robot. 2018, 10, 421–437. [Google Scholar] [CrossRef]
  46. Belay Tuli, T.; Olana Terefe, T.; Ur Rashid, M.M. Telepresence Mobile Robots Design and Control for Social Interaction. Int. J. Soc. Robot. 2021, 13, 877–886. [Google Scholar] [CrossRef]
  47. Double Robotics—Telepresence Robot for the Hybrid Office. Available online: https://www.doublerobotics.com/ (accessed on 1 March 2022).
  48. Mubin, O.; Alhashmi, M.; Baroud, R.; Alnajjar, F.S. Humanoid Robots as Teaching Asistants in an Arab School. In Proceedings of the 31st Australian Conference on Human-Computer Interaction, Fremantle, Australia, 2–5 December 2019. [Google Scholar]
  49. Mispa, T.A.; Sojib, N. Educational Robot Kiddo Learns to Draw to Enhance Interactive Handwriting Scenario for Primary School Children. In Proceedings of the 3rd Intrernational Conference of Intelligent Robotic and Control Engineering (IRCE), Oxford, UK, 10–12 August 2020. [Google Scholar]
  50. Schodde, T.; Bergmann, K.; Kopp, S. Adaptive Robot Language Tutoring Based on Bayesian Knowledge Tracing and Predictive Decision-Making. In Proceedings of the 12th ACM/IEEE International Conference on Human-Robot Interaction, Vienna, Austria, 6–9 March 2017. [Google Scholar]
  51. Mubin, O.; Stevens, C.J.; Shahid, S.; Al Mahmud, A.; Dong, J.J. A Review of the Applicability of Robots in Education. Technol. Educ. Learn. 2013, 1, 13. [Google Scholar] [CrossRef] [Green Version]
  52. Xia, Y.; LeTendre, G. Robots for Future Classrooms: A Cross-Cultural Validation Study of “Negative Attitudes Toward Robots Scale” in the U.S. Context. Int. J. Soc. Robot. 2021, 13, 703–714. [Google Scholar] [CrossRef]
  53. Nomura, T.; Kanda, T.; Suzuki, T. Experimental investigation into influence of negative attitudes toward robots on human-robot interaction. Ai Soc. 2006, 20, 138–150. [Google Scholar] [CrossRef]
  54. Vogt, P.; van den Berghe, R.; de Haas, M.; Hoffman, L.; Kanero, J.; Mamus, E.; Montanier, J.M.; Oranc, C.; Oudgenoeg-Paz, O.; Hernandez Garcia, D.; et al. Second Language Turoting using Social Robots: A Large-Scale Study. In Proceedings of the 14th ACM/IEEE International Conference on Human-Robot Interaction (HRI), Daegu, Korea, 11–14 March 2019. [Google Scholar]
  55. Abdi, J.; Al-Hindawi, A.; Ng, T.; Vizcaychipi, M.P. Scoping review on the use of socially assistive robot technology in elderly care. BMJ Open 2017, 8, e018815. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  56. Lin, C.; Ogata, T.; Zhong, Z.; Kanai-Pak, M.; Maeda, J.; Kitajima, Y.; Nakamura, M.; Kuwahara, N.; Ota, J. Development and Validation of Robot Patient Equipped with an Inertial Measurement Unit and Angular Position Sensors to Evaluate Transfer Skills of Nurses. Int. J. Soc. Robot. 2021, 13, 899–917. [Google Scholar] [CrossRef]
  57. Meia, C.T.; Scheutz, M. Assistive Robots for the Social Management of Health: A Framework for Robot Design and Human–Robot Interaction Research. Int. J. Soc. Robot. 2021, 13, 197–217. [Google Scholar]
  58. Bardaro, G.; Antonini, A.; Motta, E. Robots for Elderly Care in the Home: A Landscape Analysis and Co-Design Toolkit. Int. J. Soc. Robot. 2022, 14, 657–681. [Google Scholar] [CrossRef]
  59. Obayashi, K.; Kodate, N.; Masuyama, S. Enhancing older people’s activity and participation with socially asisstive robots: A multicentre quasi-experimental study using the ICF framework. Adv. Robot. 2018, 32, 1207–1216. [Google Scholar] [CrossRef]
  60. Broadbent, E.; Stafford, R.; MacDonald, B. Acceptance of Healthcare Robots for the Older Population: Review and Future Directions. Int. J. Soc. Robot. 2009, 1, 319–330. [Google Scholar] [CrossRef]
  61. Frennert, S.; Aminoff, H.; Ostlund, B. Technological Framces and Care Robots in Eldercare. Int. J. Soc. Robot. 2021, 13, 317–325. [Google Scholar] [CrossRef] [Green Version]
  62. McGinn, C.; Bourke, E.; Murtagh, A.; Donovan, C.; Cullinan, M.F. Meeting Stevie: Perceptions of a Socially Assistive Robot by Residents and Staff in a Long-term Care Facility. In Proceedings of the 14th ACM/IEEE International Conference on Human-Robot Interaction (HRI), Daegu, Korea, 11–14 March 2019. [Google Scholar]
  63. Williams, A.B.; Williams, R.M.; Moore, R.E.; McFarlane, M. AIDA: A Social Co-Robot to Uplift Workers with Intellectual and Developmental Disabilities. In Proceedings of the 14th ACM/IEEE International Conference on Human-Robot Interaction (HRI), Daegu, Korea, 11–14 March 2019. [Google Scholar]
  64. Monekosso, D.; Florez-Revuelta, F.; Remagnino, P. Ambient Assisted Living [Guest editors’ introduction]. IEEE Intell. Syst. 2015, 30, 2–6. [Google Scholar] [CrossRef]
  65. AAL Home 2020—AAL Programme. Available online: www.aal-europe.eu (accessed on 15 February 2022).
  66. Casiddu, N.; Cesta, A.; Cortellessa, G.; Orlandini, A.; Porfirione, C.; Divano, A.; Micheli, E.; Zallio, M. Robot Interface Design: The Giraff Telepresence Robot for Social Interaction. Biosyst. Biorobot. 2015, 11, 499–509. [Google Scholar] [CrossRef]
  67. Coradeschi, S.; Cesta, A.; Cortellessa, G.; Coraci, L.; Galindo, C.; González-Jiménez, J.; Karlsson, L.; Forsberg, A.; Frennert, S.; Furfari, F.; et al. GiraffPlus: A System for Monitoring Activities and Physiological Parameters and Promoting Social Interaction for Elderly. Adv. Intell. Syst. Comput. 2014, 300, 261–271. [Google Scholar] [CrossRef]
  68. Kabacinska, K.; Prescott, T.J.; Robillard, J.M. Socially Assistive Robots as Mental Health Interventions for Children: A Scoping Review. Int. J. Soc. Robot. 2021, 13, 919–935. [Google Scholar] [CrossRef]
  69. Rasouli, S.; Gupta, G.; Nilsen, E.; Dautenhahn, K. Potential Applications of Social Robots in Robot-Assisted Interventions for Social Anxiety. Int. J. Soc. Robot. 2022. ahead of printing. [Google Scholar] [CrossRef] [PubMed]
  70. Nielsen, C.; Mathiesen, M.; Nielsen, J.; Jensen, L.C. Changes in Heart Rate and Feeling of Safety when Led by a Rehabilitation Robot. In Proceedings of the 14th ACM/IEEE International Conference on Human-Robot Interaction (HRI), Daegu, Korea, 11–14 March 2019. [Google Scholar]
  71. Wilson, J.R.; Lee, N.Y.; Saechao, A.; Tickle-Degnen, L.; Scheutz, M. Supporting Human Autonomy in a Robot-Assisted Medication Sorting Task. Int. J. Soc. Robot. 2018, 10, 621–641. [Google Scholar] [CrossRef]
  72. Chatbots|GPT-3 Demo. Available online: https://gpt3demo.com/category/chatbots (accessed on 13 April 2022).
  73. Delaherche, E.; Chetouani, M.; Bigouret, F.; Xavier, J.; Plaza, M.; Cohen, D. Assessment of the communicative and coordination skills of children with Autism Spectrum Disorders and typically developing children using social signal processing. Res. Autism Spectr. Disord. 2013, 7, 741–756. [Google Scholar] [CrossRef]
  74. Boucenna, S.; Narzisi, A.; Tilmont, E.; Muratori, F.; Pioggia, G.; Cohen, D.; Chetouani, M. Interactive Technologies for Autistic Children: A Review. Cogn. Comput. 2014, 6, 722–740. [Google Scholar] [CrossRef] [Green Version]
  75. Chetouani, M.; Boucenna, S.; Chaby, L.; Plaza, M.; Cohen, D. Social Signal Processing and Socially Assistive Robotics in Developmental Disorders; Cambrige University Press: Cambrige, UK, 2017; pp. 389–403. [Google Scholar] [CrossRef]
  76. Emery, N. The eyes have it: The neuroethology, function and evolution of social gaze. Neurosci. Biobehav. Rev. 2000, 24, 581–604. [Google Scholar] [CrossRef]
  77. Wood, L.J.; Zaraki, A.; Robins, B.; Dautenhahn, K. Developing Kaspar: A Humanoid Robot for Children with Autism. Int. J. Soc. Robot. 2021, 13, 491–508. [Google Scholar] [CrossRef] [Green Version]
  78. Lee, J.; Lee, D.; Lee, J.G. Can Robots Help Working Parents with Childcare? Optimizing Childcare Functions for Different Parenting Characteristics. Int. J. Soc. Robot. 2022, 14, 193–201. [Google Scholar] [CrossRef]
  79. de Oliveira, E.; Donadoni, L.; Boriero, S.; Bonarini, A. Deceptive Actions to Improve the Attribution of Rationality to Playing Robotic Agents. Int. J. Soc. Robot. 2021, 13, 391–405. [Google Scholar] [CrossRef]
  80. Wu, C.H.; Huang, Y.M.; Hwang, J.P. Review of affective computing in education/learning: Trends and challenges. Br. J. Educ. Technol. 2016, 47, 1304–1323. [Google Scholar] [CrossRef]
  81. Zheng, M.; She, Y.; Chen, J.; Shu, Y.; XiaHou, J. BabeBay—A Companion Robot for Children Based on Multimodal Affective Computing. In Proceedings of the 14th ACM/IEEE International Conference on Human-Robot Interaction (HRI), Daegu, Korea, 11–14 March 2019. [Google Scholar]
  82. Bjorling, E.A.; Rose, E.; Davidson, A.; Ren, R.; Wong, D. Can We Keep Him Forever? Teens’ Engagement and Desire for Emotional Connection with a Social Robot. Int. J. Soc. Robot. 2020, 12, 65–77. [Google Scholar] [CrossRef]
  83. Gonzalez-Pacheco, V.; Ramey, A.; Alonso-Martin, F.; Castro-Gonzalez, A.; Salichs, M.A. Maggie: A Social Robot as a Gaming Platform. Int. J. Soc. Robot. 2011, 3, 371–381. [Google Scholar] [CrossRef] [Green Version]
  84. Mutlu, B.; Forlizzi, J.; Hodgins, J. A Storytelling Robot: Modeling and Evaluation of Human-like Gaze Behavior. In Proceedings of the IEEE-RAS Intrernational Conference on Humanoid Robots, Genova, Italy, 4–6 December 2006. [Google Scholar]
  85. Hsieh, W.F.; Sato-Shimokawara, E.; Yamaguchi, T. Enhancing the Familiarity for Humanoid Robot Pepper by Adopting Customizable Motion. In Proceedings of the IECON 2017—43rd Annual Conference of the IEEE Industrial Electronics Society, Beijing, China, 29 October–1 November 2017. [Google Scholar]
  86. Pasquali, D.; Gonzalez-Billandon, J.; Aroyo, A.M.; Sandini, G.; Sciutti, A.; Rea, F. Detecting Lies is a Child (Robot)’s Play: Gaze-Based Lie Detection in GRI. Int. J. Soc. Robot. 2021. [Google Scholar] [CrossRef]
  87. Youssef, K.; Said, S.; Beyrouthy, T.; Alkork, S. A Social Robot with Conversational Capabilities for Visitor Reception: Design and Framework. In Proceedings of the 2021 4th International Conference on Bio-Engineering for Smart Technologies (BioSMART), Paris/Créteil, France, 8–10 December 2021; pp. 1–4. [Google Scholar] [CrossRef]
  88. Novanda, O.; Salem, M.; Saunders, J.; Walters, M.L.; Dautenhahn, K. What Communication Modalities Do Users Prefer in Real Time HRI? arXiv 2016, arXiv:1606.03992. [Google Scholar]
  89. Tatarian, K.; Stower, R.; Rudaz, D.; Chamoux, M.; Kappas, A.; Chetouani, M. How does Modality Matter? Investigating the Synthesis and Effects of Multi-modal Robot Behavior on Social Intelligence. Int. J. Soc. Robot. 2021, 14, 893–911. [Google Scholar] [CrossRef]
  90. Tsiourti, C.; Weiss, A.; Wac, K.; Vincze, M. Multimodal Integration of Emotional Signals from Voice, Body, and Context: Effects of (In)Congruence on Emotion Recognition and Attitudes Towards Robots. Int. J. Soc. Robot. 2019, 11, 555–573. [Google Scholar] [CrossRef] [Green Version]
  91. Feng, Y.; Perugia, G.; Yu, S.; Barakova, E.I.; Hu, J.; Rauterberg, G.W.M. Context-Enhanced Human-Robot Interaction: Exploring the Role of System Interactivity and Multimodal Stimuli on the Engagement of People with Dementia. Int. J. Soc. Robot. 2021, 14, 807–826. [Google Scholar] [CrossRef]
  92. Friedman, N.; Goedicke, D.; Zhang, V.; Rivkin, D.; Jenkin, M.R.M.; Degutyte, Z.; Astell, A.J.; Liu, X.; Dudek, G. Out of My Way! Exploring Different Modalities for Robots to Ask People to Move Out of the Way. 2020. Available online: https://www.semanticscholar.org/paper/Out-of-my-way!-Exploring-Different-Modalities-for-Friedman-Goedicke/c7467ad74a5f72871019d6eb2e24c907b6de108e (accessed on 13 April 2022).
  93. Johnson, D.O.; Agah, A. Human Robot Interaction Through Semantic Integration of Multiple Modalities, Dialog Management, and Contexts. Int. J. Soc. Robot. 2009, 1, 283–305. [Google Scholar] [CrossRef] [Green Version]
  94. Kang, S.H.; Han, J.H. Video Captioning Based on Both Egocentric and Exocentric Views of Robot Vision for Human-Robot Interaction. Int. J. Soc. Robot. 2021. [Google Scholar] [CrossRef]
  95. Kragic, D.; Vincze, M. Vision for Robotics. Found. Trends Robot. 2010, 1, 1–78. [Google Scholar] [CrossRef]
  96. Ronchi, M.R. Vision for Social Robots: Human Perception and Pose Estimation. Ph.D. Thesis, California Institute of Technology, Pasadena, CA, USA, 2020. [Google Scholar]
  97. Garcia-Salguero, M.; Gonzalez-Jimenez, J.; Moreno, F.A. Human 3D Pose Estimation with a Tilting Camera for Social Mobile Robot Interaction. Sensors 2019, 19, 4943. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  98. Pathi, S.K.; Kiselev, A.; Kristoffersson, A.; Repsilber, D.; Loutfi, A. A Novel Method for Estimating Distances from a Robot to Humans Using Egocentric RGB Camera. Sensors 2019, 19, 3142. [Google Scholar] [CrossRef] [Green Version]
  99. Kostavelis, I.; Vasileiadis, M.; Skartados, E.; Kargakos, A.; Giakoumis, D.; Bouganis, C.S.; Tzovaras, D. Understanding of Human Behavior with a Robotic Agent through Daily Activity Analysis. Int. J. Soc. Robot. 2019, 11, 437–462. [Google Scholar] [CrossRef] [Green Version]
  100. Gurpinar, C.; Uluer, P.; Akalin, N.; Kose, H. Sign Recognition System for an Assistive Robot Sign Tutor for Children. Int. J. Soc. Robot. 2020, 2, 355–369. [Google Scholar] [CrossRef]
  101. Cosar, S.; Fernandez-Carmona, M.; Agrigoroaie, R.; Pages, J.; Ferland, F.; Zhao, F.; Yue, S.; Bellotto, N.; Tapus, A. ENRICHME: Perception and Interaction of an Assistive Robot for the Elderly at Home. Int. J. Soc. Robot. 2020, 12, 779–805. [Google Scholar] [CrossRef] [Green Version]
  102. Al-Abdullah, A.; Al-Ajmi, A.; Al-Mutairi, A.; Al-Mousa, N.; Al-Daihani, S.; Karar, A.S.; alkork, S. Artificial Neural Network for Arabic Speech Recognition in Humanoid Robotic Systems. In Proceedings of the 2019 3rd International Conference on Bio-engineering for Smart Technologies (BioSMART), Paris, France, 24–26 April 2019; pp. 1–4. [Google Scholar] [CrossRef]
  103. Gao, J.; Galley, M.; Li, L. Neural Approaches to Conversational AI, Question Answering, 1052 Task-Oriented Dialogues and Social Chatbots; Now Foundations and Trends: Hanover, MA, USA, 2019. [Google Scholar]
  104. Dzakwan, G.; Purwarianti, A. Comparative Study of Topology and Feature Variants for Non-Task-Oriented Chatbot using Sequence to Sequence Learning. In Proceedings of the 5th International Conference on Advanced Informatics: Concept Theory and Applications (ICAICTA), Krabi, Thailand, 14–17 August 2018. [Google Scholar]
  105. Pham, K.T.; Nabizadeh, A.; Selek, S. Artificial Intelligence and Chatbots in Psychiatry. Psychiatr. Q. 2022, 93, 249–253. [Google Scholar] [CrossRef]
  106. Grassi, L.; Recchiuto, C.T.; Sgorbissa, A. Knowledge-Grounded Dialogue Flow Management for Social Robots and Conversational Agents. Int. J. Soc. Robot. 2022. [Google Scholar] [CrossRef]
  107. Briggs, G.; Williams, T.; Jackson, R.B.; Scheutz, M. Why and How Robots Should Say ‘No’. Int. J. Soc. Robot. 2022, 14, 323–339. [Google Scholar] [CrossRef]
  108. Brown, T.B.; Mann, B.; Ryder, N.; Subbiah, M.; Kaplan, J.; Dhariwal, P.; Neelakantan, A.; Shyam, P.; Sastry, G.; Askell, A.; et al. Language Models are Few-Shot Learners. Adv. Neural Inf. Process. Syst. 2020, 33, 1877–1901. [Google Scholar]
  109. Psychiatry.org—DSM. Available online: www.dsm5.org (accessed on 5 January 2022).
  110. Shuster, K.; Poff, S.; Chen, M.; Kiela, D.; Weston, J. Retrieval Augmentation Reduces Hallucination in Conversation. arXiv 2021, arXiv:2104.07567. [Google Scholar]
  111. Maynez, J.; Narayan, S.; Bohnet, B.; McDonald, R. On Faithfulness and Factuality in Abstractive Summarization. In Proceedings of the 58th Annual Meeting of the Association for Computational Linguistics; Association for Computational Linguistics, Online, 5–10 July 2020; pp. 1906–1919. [Google Scholar] [CrossRef]
  112. Hazourli, A.; Djeghri, A.; Salam, H.; Othmani Morgan, A. Multi-facial patches aggregation network for facial expression recognition and facial regions contributions to emotion display. Multimed. Tools Appl. 2021, 80, 13639–13662. [Google Scholar] [CrossRef]
  113. Aly, A.; Tapus, A. On Designing Expressive Robot Behavior: The Effect of Affective Cues on Interaction. SN Comput. Sci. 2020, 1, 314. [Google Scholar] [CrossRef]
  114. Boucenna, S.; Gaussier, P.; Andry, P.; Hafemeister, L. Imitation as a Communication Tool for Online Facial Expression Learning and Recognition. In Proceedings of the 2010 IEEE/RSJ International Conference on Intelligent Robots and Systems, Taipei, Taiwan, 18–22 October 2010; pp. 5323–5328. [Google Scholar] [CrossRef] [Green Version]
  115. Ashraf, A.B.; Lucey, S.; Cohn, J.F.; Chen, T.; Ambadar, Z.; Prkachin, K.M.; Solomon, P.E. The painful face–pain expression recognition using active appearance models. Image Vis. Comput. 2009, 27, 1788–1796. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  116. Faraj, Z.; Selamet, M.; Morales, C.; Torres, P.; Hossain, M.; Chen, B.; Lipson, H. Facially expressive humanoid robotic face. HardwareX 2021, 9, e00117. [Google Scholar] [CrossRef]
  117. SoftBank Robotics|Humanoid and Programmable Robots. Available online: https://www.softbankrobotics.com/emea/en (accessed on 5 March 2022).
  118. NAO the Humanoid and Programmable Robot|SoftBank Robotics. Available online: https://www.softbankrobotics.com/emea/en/nao (accessed on 5 March 2022).
  119. Frid, E.; Bresin, R. Perceptual Evaluation of Blended Sonification of Mechanical Robot Sounds Produced by Emotionally Expressive Gestures: Augmenting Consequential Sounds to Improve Non-verbal Robot Communication. Int. J. Soc. Robot. 2021, 14, 357–372. [Google Scholar] [CrossRef]
  120. Johnson, D.O.; Cuijpers, R.H. Investigating the Effect of a Humanoid Robot’s Head Position on Imitating Human Emotions. Int. J. Soc. Robot. 2018, 11, 65–74. [Google Scholar] [CrossRef]
  121. Pepper the Humanoid and Programmable Robot|SoftBank Robotics. Available online: https://www.softbankrobotics.com/emea/en/pepper (accessed on 5 March 2022).
  122. ASIMO by Honda|The World’s Most Advanced Humanoid Robot. Available online: https://asimo.honda.com/ (accessed on 5 March 2022).
  123. Sakagami, Y.; Watanabe, R.; Aoyama, C.; Matsunaga, S.; Higaki, N.; FujiMura, K. The intelligent ASIMO: System overview and integration. In Proceedings of the IEEE/RSJ International Conference on Intelligent Robots and Systems, Lausanne, Switzerland, 30 September–4 October 2002; Volume 3, pp. 247–2483. [Google Scholar] [CrossRef]
  124. Metta, G.; Natale, L.; Nori, F.; Sandini, G.; Vernon, D.; Fadiga, L.; von Hofsten, C.; Rosandwr, K.; Lopes, M.; Santos-Victor, J.; et al. The iCub humanoid robot: An open-systems platform for research in cognitive development. Neural Netw. 2010, 23, 1125–1134. [Google Scholar] [CrossRef]
  125. Al Moubayed, S.; Beskow, J.; Skantze, G.; Granstrom, B. Furhat: A back-projected human-like robot head for multiparty human-machine interaction. In Cognitive Behavioural Systems; Lecture Notes in Computer Science; Springer: Berlin/Heidelberg, Germany, 2012. [Google Scholar]
  126. McGinn, C.; Bourke, E.; Murtagh, A.; Cullinan, M.; Kelly, K. Exploring the application of design thinking to the development of service robot technology. In Proceedings of the ICRA2018 Workshop on Elderly Care Robotics-Technology and Ethics (WELCARO), Brisbane, Australia, 21–25 May 2018. [Google Scholar]
  127. OPSORO · GitHub. Available online: https://github.com/OPSORO (accessed on 5 March 2022).
  128. Vandevelde, C.; Wyffels, F.; Vanderborght, B.; Saldien, J. Do-It-Yourself Design for Social Robots: An Open-Source Hardware Platform to Encourage Innovation. IEEE Robot. Autom. Mag. 2017, 24, 86–94. [Google Scholar] [CrossRef] [Green Version]
  129. Reeti|Robot. Available online: https://robots.nu/en/robot/reeti (accessed on 12 June 2022).
  130. PARO Therapeutic Robot. Available online: http://www.parorobots.com/index.asp (accessed on 5 March 2022).
  131. Paro—ROBOTS: Your Guide to the World of Robotics. Available online: https://robots.ieee.org/robots/paro/ (accessed on 5 March 2022).
  132. Tega Robot. Available online: https://www.wevolver.com/wevolver.staff/tega.robot (accessed on 5 March 2022).
  133. Overview <Huggable: A Social Robot for PEDIATRIC Care—MIT Media Lab. Available online: https://www.media.mit.edu/projects/huggable-a-social-robot-for-pediatric-care/overview/ (accessed on 5 March 2022).
  134. Social Robots can Benefit Hospitalized Children|MIT News|Massachusetts Institute of Technology. Available online: https://news.mit.edu/2019/social-robots-benefit-sick-children-0626 (accessed on 15 March 2022).
  135. Sophia—Hanson Robotics. Available online: https://www.hansonrobotics.com/sophia/ (accessed on 25 January 2022).
  136. Park, S.; Lee, H.; Hanson, D.; Oh, P.Y. Sophia-Hubo’s Arm Motion Generation for a Handshake and Gestures. In Proceedings of the 15th International Conference on Ubiquitous Robots (UR), Honolulu, HI, USA, 26–30 June 2018. [Google Scholar]
  137. The Furhat Robot|Furhat Robotics. Available online: https://furhatrobotics.com/furhat-robot/ (accessed on 25 January 2022).
  138. Aibo. Available online: https://us.aibo.com/ (accessed on 17 March 2022).
  139. Aibo—ROBOTS: Your Guide to the World of Robotics. Available online: https://robots.ieee.org/robots/aibo2018/ (accessed on 17 March 2022).
  140. Schellin, H.; Oberley, T.; Patterson, K.; Kim, B.; Haring, K.S.; Tossell, C.C.; Phillips, E.; Visser, E.J.d. Man’s New Best Friend? Strengthening Human-Robot Dog Bonding by Enhancing the Doglikeness of Sony’s Aibo. In Proceedings of the 2020 Systems and Information Engineering Design Symposium (SIEDS), Charlottesville, VA, USA, 24 April 2020; pp. 1–6. [Google Scholar] [CrossRef]
  141. Nao—ROBOTS: Your Guide to the World of Robotics. Available online: https://robots.ieee.org/robots/nao/ (accessed on 5 March 2022).
  142. Honda Global|ASIMO. Available online: https://global.honda/innovation/robotics/ASIMO.html (accessed on 5 March 2022).
  143. Asimo—ROBOTS: Your Guide to the World of Robotics. Available online: https://robots.ieee.org/robots/asimo/ (accessed on 5 March 2022).
  144. Giraff. Available online: https://robots.nu/en/robot/giraff-telepresence-robot (accessed on 15 March 2022).
  145. Westlund, J.K.; Lee, J.J.; Plummer, L.; Faridi, F.; Gray, J.; Berlin, M.; Quintus-Bosz, H.; Hartmann, R.; Hess, M.; Dyer, S.; et al. Tega: A social robot. In Proceedings of the 2016 11th ACM/IEEE International Conference on Human-Robot Interaction (HRI), Christchurch, New Zealand, 7–10 March 2016; p. 561-561. [Google Scholar] [CrossRef] [Green Version]
  146. Sooyeon, J.; Dos Santos, K.; Graca, S.; O’Connell, B.; Anderson, L.; Stenquist, N.; Fitzpatrick, K.; Goodenough, H.; Logan, D.; Weinstock, P.; et al. Designing a Socially Assistive Robot for Pediatric Care. In Proceedings of the 14th International Conference on Interaction Design and Children, Medford, MA, USA, 21–25 June 2015. [Google Scholar]
  147. Said, S.; AlKork, S.; Beyrouthy, T.; Abdrabbo, M.F. Wearable bio-sensors bracelet for driveras health emergency detection. In Proceedings of the 2017 2nd International Conference on Bio-Engineering for Smart Technologies (BioSMART), Paris, France, 30 August–1 September 2017; pp. 1–4. [Google Scholar] [CrossRef]
  148. Said, S.; Boulkaibet, I.; Sheikh, M.; Karar, A.S.; Alkork, S.; Nait-ali, A. Machine-Learning-Based Muscle Control of a 3D-Printed Bionic Arm. Sensors 2020, 20, 3144. [Google Scholar] [CrossRef] [PubMed]
  149. Zhe, C.; Simon, T.; Wei, S.E.; Sheikh, Y. Realtime Multi-Person 2D Pose Estimation using Part Affinity Fields. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Honolulu, HI, USA, 21–26 July 2017. [Google Scholar]
  150. Pennington, J.; Socher, R.; Manning, C.D. Glove: Global Vectors for Word Representation. In Proceedings of the Empirical Methods in Natural Language Processing (EMNLP), Doha, Qatar, 25–29 October 2014. [Google Scholar]
  151. Dibia, V. HandTrack: A library for prototyping real-time hand tracking interfaces using convolutional neural networks. Github Repos. 2017, 3, 6. [Google Scholar]
  152. Roshdy, A.; Karar, A.S.; Al-Sabi, A.; Barakeh, Z.A.; El-Sayed, F.; Alkork, S.; Beyrouthy, T.; Nait-ali, A. Towards Human Brain Image Mapping for Emotion Digitization in Robotics. In Proceedings of the 2019 3rd International Conference on Bio-engineering for Smart Technologies (BioSMART), Paris, France, 24–26 April 2019; pp. 1–5. [Google Scholar] [CrossRef]
  153. Roshdy, A.; Al Kork, S.; Karar, A.; Al Sabi, A.; Al Barakeh, Z.; ElSayed, F.; Beyrouthy, T.; NAIT-ALI, A. Machine Empathy: Digitizing Human Emotions. In Proceedings of the 2021 International Symposium on Electrical, Electronics and Information Engineering, Seoul, Korea, 19–21 February 2021; Association for Computing Machinery: New York, NY, USA, 2021; pp. 307–311. [Google Scholar] [CrossRef]
  154. Sakhineti, M.; Jayabalan, S. Design and Fabrication of SHRALA: Social Humanoid Robot Based on Autonomous Learning Algorithm. In Proceedings of the Third International Conference on Computing and Network Communications (CoCoNet), Trivandrum, India, 18–21 December 2019. [Google Scholar]
  155. Tiansong, L.; Feng, G.; Yilong, Y. Design of Low-cost Desktop Robot Based on 3D Printing Technology and Open-source Control System. In Proceedings of the IEEE 3rd Informatin Technology, Networking, Electronic and Automation Control Conference (ITNEC), Chengdu, China, 15–17 March 2019. [Google Scholar]
  156. Potnuru, A.; Jafarzadeh, M.; Tadese, Y. 3D printed dancing humanoid robot “Buddy” for homecare. In Proceedings of the IEEE International Conference on Automation Science and Engineering (CASE), Fort Worth, TX, USA, 21–25 August 2016. [Google Scholar]
  157. Al-Omary, A.; Akram, M.M.; Dhamodharan, V. Design and Implementation of Intelligent Socializing 3D Humanoid Robot. In Proceedings of the International Conference on Innovation and Intelligence for Informatics, Computing, and Technologies (3ICT), Virtual, 29–30 September 2021. [Google Scholar]
  158. Romeo, J. Why Additive Manufacturing and 3D Printing Benefits Robot Creators. Robot. Bus. Rev. 2019. Available online: https://www.roboticsbusinessreview.com/wp-content/uploads/2019/04/RBR-AdditiveManufacturing-RobotCreators-Final.pdf (accessed on 15 March 2022).
  159. Saini, J.; Chew, E. Recent Trends in Mechatronics Towards Industry; Springer: Berlin/Heidelberg, Germany, 2021; pp. 275–287. [Google Scholar]
  160. Cheng, H.; Ji, G. Design and implementation of a low cost 3D printed humanoid robotic platform. In Proceedings of the 2016 IEEE International Conference on Cyber Technology in Automation, Control, and Intelligent Systems (CYBER), Chengdu, China, 19–22 June 2016; pp. 86–91. [Google Scholar] [CrossRef]
  161. Sidher, A.; Shen, Y. Improving a 3D-printed artificial anthropomorphic hand using the human hand model. In Proceedings of the 2017 IEEE International Conference on Real-time Computing and Robotics (RCAR), Okinawa, Japan, 14–18 July 2017; pp. 739–744. [Google Scholar] [CrossRef]
  162. Berra, R.; Setti, F.; Cristani, M. Berrick: A low-cost robotic head platform for human-robot interaction. In Proceedings of the 2019 IEEE International Conference on Systems, Man and Cybernetics (SMC), Bari, Italy, 6–9 October 2019; pp. 559–566. [Google Scholar] [CrossRef]
  163. Delda, R.N.M.; Basuel, R.B.; Hacla, R.P.; Martinez, D.W.C.; Cabibihan, J.J.; Dizon, J.R.C. 3D Printing Polymeric Materials for Robots with Embedded Systems. Technologies 2021, 9, 82. [Google Scholar] [CrossRef]
  164. Netzev, M.; Houbre, Q.; Airaksinen, E.; Angleraud, A.; Pieters, R. Many Faced Robot - Design and Manufacturing of a Parametric, Modular and Open Source Robot Head. In Proceedings of the 16th International Conference on Ubiquitous Robots (UR), Jeju, Korea, 24–27 June 2019. [Google Scholar]
  165. Harrison, A.M.; Xu, W.M.; Trafton, J.G. User-Centered Robot Head Design: A Sensing Computing Interaction Platform for Robotics Research (SCIPRR). In Proceedings of the 13th ACM/IEEE International Conference on Human-Robot Interaction (HRI), Chicago, IL, USA, 5–8 March 2018. [Google Scholar]
  166. Chen, H.; Mooring, B.; Stern, H. Dynamic wafer handling process in semiconductor manufacturing. In Proceedings of the IEEE International Conference on Robotics and Biomimetics (ROBIO), Karon Beach, Thailand, 7–11 December 2011. [Google Scholar]
  167. How Robotics Have Revolutionized Semiconductor Manufacturing. Available online: https://www.modutek.com/how-robotics-have-revolutionized-semiconductor-manufacturing/ (accessed on 24 January 2022).
  168. Ruiz, P.Y. Semiconductor solutions for robotics, a special focus on drives technologies. Infineon Technol. AG 2018. [Google Scholar]
  169. GPUs and the Future of Robotics—Robotics Business Review. Available online: https://www.roboticsbusinessreview.com/rbr/gpus-and-the-future-of-robotics/ (accessed on 15 March 2022).
  170. Yan, F.; Tran, D.M.; He, H. Robotic Understanding of Object Semantics by Referringto a Dictionary. Int. J. Soc. Robot. 2020, 12, 1251–1263. [Google Scholar] [CrossRef]
  171. Kim, M.; Kwon, T.; Kim, K. Can Human–Robot Interaction Promote the Same Depth of Social Information Processing as Human–Human Interaction? Int. J. Soc. Robot. 2017, 10, 33–42. [Google Scholar] [CrossRef]
  172. Savery, R.; Rose, R.; Weinberg, G. Finding Shimi’S Voice: Fostering Human-Robot Communication with Music and a Nvidia Jetson TX2. In Proceedings of the 17th Linux Audio Conference, Stanford, CA, USA, 23–26 March 2019. [Google Scholar]
  173. NVIDIA Isaac SDK|NVIDIA Developer. Available online: https://developer.nvidia.com/isaac-sdk (accessed on 15 March 2022).
  174. SoftBank Robotics Documentation. Available online: http://doc.aldebaran.com/2-1/family/robots/motherboard_robot.html#robot-motherboard (accessed on 12 March 2022).
  175. SoftBank Robotics Documentation. Available online: http://doc.aldebaran.com/2-4/family/pepper_technical/motherboard_pep.html (accessed on 12 March 2022).
  176. Sophia—ROBOTS: Your Guide to the World of Robotics. Available online: https://robots.ieee.org/robots/sophia/ (accessed on 5 March 2022).
  177. ARI—PAL Robotics: Leading Service Robotics. Available online: https://pal-robotics.com/robots/ari/ (accessed on 13 March 2022).
  178. QTrobot—ROBOTS: Your Guide to the World of Robotics. Available online: https://robots.ieee.org/robots/qtrobot/ (accessed on 15 March 2022).
  179. Furhat Platform; Technical Report; Furhat Robotics: Stockholm, Sweden, 2021.
  180. ROS: Home. Available online: https://www.ros.org/ (accessed on 5 March 2022).
  181. #Tags. Available online: https://robots.ros.org/tags/#social (accessed on 10 March 2022).
  182. QTrobot. Available online: https://robots.ros.org/qtrobot/ (accessed on 15 March 2022).
  183. ARI. Available online: https://robots.ros.org/ari/ (accessed on 15 March 2022).
  184. Fu, G.; Zhang, X. ROSBOT: A low-cost autonomous social robot. In Proceedings of the 2015 IEEE International Conference on Advanced Intelligent Mechatronics (AIM), Busan, Korea, 7–11 July 2015; pp. 1789–1794. [Google Scholar] [CrossRef]
  185. Infantino, I.; Augello, A.; Maniscalto, U.; Pilato, G.; Storniolo, P.; Vella, F. A ROS Architecture for a Cognitive Social Robot. 2018. Available online: https://www.semanticscholar.org/paper/A-ROS-architecture-for-a-Cognitive-Social-Robot-Infantino-Augello/306f90831a6db3e9f425a3c9a5cbdc3ebbd7b7a6 (accessed on 15 March 2022).
  186. Martín, F.; Rodríguez Lera, F.J.; Ginés, J.; Matellán, V. Evolution of a Cognitive Architecture for Social Robots: Integrating Behaviors and Symbolic Knowledge. Appl. Sci. 2020, 10, 6067. [Google Scholar] [CrossRef]
  187. Adam, C.; Johal, W.; Pellier, D.; Fiorino, H.; Pesty, S. Social Human-Robot Interaction: A New Cognitive and Affective Interaction-Oriented Architecture. In Proceedings of the International Conference on Social Robotics, Kansas City, MO, USA, 1–3 November 2016; Volume 9979, pp. 253–263. [Google Scholar] [CrossRef] [Green Version]
  188. UXA-90 Humanoid Robot|Roobuilder Co., Ltd. Available online: https://www.robobuilder.net/uxa-90 (accessed on 2 June 2022).
  189. SoftBank Robotics Documentation. Available online: http://doc.aldebaran.com/2-5/index_dev_guide.html# (accessed on 12 March 2022).
  190. OpenNAO—NAO OS—NAO Software 1.14.5 Documentation. Available online: http://doc.aldebaran.com/1-14/dev/tools/opennao.html (accessed on 12 April 2022).
  191. Tsardoulias, E.; Mitkas, P. Robotic frameworks, architectures and middleware comparison. arXiv 2017, arXiv:1711.06842. [Google Scholar]
  192. Mohamed, Y.; Lemaignan, S. ROS for Human-Robot Interaction. In Proceedings of the 2021 IEEE/RSJ International Conference on Intelligent Robots and Systems (IROS), Prague, Czech Republic, 27 September–1 October 2021. [Google Scholar]
  193. People—ROS Wiki. Available online: https://wiki.ros.org/people (accessed on 10 March 2022).
  194. Cob_People_Detection—ROS Wiki. Available online: http://wiki.ros.org/cob_people_detection (accessed on 10 March 2022).
  195. Fong, T.; Kunz, C.; Hiatt, L.M.; Bugajska, M. The Human-Robot Interaction Operating System. In Proceedings of the 1st ACM SIGCHI/SIGART Conference on Human-Robot Interaction, Salt Lake City, UT, USA, 2–3 March 2006; Association for Computing Machinery: New York, NY, USA, 2006; pp. 41–48. [Google Scholar] [CrossRef] [Green Version]
  196. Trafton, G.; Hiatt, L.; Harrison, A.; Tanborello, F.; Khemlani, S.; Schultz, A. ACT-R/E: An embodied cognitive architecture for human-robot interaction. J. Hum. Robot Interact. 2013, 2, 30–55. [Google Scholar] [CrossRef]
  197. Lim, V.; Rooksby, M.; Cross, E.S. Social Robots on a Global Stage: Establishing a Role for Culture During Human–Robot Interaction. Int. J. Soc. Robot. 2021, 13, 1307–1333. [Google Scholar] [CrossRef]
  198. Fortunati, L.; Manganelli, A.M.; Hoflich, J.; Ferrin, G. Exploring the Perceptions of Cognitive and Affective Capabilities of Four, Real, Physical Robots with a Decreasing Degree of Morphological Human Likeness. Int. J. Soc. Robot. 2021. [Google Scholar] [CrossRef]
  199. Dautenhahn, K.; Woods, S.; Kaouri, C.; Walters, M.; Koay, K.; Werry, I. What is a robot companion-Friend, assistant or butler? In Proceedings of the 2005 IEEE/RSJ International Conference on Intelligent Robots and Systems, Edmonton, AB, Canada, 2–6 August 2005; pp. 1192–1197. [Google Scholar] [CrossRef] [Green Version]
  200. Bartneck, C.; Kulic, D.; Croft, E.; Zoghbi, S. Measurement Instruments for the Anthropomorphism, Animacy, Likeability, Perceived Intelligence, and Perceived Safety of Robots. Int. J. Soc. Robot. 2009, 1, 71–81. [Google Scholar] [CrossRef] [Green Version]
  201. Heerink, M.; Krose, B.; Evers, V.; Wielinga, B. Assessing Acceptance of Assistive Social Agent Technology by Older Adults: The Almere Model. Int. J. Soc. Robot. 2010, 2, 361–375. [Google Scholar] [CrossRef] [Green Version]
  202. Venkatesh, V.; Morris, M.G.; Davis, G.B.; Davis, F.D. User Acceptance of Information Technology: Toward a Unified View. Manag. Inf. Syst. Q. 2003, 27, 425–478. [Google Scholar] [CrossRef] [Green Version]
  203. Anzalone, S.; Boucenna, S.; Ivaldi, S.; Chetouani, M. Evaluating the Engagement with Social Robots. Int. J. Soc. Robot. 2015, 7, 465–478. [Google Scholar] [CrossRef]
  204. Rueben, M.; Elprama, S.A.; Chrysostomou, D.; Jacobs, A. Introduction to (re)using questionnaires in human-robot interaction research. In Human-Robot Interaction: Evaluation Methods and Their Standardization; Jost, C., Le Pévédic, B., Belpaeme, T., Bethel, C., Chrysostomou, D., Crook, N., Grandgeorge, M., Mirnig, N., Eds.; Springer International Publishing: Cham, Switzerland, 2020; pp. 125–144. [Google Scholar] [CrossRef]
  205. Avioz-Sarig, O.; Olatunju, S.; Sarne-Fleischmann, V.; Edan, Y. Robotic System for Physical Training of Older Adults. Int. J. Soc. Robot. 2020, 13, 1109–1124. [Google Scholar] [CrossRef] [PubMed]
  206. White Paper on Artificial Intelligence: A European Approach to Excellence and Trust. Available online: https://ec.europa.eu/info/publications/white-paper-artificial-intelligence-european-approach-excellence-and-trust_en (accessed on 4 July 2022).
  207. Mazzeo, G.; Staffa, M. TROS: Protecting Humanoids ROS from Privileged Attackers. Int. J. Soc. Robot. 2020, 12, 827–841. [Google Scholar] [CrossRef]
  208. Yaacoub, J.P.A.; Noura, H.N.; Salman, O.; Chehab, A. Robotics cyber security: Vulnerabilities, attacks, countermeasures, and recommendations. Int. J. Inf. Secur. 2022, 21, 115–158. [Google Scholar] [CrossRef]
  209. Miller, J.; Williams, A.; Perouli, D. A Case Study on the Cybersecurity of Social Robots. In Proceedings of the Companion of the 2018 ACM/IEEE International Conference on Human-Robot Interaction, Chicago, IL, USA, 5–8 March 2018; pp. 195–196. [Google Scholar] [CrossRef]
  210. Fortunati, L.; Sorrentino, A.; Fiorini, L.; Cavallo, F. The Rise of the Roboid. Int. J. Soc. Robot. 2021, 13, 1457–1471. [Google Scholar] [CrossRef]
Figure 1. Points addressed in the survey.
Figure 1. Points addressed in the survey.
Robotics 11 00075 g001
Figure 2. A Double telepresence robot in the American University of the Middle east.
Figure 2. A Double telepresence robot in the American University of the Middle east.
Robotics 11 00075 g002
Figure 3. Softbank Robotics Nao (left) and Pepper (right) robots at the American University of the Middle East.
Figure 3. Softbank Robotics Nao (left) and Pepper (right) robots at the American University of the Middle East.
Robotics 11 00075 g003
Figure 4. Robot head designed for 3D printing with rich expressiveness. Reproduced with permission from [K Youssef, S Said, T Beyrouthy, S Alkork], [BioSMART] ; published by [IEEE], [2022] [87].
Figure 4. Robot head designed for 3D printing with rich expressiveness. Reproduced with permission from [K Youssef, S Said, T Beyrouthy, S Alkork], [BioSMART] ; published by [IEEE], [2022] [87].
Robotics 11 00075 g004
Figure 5. The humanoid robots UXA-90 with different body postures at the American University of the Middle East.
Figure 5. The humanoid robots UXA-90 with different body postures at the American University of the Middle East.
Robotics 11 00075 g005
Table 1. Examples of applications and experiments where social robots were used, with the respective end users.
Table 1. Examples of applications and experiments where social robots were used, with the respective end users.
StudyRobotResearch Goal/ApplicationTargets/End Users
Shiarlis et al. [15]TeresaTelepresence-behavior in interactionUnspecified
Shiarlis et al. [22]TeresaTelepresence-participation in social eventsElderly
Niemelä et al. [23]DoubleTelepresence-communication with family membersElderly
Zhang et al. [24]UnspecifiedTelepresence-Control with eye gazePersons with motor disabilities
Hood et al. [25]NaoHandwriting learningChildren
Engwall et al. [26]FurhatSecond language learningVarious
Kanero et al. [27]NaoSecond language learningAdults
Shimaya et al. [28]CommUCommunication in lecturesStudents and lecturers
Reyes et al. [29]NaoAssistance in classStudents and lecturers
Vogt et al. [30]NaoSecond language learningChildren
Schicchi et al. [21]PepperVocabulary enhancementChildren
Obayashi et al. [31]Mon-chanCare in nursing homesElderly
McGinn et al. [31]StevieCare in a care facilityResidents and staff
Luperto et al. [32]Giraff-XAssistance at homeElders
Ismail et al. [33]LUCAAnalysis of attentionChildren with cognitive impairment
van der Putte et al. [20]PepperData acquisitionHospitalized patients
Schrum et al. [34]PepperEncouraging physical exerciseDementia patients
Moharana et al. [35]Different robots designedDesigning robots for dementia caregivingDementia caregiver support groups
Anzalone et al. [36]NaoEnvironment perceptionChildren with ASD
Huijnen et al. [37]KasparMaking contact and catching attentionChildren with ASD
Taheri et al. [38]NaoBody gesture imitationChildren with ASD
Chung [39]NaoEnhancement of social skillsChildren with ASD
Striepe et al. [40]ReetiImplementing behaviors of a robot storytellerPersons aged from 18 to 30
Desideri et al. [41]NaoStudying gaze aversion in human-robot interactionChildren
Filippini et al. [42]Mio AmicoAssessing the emotional state of robot interlocutorChildren
Uluer et al. [43]PepperAssistance for hearing disabilitiesChildren
Castellano et al. [19]PepperImproving attitudes toward recyclingChildren
lio et al. [44]ASIMOGuidance for a science museumVarious, museum visitors
Shi et al. [45]RobovieFlyer distributionVarious, pedestrians in a shopping mall
Table 2. Examples of robots used in social contexts and some of their characteristics.
Table 2. Examples of robots used in social contexts and some of their characteristics.
RobotAppearanceHeight (cm)D.o.F.Features
Nao [118,141]Humanoid5825Touch sensors, directional microphones and speakers, 2D cameras, embedded speech recognition and dialogue, programmable, etc.
Pepper [121]Humanoid12020Touch sensors and microphones and speakers, 2D and 3D cameras, embedded speech recognition and dialogue, programmable, etc.
Asimo [142,143]Humanoid13057Different proprioceptive and exteroceptive sensors for motion tracking, obstacle detection, image and sound acquisition, etc.
Kaspar [77]Humanoid5522Color camera, Kinect and IMU sensor, semi-autonomous, Wi-Fi/Ethernet connection.
TERESA [15]UnspecifiedUnspecifiedUnspecifiedsemi-autonomous navigation, different proprioceptive and exteroceptive sensors for motion tracking, obstacle detection, image and sound acquisition, etc.
Furhat [125]Human-like face413Onboard camera and microphones, speech recognition and synthesis, eye contact, etc.
Sophia [135]Humanoid16783several cameras, audio localization array, complex and emotional expressions, natural language processing, visual tracking, etc.
Giraff [144]TelepresenceUnspecifiedUnspecifiedLCD screen, remote control, provides audio and visual cues to the user, data collection for health professionals, etc.
Paro [130,131]Animal (Seal)UnspecifiedUnspecifiedDifferent kinds of sensors, learning to behave as the user prefers, can move its head and legs, etc.
Tega [132,145]Unspecified34.545Microphone, camera and accelerometer sensors, autonomous or remote operation, ability to generate behaviors and facial expressions, etc.
Huggable [133,146]Teddy bearN/A12Can perceive physical touch, different other sensors, controlled by an application on a smart phone, teleoperation interface, etc.
Aibo [138,139]Dog-like29.322Different sensors, cameras, microphones, can recognize faces and voices, capable of simultaneous localization and mapping, etc.
Table 3. Examples of social robots and some of their processing characteristics.
Table 3. Examples of social robots and some of their processing characteristics.
RobotProcessorProcessor FeaturesRAMGPU
Nao V5 & V4 [174]ATOM Z5301.6 GHZ clock speed1 GBNone
Pepper V1.6 [175]ATOM E38451.91 GHZ clock speed, quadcore4 GB DDR3None
Sophia [176]Intel i73 GHZ32 GBintegrated GPU
ARI [177]Intel i5/i7Unspecified8 GB/16 GBNVIDIA Jetson TX2
QTrobot [178]Intel NUC i7Unspecified16 GBNone
Furhat [179]Intel i5up to 3.4 GHz8 GBNone
Giraff [144]Intel i7Unspecified8 GBNVIDIA Jetson TX2 in [32]
Asimo [122]Unspecified
Huggable [146]Computational power of an Android phone
Shimi in [172]ARMQuadcore8 GBNVIDIA Jetson TX2
Aibo [139]Qualcomm Snapdragon 82064-bit Quadcore4 GBNone
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Share and Cite

MDPI and ACS Style

Youssef, K.; Said, S.; Alkork, S.; Beyrouthy, T. A Survey on Recent Advances in Social Robotics. Robotics 2022, 11, 75. https://doi.org/10.3390/robotics11040075

AMA Style

Youssef K, Said S, Alkork S, Beyrouthy T. A Survey on Recent Advances in Social Robotics. Robotics. 2022; 11(4):75. https://doi.org/10.3390/robotics11040075

Chicago/Turabian Style

Youssef, Karim, Sherif Said, Samer Alkork, and Taha Beyrouthy. 2022. "A Survey on Recent Advances in Social Robotics" Robotics 11, no. 4: 75. https://doi.org/10.3390/robotics11040075

APA Style

Youssef, K., Said, S., Alkork, S., & Beyrouthy, T. (2022). A Survey on Recent Advances in Social Robotics. Robotics, 11(4), 75. https://doi.org/10.3390/robotics11040075

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop