Next Article in Journal
Synthesis and Surface Modification of Nanostructured F-Doped ZnO: Toward a Transducer for Label-Free Optical Biosensing
Previous Article in Journal
The Zerotrope, a Dynamic Holographic Display: Design and Implementation
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Optimizing Android Facial Expressions Using Genetic Algorithms

1
Intelligent Robot Engineering, Korea University of Science and Technology (UST), Ansan 15588, Korea
2
Robotics R&D Group, Korea Institute of Industrial Technology (KITECH), Ansan 15588, Korea
*
Author to whom correspondence should be addressed.
Appl. Sci. 2019, 9(16), 3379; https://doi.org/10.3390/app9163379
Submission received: 30 July 2019 / Accepted: 14 August 2019 / Published: 16 August 2019
(This article belongs to the Section Computing and Artificial Intelligence)

Abstract

:
Because the internal structure, degree of freedom, skin control position and range of the android face are different, it is very difficult to generate facial expressions by applying existing facial expression generation methods. In addition, facial expressions differ among robots because they are designed subjectively. To address these problems, we developed a system that can automatically generate robot facial expressions by combining an android, a recognizer capable of classifying facial expressions and a genetic algorithm. We have developed two types (older men and young women) of android face robots that can simulate human skin movements. We selected 16 control positions to generate the facial expressions of these robots. The expressions were generated by combining the displacements of 16 motors. A chromosome comprising 16 genes (motor displacements) was generated by applying real-coded genetic algorithms; subsequently, it was used to generate robot facial expressions. To determine the fitness of the generated facial expressions, expression intensity was evaluated through a facial expression recognizer. The proposed system was used to generate six facial expressions (angry, disgust, fear, happy, sad, surprised); the results confirmed that they were more appropriate than manually generated facial expressions.

1. Introduction

A social robot can deliver information required for everyday life to humans and can interact with humans via two-way communication. Natural interaction with humans is enabled only by combining research on topics such as robot control, object recognition and dialog engines [1,2,3]. Recently, a hotel in Japan introduced an android to confirm the reservation information of clients and guide them to their rooms [4,5]. This system combines interactive technology that responds to clients, vision/recognition technology that can recognize an approaching person and memorize their appearance, voice recognition and dialog technology that can understand what people say and have a conversation and robot control technology that can inform people of their room location via hand gestures.
The appearance of social robots takes various forms, depending on their roles and purposes. In homes, domestic robots that perform the role of a personal secretary have been introduced [6,7]. These robots play music, read either texts or schedules and show them on screens upon receiving voice commands from their owners. Unlike home service robots, robots that guide people through spaces, such as exhibitions, are relatively larger and they have additional mobile functions and manipulator technology. These robots stretch their arms to lead people through areas or communicate the location of an object and can directly guide people to a requested location by using mobile technology [8,9,10]. In particular, when used in roles such as hotel receptionist and announcer positions, androids present a human-like appearance to approach people in a friendly manner. Androids imitate humans in terms of their skin appearance, behavior and facial expressions, and are the most similar type of robots to humans [11,12].
Behavioral expression technology is currently aimed at enabling androids to exactly imitate human facial expressions and gestures [13,14,15]. The behavioral expressions of robots can largely be classified as facial expressions, gestures, lip syncing and eye contact [16,17], which also vary depending on if there are one or multiple interaction targets. These behavioral expressions serve different functional purposes; the facial expression of a robot can express emotion, while lip sync technology can improve the delivery of information in a conversational situation [18,19,20]. In addition, robots should be able to gaze naturally at people when interacting with them and maintain natural eye contact when interacting with a large number of people. Finally, gestures facilitate the delivery of more accurate information during a conversation with humans [21,22,23]. For example, when a robot guides people to a location, these people can recognize the place more accurately if the robot gives directions with its hand. Additionally, a variety of behavioral expressions enables people to communicate with robots more comfortably.
Facial expressions are important nonverbal behavioral expressions that can convey thoughts and feelings to others effectively. According to Mehrabian, nonverbal behavior such as expressions and attitudes are more important for communication than words [24]. He stated that, for communication, 55% of information is conveyed through body language, 38% through tone of voice and 7% through words. Therefore, facial expressions, which are nonverbal behavior, are crucial for communication. Additionally, Ekman proved that the six basic expressions of human beings are similar regardless of religion, region and culture. This indicates that basic emotions can be communicated through facial expressions even between different cultures or languages [25].
Research is underway to analyze faces in 3D to model them in a virtual environment. Vezzetti proposed a new approach to 3D face analysis using the landmarking method [26,27]. Marcolin proposed 105 geometrical descriptors for face analysis. The descriptors they proposed can be Unlike a virtual face, which directly controls virtual skin, the robot must control a motor to move real skin. Because the correlation of internal wiring and the intensity of skin differ among robots, it is very difficult to predict the motion of the skin and control the motor. In other words, the robot is deformed as the motor moves. It is difficult to use the 3D facial analysis data because the motion of the motor due to the deformation of the skin depends upon the characteristics of the robot.
Thus, the facial expressions of an android are important for delivering information from the robot to humans. Many studies have been conducted on the generation of natural robot facial expressions. P. K. Dick, an android developed by Hanson Robotics, was trained to generate facial expressions through the inverse nonlinear mapping of the human facial feature space to actuator space [28]. ERICA, an android developed by ATR, achieved behavioral expressions that could be utilized in polite situations. The developers generated natural smiling behavioral expressions by considering the android’s facial expressions, as well as movements of the eyes, neck and waist [29]. Moreover, Lin proposed a mechanism that can generate various facial expressions by using the minimum number of actuators in consideration of production cost [30]. Hashimoto developed a facial expression generation system capable of imitating human muscle structures based on the EMG signals used for facial expressions [31]. Furthermore, Huang et al. created an RNN-based forward kinematics model by measuring facial geometric deformation features from human facial expression sequence data and then generated facial expressions by producing an IK solver and applying it to an android [32].
As android faces have different presentational forms, internal mechanism structures, numbers of motors and muscle expression mechanisms, their facial expression generation methods also differ, making it difficult to mathematically define these generation methods. Most android developers generate facial expressions by controlling the motors and presenting subjective interpretations of facial expression characteristics. Thus, the results will differ depending on the skills of the developer and the same facial expression may be interpreted differently by different people. Therefore, an index is required to objectively evaluate the facial expressions generated by an android.
To address these problems, this study proposes a system for androids that can automatically generate the facial expressions of an android based on a genetic algorithm. Because the face of the android resembles a human being, it is possible to utilize a human facial expression recognizer directly. This system comprises a facial expression recognizer, an android capable of generating facial expressions and a genetic algorithm. The system features a number of prerequisites. First, the facial expression recognizer should be able to identify the android’s facial expressions as effectively as it can recognize those of a human. Second, the android must be able to generate facial expressions by using a combination of motor displacements inside the mechanism. Finally, to utilize the genetic algorithm, the motor displacements must be able to be expressed as genes to evaluate the fitness of the facial expressions using recognizers. By doing so, the android’s facial expressions can evolve through processes such as selection, crossover and mutation, thereby generating natural facial expressions.
The rest of this paper is structured as follows. In Section 2, we describe the EveR series of androids, to which the proposed system is applied. Section 3 describes how to encode facial expressions into genes and how to measure the fitness of the generated facial expressions, as well as our complete optimization system that comprises the android, the facial expression recognizer and the genetic algorithm for facial expression optimization. In Section 4, we describe the system for implementing the proposed algorithm and analyze the evolution and fitness of facial expressions using this system. Finally, we present conclusions in Section 5.

2. EveR Androids

Androids are humanoid-type robots with similar appearances, skin and eyes to human beings. EveR, an android capable of expressing human movements, was developed by KITECH [12]. The name was generated using the name of the “first” woman, Eve and the first letter of “robot”. EveR is divided into three sections for control: the head, upper body and lower body. Twenty-three motors are arranged in layers in the head and connected to major control positions through wires to generate skin movements. Human behavioral expressions, such as eye blinking, eye movements, lip shape formation and other facial expressions, are imitated through a combination of movements of the 23 motors. The waist and neck have three degrees of freedom to allow EveR to rotate its waist and neck like humans. Finally, both arms can express a range of gestures by using 12 motors [13,14].
According to Mori, who proposed the uncanny valley theory, human beings have a more favorable impression of robots when they are more similar to humans. Conversely, they have feelings of extreme rejection when the behavioral expressions of robots are awkward [33]. Therefore, robots resembling humans must act as much like humans as possible to reduce this feeling of rejection and enhance favorable impressions, necessitating behavioral expressions with natural presentations.
EveR has been utilized in different exhibitions and performances (Figure 1) [18]. As the majority of exhibitions are based on predictable scenarios, predefined dialogs and gestures must be prepared in advance. For this, various gesture expressions using the upper body of the android (arms, neck and waist), as well as behavioral expressions using the head (lip syncing and facial expressions), must also be generated naturally in advance. For gestures, the motion of an actor is captured using marker-based motion capture equipment. Robot joint angles are extracted from these data and utilized for the android gestures. Unlike upper body movements, the facial expressions and lip shapes are generated by controlling individual motors.

3. Facial Expression Generation Based on Genetic Algorithms

3.1. Encoding of Facial Expressions

Genetic algorithms are optimization algorithms created by Holland in 1975 that imitate the evolutionary process of natural ecosystems [34], where genes evolve adaptably for survival through the processes of selection, crossover and mutation. Genetic algorithms that model this process from an engineering perspective can be applied to various problems that are difficult to define mathematically, such as project scheduling problems, cost problems for constructing network environments and optimization problems of multipurpose buildings, to obtain optimal solutions.
In this study, a facial expression optimization system based on a genetic algorithm was developed for facial expression generation methods that are difficult to define mathematically. DEAP, a Python-based open source library, was employed as the genetic algorithm [35]. The DEAP framework is capable of rapidly and simply testing various evolutionary computations, such as genetic algorithms (GA), generic programming (GP) and evolution strategies (ES). In general, the genes of a genetic algorithm consist of a binary code of 0 or 1. Although the motor value can be expressed as a chromosome composed of 0 s and 1 s, the length of the chromosome becomes cumbersome and the motor value cannot be interpreted intuitively. The chromosome was therefore constructed with real numbers between −1 and 1 by using real-coded GA, in consideration of the length of the chromosome, so that the motor value could be intuitively identified [36]. The gene position in a chromosome is defined in Figure 2.
Figure 3 shows the face control positions for EveR. There are 23 motors for the face; however, movements of the pupils and tongue that do not significantly affect the facial expression recognition rate were excluded, leaving 16 control positions (Figure 3). The double arrows indicate movements that can be controlled on both sides (e.g., number 1) and single arrows indicate motion that can be controlled on a single side (e.g., number 3). The robot’s eyes are bilaterally symmetrical and have six degrees of freedom. Emotions and eye blinking can be expressed by controlling the eyebrows and eyelids. The area around the mouth has 10 degrees of freedom to express emotions as well as perform lip syncing for conversation. Table 1 summarizes the identification numbers for motors in the face, gene positions in the chromosome and face skin control positions used in this study. For example, motor number 1 corresponds to the right eyelid of EveR and is presented in the g 1 position of the chromosome.
Figure 4 shows the facial expression made by EveR from randomly generated genes. For g 1 , g 2 , g 9 and g 16 , which can be controlled in both directions, a value between −1 and 1 was randomly generated and values between 0 and 1 were randomly generated for single-direction genes. For example, the right eyelid value, g 1 , was −0.64 and the left eyelid value, g 2 , was 0.35. The difference between the genes for the two eyelids demonstrates the different actions being controlled; g 1 expresses the action for closing the eyes (less than zero) and g 2 expresses the action for opening the eyes widely (greater than zero). For g 16 , which controls the jaw, a value less than zero resulted in a closed mouth gesture whereas a value greater than zero resulted in an open mouth gesture.

3.2. Evaluation

The fitness of facial expressions generated with genes was evaluated using FaceReader, a facial expression recognizer software package developed by Noldus in the Netherlands. In addition to neutral emotion, FaceReader can recognize six basic emotions in real time—happy, sad, angry, surprised, fear and disgust. According to the results of a verification study, the emotion classification tool in FaceReader 6 has an average classification accuracy of 88%. They claimed that the latest version, FaceReader 7.1, achieved an emotion classification accuracy of 93%.
Figure 5 shows the results of human facial expression recognition by FaceReader as a bar graph. The Neutral bar was highest when the person was not performing a facial expression (Figure 5a) and the Happy bar was highest when the person was laughing (Figure 5b).
The facial features of the EveR android, which are similar to those of a person, can also be recognized as emotional changes expressed in real time through FaceReader. Figure 6 shows the results of facial expression recognition for EveR in the Neutral state; FaceReader recognized the android’s face accurately (shown by the rectangular box) and confirmed its Neutral facial expression (gray bar).

3.3. Facial Expression Generation Optimization System

Figure 7 shows the GA-based facial expression generation algorithm used in this study. The proposed system consists of EveR, FaceReader and the GA for the evolution of facial expressions. m is a vector composed of the 16 motor values used for generating the facial expressions, defined as m = [ m 1 , m 2 , , m 16 ] T R 16 . The facial expression can be expressed as F ( m ) . FaceReader’s recognition rate for the F ( m ) facial expression is defined as Y R 1 . The target recognition rate is defined as Y d , such that the error becomes e = Y d Y . Finally, the face motor value m is repeatedly updated by evolving the chromosome through the GA until the smallest error value in the generation reaches the set threshold value.

4. Experiment and Analysis

4.1. Experimental Setup

In this study, a system for generating the facial expressions of an android based on a GA was proposed. To verify this system, we attempted to generate six basic facial expressions (angry, happy, sad, surprised, disgust, fear) in EveR. For this purpose, the system was constructed as shown in Figure 8. Each software program could exchange data through TCP/IP communication.
The main control software of the system was developed based on genetic algorithms using DEAP’s library. The GA parameters used for the experiment are as follows. Each chromosome was composed of 16 genes (16 motors) and each generation was composed of 20 chromosomes (20 facial expressions). The tournament method was used for selection with a tournament size of 3. The two-point exchange method was used for crossover and the Gaussian method was used for mutation. The coefficients were set to 0.1 for μ and 0.2 for σ , to ensure that the mutation did not cause serious displacements in the gene. The probabilities of crossover and mutation were set to 50% and 30%, respectively. The facial expression intensity received from the facial expression recognizer was utilized to verify the fitness by generation and the accumulated fitness score was used for parent generation selection.
Facial expressions for the robot head were generated after receiving the chromosomes (motor values) from the main control software and controlling the motors. The control period of the proposed system was set to 10 s. For 7 s, chromosome transmission from DEAP to the robot control system, facial expression generation and maintenance in the robot head, facial expression recognition and fitness transmission from FaceReader to DEAP were performed. We recognized the facial expression by recording a snapshot 3 s after the facial expression generation. When the expression changes in the neutral state, the result of the recognizer varies greatly. To minimize the error, we maintain the expression until the recognition result is stable and then record the result. The remaining 3 s were used as rest time by setting all motor values to zero to minimize the failure of the robot. Finally, the expression intensity of the facial expressions of EveR was transmitted to the main control software to be utilized as the fitness value of the system.

4.2. Facial Expression Analysis

In this study, six basic facial expressions (angry, disgust, fear, happy, sad, surprised) were generated to evaluate the performance of the proposed system. To minimize failure of the robot skin and motors in the proposed system, evolution was limited to a maximum of 20 generations. Evolution was stopped when the average expression intensities of the evolved generations did not increase by more than 5% for four consecutive instances or when the set threshold value was reached.
Figure 9 shows the evolution of facial expressions by emotion. It shows the expression of chromosomes with the highest fitness for each generation. The surprised facial expression completed evolution in the 6th generation, which was the fastest evolution observed. In contrast, the expression intensity for angry, disgust and fear facial expressions did not significantly increase over 20 generations of evolution. Finally, evolution of the sad and happy facial expressions was completed by the 18th and 12th generations, respectively.
Figure 10 summarizes the fitness value for each generation of an evolving facial expression. Figure 10a–f show the fitness values for angry, happy, sad, surprised, disgust and fear expressions, respectively. The yellow line represents the fitness of the expression created manually by controlling the motor in one axis, whereas the blue line indicates the maximum fitness of the expression created by the proposed system following each generation and the red line indicates the average fitness following each generation.
The facial expressions generated using the proposed system had higher fitness values than manually generated facial expressions. For sad expressions, the system’s evolution produced more suitable facial expressions than manual expressions by the 5th generation. From the 8th generation, the average fitness of the evolving generation was higher than that of the manual expressions. Moreover, the fitness of every facial expressions increased with evolution. However, the recognizer did not effectively identify angry and happy facial expressions. Although more appropriate facial expressions were generated through this system, with respect to their fit to an emotion, future research should increase the expressiveness of the android by changing the skin control position of the robot and adding degrees of freedom to generate more facial expressions.
We generated facial expressions by applying the proposed system to the older male face robot. The internal structure of this robot and the number of motors are similar to those of ’EveR’ but the skin control position, external shape and skin thickness are different. We applied the system to the male robot by mapping the motor value to the gene to generate facial expressions for the robot. We generate facial expressions expressing six emotions through the same process as that when generating ’EveR’ facial expressions.
Figure 11 and Figure 12 show the evolution process of facial expression generation and the results of the facial expression generation experiment of the male robot. Each facial expression in Figure 11 was generated using chromosomes with the highest fitness of the generation. In the case of the male robot’s facial expression, all 20 generations made a facial expression. Similar to the EveR facial expressions, surprise and sad expressions were well expressed but the other facial expressions could not be generated well.
Finally, we analyzed the GA parameters affecting facial expression generation. The GA parameters that were previously applied are as follows. Each generation comprised 20 chromosomes. We used tournament selection and Gaussian mutation; the tournament size was 3 and the sigma value of the mutation was 0.2. We generated a happy facial expression by varying the number of chromosomes, tournament size and the sigma value of the Gaussian mutation. Figure 13 shows the evolutionary result of the parameter difference. The number of chromosomes is 10, the sigma value is 0.02 and the tournament size is 3 in Figure 13a. The number of chromosomes is 10, the sigma value is 0.2 and the tournament size is 2 in Figure 13b. The number of chromosomes is 10, the sigma value is 0.2 and the tournament size is 3 in Figure 13c. The number of chromosomes is 20, the sigma value is 0.2 and the tournament size is 2 in Figure 13d. We confirmed that the faster the evolution (owing to a larger sigma value of the mutation), the larger the change in motor value through Figure 13a,b. Similarly, through Figure 13b,c, the evolution speed is fast when the tournament size is large. And we confirmed that the higher the number of chromosomes, the faster the evolutionary rate through Figure 13d.

5. Discussion

5.1. Difference of Facial Expression According to Recognition Performance

The facial expression recognizer has the biggest influence on the facial expression generation of the android. A recognizer built by learning the face of a Westerner cannot recognize the facial expressions of Asian people. Based on recent research results, Wataru proved that the basic emotional expression is not related to pleasure, surprise expression and culture, unlike Ekman’s claim that it is similar regardless of culture [37]. This implies that facial expressions are not generated well when Western facial datasets are used to create facial expressions for Eastern facial robots.
To generate facial expressions naturally, the performance of the recognizer should be supported. By using a recognizer that accounts for the characteristics of the robot, more suitable facial expressions can be generated. For example, to make an angry facial expression for an Asian male robot, an Asian facial expression recognizer based on an angry data set would be ideal. This method can generate more suitable facial expressions than those created using commercial software.

5.2. Limitations

We aimed to generate facial expressions for androids that provide guidance and personal services. Depending on the purpose of the service, androids have different genders, facial features and functions. The proposed system has the advantage that facial expressions of various appearances can be easily generated by using recognizers learned based on human facial data. If an artist with aesthetic sensibilities creates a facial expression while considering the director’s intention, a more suitable facial expression than that generated using this system can be made. However, it is very difficult for an artist to generate facial expressions because they cannot accurately predict the movement of the skin when the motor is moved. Therefore, facial expressions are generated easily using our proposed system. It would be more efficient to create facial expressions after generating a draft facial expression by using this system.
The expressiveness of facial expressions that can be made varies according to the characteristics of each face robot. There are approximately 80 muscles in the human face. However, face robots only have between 10 and 25 degrees of freedom. This makes it very difficult to generate facial expressions of the robot that are identical to those of human beings. The number of facial expressions of the android that can be generated will increase as the number of motors increases but this is accompanied by increased interference among the wires.

6. Conclusions

In this study, we proposed a system that can automatically generate facial expressions for an android by combining three components—an android head, a facial expression recognizer capable of interpreting human and android faces and a genetic algorithm. A facial expression recognizer was employed to identify when androids exhibited angry, happy, sad, surprised, disgust and fear facial expressions. These expressions were controlled using one chromosome (a set of genes), through 1:1 mapping of the motors and genes. Each generation of facial expressions comprised 20 chromosomes. Crossover probabilities were set to 50% and mutation probabilities were set to 30%. For evolution into the next generation, the tournament method was used for selection, the two-point method was used for crossover and the Gaussian method was used for mutation. The results from the facial expression recognizer showed that the fitness of all facial expressions generated using the proposed system was higher than that of manually generated facial expressions, with the surprised facial expression evolving quickest. Although the face of the EveR android had insufficient muscle expression ability to effectively generate angry and happy facial expressions, the expression ability for surprised and sad expressions was excellent.
We suggest that the facial expressions of a range of androids can be conveniently generated using the proposed system. In particular, the main advantage of this system is that it does not need to take factors such as degree of freedom, internal structure, skin control position and range into consideration to generate a robot’s facial expression. To utilize this system, it is only necessary to map the motor value of face robot to a gene. The system will determine the optimal facial expression by reflecting the characteristics of the face robot. The higher the degree of freedom of the face robot, the more deformable the skin can be, allowing a natural facial expression to be generated. In addition, a face robot with the appearance of an Asian person can generate a natural facial expression by using a recognizer learned from Asian facial expression data.

Author Contributions

Conceptualization, H.-J.H., H.U.Y. and D.-W.L.; Funding acquisition, D.-W.L.; Methodology, H.-J.H.; Project administration, D.-W.L.; Resources, D.C. and D.-Y.L.; Software, H.-J.H.; Supervision, D.-W.L.; Writing—original draft, H.-J.H.; Writing—review & editing, H.U.Y. and D.-W.L.

Funding

This research is supported by Ministry of Culture, Sports and Tourism(MCST) and Korea Creative Content Agency(KOCCA) in the Culture Technology(CT) Research & Program 2017.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Duffy, B.D. Anthropomorphism and the Social Robot. Robot. Auton. Syst. 2003, 42, 177–190. [Google Scholar] [CrossRef]
  2. Scheeff, M.; Pinto, J.; Rahardja, K.; Snibbe, T.; Tow, R. Experiences with Sparky, a Social Robot. In Proceedings of the Workshop on Interactive Robotics and Entertainment, Pittsburgh, PA, USA, 30 April–1 May 2002. [Google Scholar]
  3. Saldien, J.; Goris, K.; Vanderborght, B.; Vanderfaeillie, J.; Lefeber, D. Expressing Emotions with The Social Robot Probo. Int. J. Soc. Robot. 2010, 4, 377–389. [Google Scholar] [CrossRef]
  4. Zalama, E.; García-Bermejo, J.G.; Marcos, S.; Domínguez, S.; Feliz, R.; Pinillos, R.; López, J. Sacarino, a Service Robot in a Hotel Environment. In Proceedings of the ROBOT2013: First Iberian Robotics Conference, Madrid, Spain, 28–29 November 2013; Springer: Cham, Switzerland, 2014; Volume 253, pp. 3–14. [Google Scholar]
  5. Pinillos, R.; Marcos, S.; Feliz, R.; Zalama, E.; García-Bermejo, J.G. Long-Term Assessment of a Service Robot in a Hotel Environment. Robot. Auton. Syst. 2016, 79, 40–57. [Google Scholar] [CrossRef]
  6. Lee, K.; Kim, H.; Yoon, W.C.; Yoon, Y.; Kwon, D. Designing a Human-Robot Interaction Framework for Home Service Robot. In Proceedings of the 2005 IEEE International Workshop on Robots and Human Interactive Communication, Nashville, TN, USA, 13–15 August 2005; pp. 286–293. [Google Scholar]
  7. Viraj, M.A.; Muthugala, J.; Buddhika, A.G.; Jayasekara, P. MIRob: An Intelligent Service Robot that Learns from Interactive Discussions while Handling Uncertain Information in User Instructions. In Proceedings of the 2016 Moratuwa Engineering Research Conference (MERCon), Moratuwa, Sri Lanka, 5–6 April 2016; pp. 397–402. [Google Scholar]
  8. Vasquez, H.A.; Vargas, H.S.; Sucar, L.E. Using Gestures to Interact with a Service Robot using Kinect 2. Res. Comput. Sci. 2015, 96, 85–93. [Google Scholar]
  9. Quan, L.; Pei, D.; Wang, B.; Ruan, W. Research on Human Target Recognition Algorithm of Home Service Robot based on Fast-RCNN. In Proceedings of the 2017 10th International Conference on Intelligent Computation Technology and Automation (ICICTA), Changsha, China, 9–10 October 2017; pp. 369–373. [Google Scholar]
  10. Lin, C.; Li, T.S.; Kuo, P.; Wang, Y. Integrated Particle Swarm Optimization Algorithm based Obstacle Avoidance Control Design for Home Service Robot. Comput. Electr. Eng. 2016, 56, 748–762. [Google Scholar] [CrossRef]
  11. Minato, T.; Shimada, M.; Ishiguro, H.; Itakura, S. Development of an Android Robot for Studying Human-Robot Interaction. In Proceedings of the International Conference on Industrial, Engineering and Other Applications of Applied Intelligent Systems, Graz, Austria, 9–11 July 2004; pp. 424–434. [Google Scholar]
  12. Lee, D.; Lee, T.; So, B.; Choi, M.; Shin, E.; Yang, K.; Baek, M.; Kim, H.; Lee, H. Development of an Android for Emotional Expression and Human Interaction. In Proceedings of the 17th World Congress the International Federation of Automatic, Control, Seoul, Korea, 6–11 July 2008; pp. 4336–4337. [Google Scholar]
  13. Hyung, H.; Yoon, H.U.; Choi, D.; Lee, D.; Hur, M.; Lee, D. Facial Expression Generation of an Android Robot based on Probabilistic Model. In Proceedings of the 2018 27th IEEE International Symposium on Robot and Human Interactive Communication (RO-MAN), Tai’an, China, 27–31 August 2018; pp. 458–460. [Google Scholar]
  14. Go, D.; Hyung, H.; Yoon, H.U.; Lee, D. Android Robot Motion Generation based on Video-Recorded Human Demonstrations. In Proceedings of the2018 27th IEEE International Symposium on Robot and Human Interactive Communication (RO-MAN), Tai’an, China, 27–31 August 2018; pp. 476–478. [Google Scholar]
  15. Noma, M.; Saiwaki, N.; Itakura, S.; Ishiguro, H. Composition and Evaluation of the Humanlike Motions of an Android. In Proceedings of the 2006 6th IEEE-RAS International Conference on Humanoid Robots, Genova, Italy, 4–6 December 2006; pp. 163–168. [Google Scholar]
  16. Khoramshahi, M.; Shukla, A.; Raffard, S.; Bardy, B.G.; Billard, A. Role of Gaze Cues in Interpersonal Motor Coordination: Towards Higher Affiliation in Human-Robot Interaction. PLoS ONE 2016, 11, e0156874. [Google Scholar] [CrossRef] [PubMed]
  17. Shimada, M.; Yoshikawa, Y.; Asada, M.; Saiwaki, N.; Ishiguro, H. Effects of Observing Eye Contact between a Robot and Another Person. Int. J. Soc. Robot. 2011, 3, 143–154. [Google Scholar] [CrossRef]
  18. Ahn, H.S.; Lee, D.; Choi, D.; Lee, D.Y.; Hur, M.H.; Lee, H.; Shon, W.H. Development of an Android for Singing with Facial Expression. In Proceedings of the IECON 2011-37th Annual Conference of the IEEE Industrial Electronics Society, Melbourne, Australia, 7–10 November 2011; pp. 104–109. [Google Scholar]
  19. Becker-Asano, C.; Ishiguro, H. Evaluating Facial Displays of Emotion for the Android Robot Geminoid F. In Proceedings of the 2011 IEEE workshop on affective computational intelligence (WACI), Paris, France, 11–15 April 2011; pp. 1–8. [Google Scholar]
  20. Oh, K.; Jung, C.; Lee, Y.; Kim, S. Real-Time Lip Synchronization between Text-To-Speech (TTS) System and Robot Mouth. In Proceedings of the 19th International Symposium in Robot and Human Interactive Communication, Viareggio, Italy, 13–15 September 2010; pp. 620–625. [Google Scholar]
  21. Kondo, Y.; Takemura, K.; Takamatsu, J.; Ogasawara, T. A Gesture-Centric Android System for Multi-Party Human-Robot Interaction. J. Hum. Robot Int. 2013, 2, 133–151. [Google Scholar] [CrossRef]
  22. Salem, M.; Kopp, S.; Wachsmuth, I.; Rohlfing, K.; Joublin, F. Generation and Evaluation of Communicative Robot Gesture. Int. J. Soc. Robot. 2012, 201–217. [Google Scholar] [CrossRef]
  23. Salem, M.; Kopp, S.; Wachsmuth, I.; Joublin, F. Towards an Integrated Model of Speech and Gesture Production for Multi-Modal Robot Behavior. In Proceedings of the 19th International Symposium in Robot and Human Interactive Communication, Viareggio, Italy, 13–15 September 2010; pp. 614–619. [Google Scholar]
  24. Mehrabian, A. Silent Messages Implicit Communication of Emotions and Attitudes, 2nd ed.; Wadsworth Publishing Company: Belmont, CA, USA, 1981. [Google Scholar]
  25. Ekman, P.; Friesen, W.V. Unmasking the Face: A Guide to Recognizing Emotions from Facial Clues; Prentice Hall: Upper Saddle River, NJ, USA, 1975. [Google Scholar]
  26. Marcolin, F.; Vezzetti, E. Novel Descriptors for Geometrical 3D Face Analysis. Multimed. Tools Appl. 2017, 76, 13805–13834. [Google Scholar] [CrossRef]
  27. Vezzetti, E.; Marcolin, F.; Stola, V. 3D human Face Soft Tissues Landmarking Method: An Advanced Approach. Comput. Ind. 2013, 64, 1326–1354. [Google Scholar] [CrossRef]
  28. Habib, A.; Das, S.K.; Bogdan, I.; Hanson, D.; Popa, D.O. Learning Human-like Facial Expressions for Android Phillip K. In Dick. In Proceedings of the 2014 IEEE International Conference on Automation Science and Engineering (CASE), New Taipei, Taiwan, 18–22 August 2014; pp. 1159–1165. [Google Scholar]
  29. Ishi, C.T.; Minato, T.; Ishiguro, H. Analysis and Generation of Laughter Motions, and Evaluation in an Android Robot. APSIPA Trans. Signal Inf. Process. 2019, 8. [Google Scholar] [CrossRef]
  30. Lin, C.; Huang, C.; Cheng, L. A Small Number Actuator Mechanism Design for Anthropomorphic Face Robot. In Proceedings of the 2011 IEEE Int. Conf. on Robotics and Biomimetics, Phuket, Thailand, 7–11 December 2011; pp. 633–638. [Google Scholar]
  31. Hashimoto, M.; Yokogawa, C.; Sadoyama, T. Development and Control of a Face Robot Imitating Human Muscular Structures. In Proceedings of the 2006 IEEE/RSJ International Conference on Intelligent Robots and Systems, Beijing, China, 9–15 October 2006; pp. 1855–1860. [Google Scholar]
  32. Huang, Z.; Ren, F.; Bao, Y.; Huang, Z.; Ren, F. Human-like Facial Expression Imitation for Humanoid Robot based on Recurrent Neural Network. In Proceedings of the 2016 International Conference on Advanced Robotics and Mechatronics (ICARM), Macau, China, 18–20 August 2016; pp. 306–311. [Google Scholar]
  33. Mori, M. The Uncanny Valley. Energy 1970, 7, 33–35. [Google Scholar]
  34. Michalewicz, Z. Genetic Algorithms + Data Structures = Evolution Programs; Springer: Berlin/Heidelberg, Germany, 1996; pp. 1–387. [Google Scholar]
  35. Fortin, F.; Rainville, F.D.; Gardner, M.; Parizeau, M.; Gagné, C. DEAP: Evolutionary Algorithms Made Easy. J. Mach. Learn. Res. 2012, 13, 2171–2175. [Google Scholar]
  36. Herrera, F.; Lozano, M.; Verdegay, J.L. Tackling Real-Coded Genetic Algorithms: Operators and Tools for Behavioural Analysis. Artif. Intell. Rev. 1998, 12, 265–319. [Google Scholar] [CrossRef]
  37. Wataru, S.; Hyniewska, S.; Minemoto, K.; Yoshikawa, S. Facial Expressions of Basic Emotions in Japanese Laypeople. Front. Psychol. 2019, 10, 1326–1354. [Google Scholar]
Figure 1. Android, EveR, in various exhibitions and performances.
Figure 1. Android, EveR, in various exhibitions and performances.
Applsci 09 03379 g001
Figure 2. Gene position in a chromosome.
Figure 2. Gene position in a chromosome.
Applsci 09 03379 g002
Figure 3. Face control positions of android.
Figure 3. Face control positions of android.
Applsci 09 03379 g003
Figure 4. Facial expression of the EveR android generated by randomly generated genes.
Figure 4. Facial expression of the EveR android generated by randomly generated genes.
Applsci 09 03379 g004
Figure 5. Results of human facial expression recognition using FaceReader for (a) a Neutral expression and (b) a Happy expression.
Figure 5. Results of human facial expression recognition using FaceReader for (a) a Neutral expression and (b) a Happy expression.
Applsci 09 03379 g005
Figure 6. Results of EveR facial expression recognition using FaceReader.
Figure 6. Results of EveR facial expression recognition using FaceReader.
Applsci 09 03379 g006
Figure 7. Genetic algorithm (GA) for generating facial expressions of the EveR android.
Figure 7. Genetic algorithm (GA) for generating facial expressions of the EveR android.
Applsci 09 03379 g007
Figure 8. Facial expression generation system based on GA.
Figure 8. Facial expression generation system based on GA.
Applsci 09 03379 g008
Figure 9. Evolution of different facial expressions in ’EveR’.
Figure 9. Evolution of different facial expressions in ’EveR’.
Applsci 09 03379 g009
Figure 10. Comparison between fitness value of facial expressions generated by the proposed system and those generated manually in ’EveR’—(a) Angry, (b) Disgust, (c) Fear, (d) Happy, (e) Sad and (f) Surprised.
Figure 10. Comparison between fitness value of facial expressions generated by the proposed system and those generated manually in ’EveR’—(a) Angry, (b) Disgust, (c) Fear, (d) Happy, (e) Sad and (f) Surprised.
Applsci 09 03379 g010
Figure 11. Evolution of different facial expressions in older male robot.
Figure 11. Evolution of different facial expressions in older male robot.
Applsci 09 03379 g011
Figure 12. Comparison between fitness value of facial expressions generated by the proposed system and those generated manually in older male robot—(a) Angry, (b) Disgust, (c) Fear, (d) Happy, (e) Sad and (f) Surprised.
Figure 12. Comparison between fitness value of facial expressions generated by the proposed system and those generated manually in older male robot—(a) Angry, (b) Disgust, (c) Fear, (d) Happy, (e) Sad and (f) Surprised.
Applsci 09 03379 g012
Figure 13. Comparison of Fitness of Facial Expression by GA Parameter.
Figure 13. Comparison of Fitness of Facial Expression by GA Parameter.
Applsci 09 03379 g013
Table 1. Relationships between face control positions and genes.
Table 1. Relationships between face control positions and genes.
Motor NumberGene PositionControl Position
1 g 1 Right eyelid
2 g 2 Left eyelid
3 g 3 Inner right eyebrow
4 g 4 Inner left eyebrow
5 g 5 Outer right eyebrow
6 g 6 Outer left eyebrow
7 g 7 Above the right lip
8 g 8 Above the left lip
9 g 9 Lower lip
10 g 10 Right cheek (top)
11 g 11 Left cheek (top)
12 g 12 Right cheek (middle)
13 g 13 Left cheek (middle)
14 g 14 Around the upper lip
15 g 15 Around the lower lip
16 g 16 Jaw

Share and Cite

MDPI and ACS Style

Hyung, H.-J.; Yoon, H.U.; Choi, D.; Lee, D.-Y.; Lee, D.-W. Optimizing Android Facial Expressions Using Genetic Algorithms. Appl. Sci. 2019, 9, 3379. https://doi.org/10.3390/app9163379

AMA Style

Hyung H-J, Yoon HU, Choi D, Lee D-Y, Lee D-W. Optimizing Android Facial Expressions Using Genetic Algorithms. Applied Sciences. 2019; 9(16):3379. https://doi.org/10.3390/app9163379

Chicago/Turabian Style

Hyung, Hyun-Jun, Han Ul Yoon, Dongwoon Choi, Duk-Yeon Lee, and Dong-Wook Lee. 2019. "Optimizing Android Facial Expressions Using Genetic Algorithms" Applied Sciences 9, no. 16: 3379. https://doi.org/10.3390/app9163379

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop