You are currently viewing a new version of our website. To view the old version click .
Healthcare
  • Article
  • Open Access

8 August 2022

Visualization and Semantic Labeling of Mood States Based on Time-Series Features of Eye Gaze and Facial Expressions by Unsupervised Learning

,
and
1
Faculty of Software and Information Science, Iwate Prefectural University, Takizawa 020-0693, Japan
2
Faculty of Systems Science and Technology, Akita Prefectural University, Akita 015-0055, Japan
*
Author to whom correspondence should be addressed.
This article belongs to the Special Issue Mental and Behavioral Healthcare

Abstract

This study is intended to develop a stress measurement and visualization system for stress management in terms of simplicity and reliability. We present a classification and visualization method of mood states based on unsupervised machine learning (ML) algorithms. Our proposed method attempts to examine the relation between mood states and extracted categories in human communication from facial expressions, gaze distribution area and density, and rapid eye movements, defined as saccades. Using a psychological check sheet and a communication video with an interlocutor, an original benchmark dataset was obtained from 20 subjects (10 male, 10 female) in their 20s for four or eight weeks at weekly intervals. We used a Profile of Mood States Second edition (POMS2) psychological check sheet to extract total mood disturbance (TMD) and friendliness (F). These two indicators were classified into five categories using self-organizing maps (SOM) and U-Matrix. The relation between gaze and facial expressions was analyzed from the extracted five categories. Data from subjects in the positive categories were found to have a positive correlation with the concentrated distributions of gaze and saccades. Regarding facial expressions, the subjects showed a constant expression time of intentional smiles. By contrast, subjects in negative categories experienced a time difference in intentional smiles. Moreover, three comparative experiment results demonstrated that the feature addition of gaze and facial expressions to TMD and F clarified category boundaries obtained from U-Matrix. We verify that the use of SOM and its two variants is the best combination for the visualization of mood states.

1. Introduction

The advanced progress of information technologies in our society provides usefulness, accessibility, and convenience to our daily lives. Particularly with the COVID-19 pandemic, the need for remote work, online meetings, and online learning has rapidly spread around the world [1,2,3,4,5]. By virtue of modern widespread internet technology, huge amounts of digital data, including big data [6], are circulating rapidly in real-time around the world, not only with global information provided as news articles from mass media but also with local information posted from bloggers and community information exchanged using social networking services (SNS) [7]. Simultaneously, regarding negative aspects, various difficulties have arisen, such as invasion of privacy, lack of computer literacy, unfounded rumors, and fake news [8]. The emergence of deepfakes [9] that can produce hyper-realistic videos using deep learning (DL) networks [10] accelerates this issue [11,12,13,14].
In addition, industrial products and computer interfaces that are unfamiliar to people inhibit satisfactory living and social activities to everyone’s desire that they be convenient. Users who try to force themselves to fit in with many situations might feel uncomfortable, frustrated, and stressed. Actually, the dominant industrial structure in modern society has changed from manufacturing industries in the last century to information industries, which process large amounts of data as digital codes in real-time [15]. Computers, tablets, and smartphones play important roles as powerful tools in marketing activities and our current digitalized society [16]. Particularly, DL technologies and applications have boosted this progress, especially during its major transition in 2012 [17]. However, these digital devices have induced numerous instances of confusion in human communication [18]. The use of irrational, difficult, and complex hardware and software often induces stress factors. Therefore, numerous people spend their daily lives and businesses coping with stressors of various types that are attributable to these influences and realities.
Stress occurs as a vital response of the brain and body to cope with stressors [19]. Individual variations exist in stress response, tolerance, and emotional patterns. Therefore, the magnitude of stress varies even among people along with individual differences [20], even in similar environments, conditions, situations, circumstances, and contexts. Usually, in a healthy condition, the brain and body respond appropriately to emphasize maintenance of physical and mental balance. However, excessive stress induces abnormalities in the mind and body. In the worst case, we are adversely affected by mental illnesses such as depression, psychosomatic disorders, and neurosis [21]. Transformation of the industrial structure and business style decreases physical illness and increases mental illness. Particularly, those who work in the service industry encounter widely various stressors in their daily work.
In Japan, a stress check program was introduced in 2014 with a revision of the Industrial Safety and Health Act [22]. Since 2015, stress check tests have been imposed as an obligation for organizations that have more than 50 employees. However, organizations employing fewer than 50 workers are simply asked to make an effort to do so. Stress check tests are conducted at one-year intervals by a medical doctor or by a public health nurse. This frequency is unsuitable for the early detection of stress accumulated in daily lives. Therefore, recognizing mild discomfort as a sign of stress plays an important role in stress management. Moreover, tools, methods, and systems that can measure stress simply, easily, readily, and frequently over a long period are expected to be necessary for modern high-stress societies. This study is intended to develop a stress measurement and visualization system for stress management in terms of simplicity and reliability. Particularly, we prototyped a mental health visualizing framework based on machine learning (ML) algorithms as a software tool that can feed back analytical signals of stress measurements [23].
Most existing studies have specifically examined subjective responses under transient stress [24]. Long-term periodic stress observations are related closely to slight changes in mental conditions. However, this study was conducted to collect an original dataset related to chronic stress obtained from university students. Usually, university students feel burdened with their daily routines and schoolwork, which might be expected to include attending lectures, club activities, working as a teaching or research assistant as a part-time job, writing reports, and research on a graduation thesis [25]. This study specifically examined the correlation between chronic stress and biometric signals obtained as physical responses. We obtained an original dataset related to stress responses from psychological and behavioral indexes. Particularly, we obtained two psychological scores related to mood states and time-series images that included facial expressions, gaze distributed patterns, and the number of saccades.
This study is intended to visualize features related to mental health using ML algorithms. We hypothesized that emotional changes resulting from different chronic stress conditions affect gaze movements and facial expression changes. Our earlier feasible research results [23] demonstrated that the degree of gaze concentration tended to be related to the psychological state. This study was conducted to verify the relation between gaze movements, including saccades, and chronic stress in daily life without using external stimuli. Feature signals of gazing and facial expressions are obtainable using a nonrestricted measurement approach. Therefore, burdens for subjects are lower than that of restricted or contact measurement methods. Experimentally obtained results obtained from 20 subjects demonstrated a tendency of feature patterns visualized on category maps for analyzing stress responses in each subject.
In this study, we used self-organizing maps (SOMs) [26] for unsupervised clustering and data visualization [27]. Because of containing both properties, SOMs have been widely used in various and numerous studies in the era of mainstream DL algorithms. Compared with DL algorithms that require a vast of data, one important advantage for SOMs is to conduct steady learning with relatively lower computational resources and calculation costs. Recent research examples of clustering, visualization, recognition, classification, and analyses using SOMs comprise medical system applications [28,29,30,31,32], social infrastructure maintenance [33,34,35,36,37,38], consumer products and services [39,40,41,42,43], food and smart farming [44,45,46], and recycling and environmental applications [47,48,49,50,51,52,53]. We employed SOMs and their variants for the task of classification and visualization of mood states.
This paper is structured as follows. Section 2 briefly reviews state-of-the-art stress measurement systems and methods, especially non-invasive and non-contact approaches. Subsequently, Section 3 and Section 4, respectively, present our original benchmark dataset and our proposed method consisting of four ML algorithms. Experiment results of classification and visualization of mood states related to gaze features and facial expressions are presented in Section 5. Finally, Section 6 presents conclusions and highlights future work.

3. Dataset

3.1. Experiment Environment

Figure 2 depicts the room used for an experiment environment to obtain benchmark datasets. The partition installed in the room of 20 m 2 separated the sections for a subject and an experimenter as an observer. The laptop computers on the desk at the front side of the room were connected to a measurement device for data collection. The experimenter monitored the progress of experimental protocols and the responses of subjects. After sitting on a chair at the back of the room, the subject watched the 50-inch monitor placed at 3 m distance across the table. The facial measurement device was set up on the table. A video for communication with an interlocutor was shown on the monitor. We took care to maintain silence in the room to allow subjects to undergo the experiment in a relaxed condition. The blinds on the windows were closed for protection from sunlight. The room temperature and humidity were kept constant using an air-conditioner.
Figure 2. The experiment room is divided into two sections for a subject and an experimenter (Left). The subject watches the monitor, which is placed at a 3 m distance across the table (Right). A video showing communication with an interlocutor is presented on the monitor.

3.2. Sensing Device

For gaze tracking and saccade extraction, we used faceLAB 5 (Seeing Machines Inc.; Fyshwick, ACT, Australia), as depicted in Figure 3. The faceLAB 5 apparatus comprises an emitter and a stereo camera with 0.5–1.0 deg angular resolution and 60 Hz data sampling. The included application software produces heatmap results calculated from gaze concentration density and the number of saccades, which are defined as rapid eye movements between fixation points.
Figure 3. Exterior of the faceLAB 5 head-and-eye tracking device.

3.3. POMS2

We used the Profile of Mood States Second edition (POMS2) [82] sheets for measuring psychological information from the respective subjects. POMS2 is used at clinical sites such as those for medical care, nursing, welfare, and counseling.
POMS2 consists of seven mood components: anger–hostility ( A H ), confusion–bewilderment ( C B ), depression–dejection ( D D ), fatigue–inertia ( F I ), tension–anxiety ( T A ), vigor–activity ( V A ), and friendliness (F). For each component, subjects give responses according to five-point scales. The total mood distance (TMD) is calculated as
TMD = A H + C B + D D + F I + T A V A ,
where V A is inverted. Component F is an index that is independent of TMD. For clustering mood states, we define TMD as the primary component and F as the secondary component.
With the different item numbers, POMS2 provides three versions. The numbers of items for a youth version for young people between 13 and 17 years old, an adult version for more than 18 years old, and a simplified adult version are, respectively, 60, 65, and 35 items. The subjects for this experiment were all university students older than 18 years old. They were applicable to the adult version of POMS2 in Japanese [83]. Regarding the total experiment time, we used the simplified version. The mean answer time was approximately five minutes.
After standardizing the prime scores, we calculated T-scores with a mean of 50 and a standard deviation of 10. The T-score conversion normalizes the metrics of assessment in terms of numerical equivalents. This normalization provides the possibility of appropriate comparison among individual examinations for the obtained scores, scales, and forms.

3.4. Obtained Datasets

Our original benchmark dataset was obtained from 20 university students, 10 male and 10 female, through volunteer sampling. Table 1 comprises the profiles of respective subjects. The data collection interval was set to one week to reduce the effects of the response for a frequently repeated stimulus. Regarding restrictions for subjects, the total measurement terms were set to two types: four weeks for 10 subjects and eight weeks for 10 subjects. Therefore, the total data volume is 120 sets.
Table 1. Profiles and measurement terms for subjects.

4. Proposed Method

Our original benchmark dataset includes no ground truth (GT) labels. Therefore, we employed unsupervised learning methods. Figure 4 depicts the entire procedure of our proposed method, comprising four ML algorithms: SOM [26], recurrent SOM (RSOM) [84], growing hierarchical SOM (GHSOM) [85], and U-Matrix [86]. The TMD and F scores, gaze features, saccades, and face images were obtained from POMS2, FaceLab, and a monocular camera. First, categories related to mood status are created using SOM and U-Matrix from TMD and F. Gaze features and saccades are used for analyzing the obtained categories. Subsequently, smile images are extracted using RSOM and GHSOM from time-series face images obtained from a monocular camera.
Figure 4. Structure and data flow of our proposed method. Category maps are created from several combinations of input features.
Figure 5 depicts the respective network structures. The core algorithms of U-Matrix, RSOM, and GHSOM were designed based on SOM.
Figure 5. Network structures of SOM, U-Matrix, RSOM, and GHSOM. SOM and U-Matrix are used for creating category maps. RSOM and GHSOM are used for extracting smile-expressed frames.

4.1. SOM

Letting x i ( t ) denote the features to input layer unit i at time t. Furthermore, letting w i j k ( t ) denote a weight from i to mapping layer unit ( j , k ) at time t. Before learning, values of w i j k ( t ) are initialized randomly. Using the Euclidean distance between x i ( t ) and w i j k ( t ) , a winner unit c j ( t ) is sought for the following as
c j ( t ) = argmin 1 j J , 1 k K i = 1 I ( x i ( t ) w i j k ( t ) ) 2 ,
where I and ( J , K ) , respectively, denote the total numbers of input layer units and mapping layer units.
A neighboring region ψ ( t ) is set from the center of c j as
ψ ( t ) = ψ ( 0 ) · J · 1 t O + 0.5 ,
where O represents the maximum of learning iterations. Subsequently, w i j k ( t ) in ψ ( t ) is updated as
w i j k ( t + 1 ) = w i j k ( t ) + α ( x i ( t ) w i j k ( t ) ) ,
where α is a learning coefficient that decreases according to the learning progress. Herein, at time t = 0 , we initialized w i j k ( 0 ) with random numbers.

4.2. U-Matrix

U-Matrix [86] is used for extracting cluster boundaries from w i j k . Based on metric distances between weights, U-Matrix visualizes the spatial distribution of categories from the similarity of neighbor units [86]. On a two-dimensional (2D) category map of square grids, a unit has eight neighbor units, except for boundary units. Letting U denote the similarity calculated using U-Matrix. For the component of the horizontal and vertical directions, U h ± and U v ± are defined as shown below.
U h ± = i = 1 I ( w i j k w i j ± 1 k ) 2 , U v ± = i = 1 I ( w i j k w i j k ± 1 ) 2 .
For components of the diagonal directions, U d ± are defined as presented below.
U d ± = 1 2 i = 1 I ( w i j k ± 1 w i j ± 1 k ) 2 + 1 2 1 = 1 I ( w i j ± 1 k w i j k ± 1 ) 2

4.3. Recurrent SOM

Our method uses RSOM for extracting smile images from time-series facial expression images [87]. As a derivative model of SOM [26], RSOM [84] incorporates an additional feedback loop for learning time-series features. Temporally changed input signals are mapped into units on the competitive layer. β 1 , β 2 , β 3 were set as denoting learning coefficients. The output y j k ( t ) from the mapping unit (j, k) at time t is presented as the following.
y j k ( t ) = β 1 y j k ( t ) + β 2 y j k ( t 2 ) + β 3 ( x i ( t ) w i j k ( t ) ) .
The weights are updated as
w i j k ( t + 1 ) = ( 1 γ ) w i j k ( t ) + γ ( x ( t ) w i j k ( t ) ) y j k ( t ) ,
where γ is a learning coefficient that decreases according to the learning progress.
The RSOM mapping size is set in advance with the number of units. This parameter controls the classification granularity of facial expression images. This method affixed 15 units based on the setting parameter of FESC in our earlier study [77].

4.4. GHSOM

As an extended SOM network and its training algorithm, GHSOM [85] incorporates a hierarchization mechanism that accommodates an increased number of mapping layers. An appropriate mapping size for solving a target problem is obtainable automatically by GHSOM. Although the weight update mechanism of GHSOM resembles that of SOM, the learning algorithm of GHSOM includes the generation of a hierarchical structure based on growing and adding mapping units in each layer, except the top layer. The respective GHSOM layers provide parallel learning as independent modules.
The growing hierarchical algorithm is launched from the top layer, which comprises a single unit [88]. Letting w 0 denote a weight between the top layer and the next layer. The top layer, which includes no growing mechanism, branches into four sub-layers. All sublayers have 2 × 2 mapping units. Growing hierarchical learning is actualized on the units of the sublayers. Letting v i denote a standard deviation for the input x i to the mapping units of the i-th sub-layer. The mean standard deviation v m is calculated as presented below.
v m = 1 I i = 0 I j = 0 J k = 0 K ( w i j k x i ) .
Letting T m represent the breadth threshold. Hierarchical growing is controlled by T m , as presented below.
T m < v m v m 1 .
A unit for growing is appended if the ratio between v m of the m-th layer and v m 1 of the ( m 1 )th layer is greater than T m .
The hierarchical growing procedure for adding new units comprises four steps. The first step is the specification of an error unit u e that indicates the maximum standard deviation between units. The second step is the selection of a dissimilar unit u d that indicates a minimum standard deviation from neighboring units around u e . The third step is the insertion of a new unit between u e and u d . The fourth step is updating of weights of the respective units based on the SOM learning algorithm. After learning, input features are classified again. The standard deviation is decreasing according to the growing progress. The growth termination is triggered by saturation of the added units as a suitable mapping size. After adding units, the addition of new layers is processed. Finally, the learning phase is completed if growing is terminated.

4.5. Parameters

Table 2 denotes the meta-parameters of SOM, RSOM, and GHSOM and their initial setting values. We set them based on our earlier study [23]. The parameter I is changed according to the input dimensions in each experiment.
Table 2. Meta-parameters and initial setting values.

5. Experiment Results

5.1. Unsupervised Classification Results of Mood States

Figure 6 presents the TMD distribution on the horizontal axis and F scores on the vertical axis. These scores are calculated from POMS2 T-scores. They can therefore be 120 plots from the dataset denoted in Table 1. The intersection of the axes corresponds to the mean scores of TMD and F obtained from [89]. On the one hand, small and large TMD scores can be interpreted, respectively, as positive and negative mental states. On the other hand, small and large F scores can be interpreted, respectively, as negative and positive mental states.
Figure 6. The 2D-distribution of TMD and F calculated from POMS2 T-scores for all subjects.
Based on unsupervised clustering of the data plots, a category map was created with SOM. Figure 7 shows the result with categorical boundaries extracted from U-Matrix. The brightness represents the depth of categorical boundaries. Lower and higher brightness scale values, respectively, indicate deeper and shallower boundaries. Deeper category boundaries appeared in the upper-left and bottom-right areas on the map. These boundaries divided the category map into three independent regions. Moreover, three categories were extracted from the left half, upper right half, and the bottom right half in the category enclosed by the solid yellow border. Regarding the relation between this classification result and the distribution in Figure 6, five categories labeled Categories A–E were obtained from Figure 7.
Figure 7. Category boundary and unsupervised classification results with U-matrix from the input of TMD and F.
Figure 8 depicts a classification result of the coordinate points in Figure 6 based on the five categories extracted from Figure 7. Category A is defined semantically as positive, which is attributable to their low TMD and high F scores. By contrast, Category E is defined semantically as negative, which is attributable to their high TMD and low F scores. Based on the positional relations, the three categories distributed around the center are defined, respectively, as positive for Category B, negative for Category D, and neutral for Category C. Fundamentally, we affixed these semantic labels based on the vertical axis associated with TMD scores. The decision boundary is located around 51 points, which is lower than the mean score of 55 points. We referred to F scores to affix semantic labels for Categories B and C, which are located in similar TMD ranges. The decision boundary is located near 52 points, which is higher than the mean score of 49 points. The data shown for respective subjects, as presented in Table 1, include intra-categorical and inter-categorical distribution patterns. For subjects with widely diverse mood states, the data plots on the horizontal axis tend to be long. By contrast, for subjects with a narrow range of mood states, data plots on the horizontal axis tend to be short. Based on these parameters, we analyzed the relation between representative subject data obtained from POMS2 and the eye-tracking device.
Figure 8. Unsupervised classification results for the data in Figure 6 based on Figure 7.

5.2. Relation between Mood State and Gaze Distribution

This evaluation experiment yielded representative results obtained from analyzing the relation between mood states and gaze distribution features, including saccades for six subjects: Subjects H, F, I, C, B, and O, in that order. Figure 9 presents the results obtained for Subject H. The mode states plots are distributed inside Category A. The heatmap results show the gaze distribution and its density gathered around the interlocutor’s face on the monitor. The tendency with small changes represents the steady mood states and gaze distribution to specific areas. The number of saccades is smaller than those of other subjects.
Figure 9. Mood states (Left) and gaze distribution (Right) for Subject H in Category A.
Figure 10 presents experiment results for Subject F. Unlike Subject H, the plots of mood states are distributed over the whole of Category A. The distribution range of F is greater than that of TMD. The experimentally obtained results demonstrate not only wider gaze distribution but also a greater number of saccades compared with those of Subject H.
Figure 10. Mood states (Left) and gaze distribution (Right) for Subject F in Category A.
Figure 11 presents experimentally obtained results for Subject I. The plots of mood states are distributed in the upper half of Category B. Although the gaze was concentrated on the interlocutor’s face on the monitor, the distribution shape was spread vertically. A high number of saccades indicates severe vertical movements of the eyes.
Figure 11. Mood states (Left) and gaze distribution (Right) for Subject I in Category B.
Figure 12 portrays experimentally obtained results for Subject C. The plots of mood states are distributed in two parts in Category C. The gaze distribution was unstable, with two of the four cases extending their range laterally. The heatmap results demonstrate that the gaze distribution of three of the four cases is divided into two clusters. The number of saccades increased with the expansion of the gaze area. Although the mood states differed from those of the other three samples in the distribution of C-1, no characteristic association with gaze was identified.
Figure 12. Mood states (left) saccade distributions (right) for Subject C in Category C.
Figure 13 depicts experiment results for Subject B. The distribution of the mood states is placed in Category D. The range of TMD and F suggests that the changes in mood states for this subject are narrow. The high-temperature heatmap results indicate that the gaze plots are gathered densely to the interlocutor’s face on the monitor. The number of saccades is smaller than for the other subjects. We consider that gaze movements are steady for questioners that are attributable to a lack of mood state changes.
Figure 13. Mood states (left) saccade distributions (right) for Subject B in Category D.
Figure 14 portrays the results obtained for Subject O. The mood states are distributed in three categories: Categories B, D, and E. Although the gaze plots are distributed widely, their concentrating area is narrow. High-temperature areas on the heatmap are gathered for the interlocutor, except for the O-3 example. We infer that gaze distribution patterns affected the mood states of this subject. The number of saccades tended to increase with wide eye movements and decrease with narrow eye movements.
Figure 14. Mood states (left) saccade distributions (right) for Subject O in Categories B, D, and E.

5.3. Smile Expression Extraction

Face regions were extracted using the Viola-Jones [90] method, which is a dominant object detection framework based on Haar-like features combined with Ada-boost cascading classifiers. Regarding the camera position, view angle, and resolutions, we extracted a fixed region of interest (RoI) of 320 × 320 pixels. The final purpose of this experiment is to visualize mental health displayed on a 2D map created from several feature combinations.
The RSOM module extracted smile-expressed frames from time-series facial images. Figure 15 presents extraction results for the representative three subjects: Subjects H, B, and O. Positive images and negative images, respectively, correspond to the smile expressions and blank expressions. The red frames show GT images labeled as smile expressions. Although mismatched images exist among the frames of low expression intensity, our method globally extracted smile images concomitantly with the GT frames.
Figure 15. Extraction result of smile expression images for three subjects: Subjects H, B, and O.
For annotators, classifying images that switch facial expressions is a difficult task. Moreover, annotating the facial expression images of women is more difficult than annotating those of men. The accuracy of extracting smile images with RSOM demonstrated that over 90% was similar to the accuracy obtained from our earlier study. The odd frames and even frames were set, respectively, to training and validation subsets. Gabor wavelets transformations [91] were applied to input data images of 160 × 160 pixels. Based on FESC [77], 15 weights corresponding to 15 units obtained from RSOM were attended to GHSOM.
Smile images were classified hierarchically using GHSOM using RSOM weights. Figure 16 depicts unsupervised classification results, presented as tree structures for Subjects H, B, and O. The RSOM mapping layer size was set to 15 units, which divides weight into 15 clusters. The maximum granularity was set to four categories in each layer. The weights of Subject H were divided into four clusters in the first depth layer. The weights of Subject B were divided into five clusters. In the fourth cluster, two weights were categorized in the second depth layer. The weights of Subject O were divided into six clusters. In the first cluster, four weights were categorized into two clusters in the second depth layer.
Figure 16. Hierarchical unsupervised classification results with RSOM weights with GHSOM for three subjects: H, B, and O.

5.4. Effects of Input Features on Visualization Results

This experiment was conducted to verify the relation between gaze patterns and facial expressions that affect changes in mood states. For experimentation, we used the visualization modules based on SOM and U-Matrix. We set E, S, R, and G as, respectively, denoting the number of pixels extracted from gaze movements, the number of saccades, the number of smile images obtained from RSOM, and the number of categories obtained from GHSOM. This experiment provides three input patterns based on TMD and F combined with E, S, R, and G, as presented below.
  • Input I: TMD + F + E + S
  • Input II: TMD + F + R + G
  • Input III: TMD + F + E + S + R + G
The combination of these input features provides different distribution patterns on category maps as visualization results. Figure 17a presents unsupervised classification results for Input I. Annotation labels, which correspond to the categories in Figure 8, are superimposed on the category map. Categories A and D were allocated, respectively, to the bottom left and upper right on the map. Categorical boundaries appeared to be deeper and more continuous than the result obtained for the input of TMD and F in Figure 7. Unsupervised classification results for Input I demonstrated that the gaze-related features are useful to delineate the boundaries of the categories, especially in Category A. Figure 17b depicts the classification result obtained for Input II. Categories A and D were allocated, respectively, to the bottom left and upper right on the map. Figure 17c depicts unsupervised classification results obtained for Input III. Categories A and D were allocated, respectively, to the bottom and upper right on the map. Categorical boundaries appeared to be deeper and more continuous than the results obtained for Inputs I and II.
Figure 17. Unsupervised classification results obtained for three input patterns. Red curves are drawn subjectively as category boundaries.

6. Conclusions

This paper presented a method of classification and visualization of mood states obtained from a psychological check sheet and facial features of gaze, saccades, and facial expressions based on unsupervised ML algorithms. The two indicators TMD and F obtained from POMS2 were classified into five categories using SOM and U-Matrix. Relations between gaze and facial expressions were analyzed from the five extracted categories. Subjects in positive categories demonstrated positive correlations between gaze concentration areas colored with high-temperature heatmaps and the number of saccades. In particular, for subjects with widely diverse mood states, the gaze data distributions on the horizontal axis tend to be long. By contrast, for subjects with a narrow range of mood states, gaze data distributions on the horizontal axis tend to be short. Regarding facial expressions, positive category subjects had a constant expression time of intentional smiles. By contrast, subjects in the negative categories exhibited a time length difference in intentional smiles. Furthermore, we examined the influences of gaze and facial expressions on category classification using RSOM and GHSOM. The results obtained from three comparative experiments indicated that adding features of gaze and facial expression to TMD and F clarified the category boundaries obtained from the U-Matrix. Compared to the result obtained for the input of TMD and F, categorical boundaries appeared to be deeper and more continuous using features of the number of pixels extracted from gaze movements, the number of saccades, the number of smile images obtained from RSOM, and the number of categories obtained from GHSOM. We verify that the use of SOM, RSOM, and GHSOM is the best combination for the visualization of mood states.
In future work, we would like to actualize stress estimation only from gaze or facial expressions. The only subjects for this experiment were university students in their 20s. We would like to expand the application range of the proposed method, especially for a wider age range of subjects. Moreover, we plan to develop apps for tablet computers and smartphones to facilitate the practical application of this method.

Author Contributions

Conceptualization, K.S.; methodology, H.M.; software, K.S.; validation, K.S.; formal analysis, S.N.; investigation, S.N.; resources, H.M.; data curation, H.M.; writing—original draft preparation, H.M.; writing—review and editing, H.M.; visualization, S.N.; supervision, K.S.; project administration, K.S. All authors have read and agreed to the published version of the manuscript.

Funding

This research received no external funding.

Institutional Review Board Statement

The study was conducted according to the guidelines approved by the Institutional Review Board of Akita Prefectural University (Kendaiken–325; 28 August 2017).

Data Availability Statement

Datasets presented as a result of this study are available on request to the corresponding author.

Acknowledgments

We would like to express our appreciation to Emi Katada and to Akari Santa, both graduates of Akita Prefectural University, for their great cooperation with experiments.

Conflicts of Interest

The authors declare that they have no conflict of interest. The funders had no role in the design of the study, in the collection, analyses, or interpretation of data, in the writing of the manuscript, or in the decision to publish the results.

Abbreviations

The following abbreviations are used in this report:
2DTwo-dimensional
DLDeep learning
EEGElectroencephalogram
FESCFacial expression spatial charts
GHSOMGrowing hierarchical self-organizing maps
GTGround truth
HbHemoglobin
HRHeart rate
MLMachine-learning
NIRSNear-infrared spectroscopy
POMS2Profile of mood states second edition
RSOMRecurrent self-organizing maps
RoIRegion of interest
SNSSocial networking services
SOMSelf-organizing maps
TMDTotal mood distance

References

  1. Toscano, F.; Zappalá, S. Social Isolation and Stress as Predictors of Productivity Perception and Remote Work Satisfaction during the COVID-19 Pandemic: The Role of Concern about the Virus in a Moderated Double Mediation. Sustainability 2020, 12, 9804. [Google Scholar] [CrossRef]
  2. Ferri, F.; Grifoni, P.; Guzzo, T. Online Learning and Emergency Remote Teaching: Opportunities and Challenges in Emergency Situations. Societies 2020, 10, 86. [Google Scholar] [CrossRef]
  3. Muller, A.M.; Goh, C.; Lim, L.Z.; Gao, X. COVID-19 Emergency eLearning and Beyond: Experiences and Perspectives of University Educators. Educ. Sci. 2021, 11, 19. [Google Scholar] [CrossRef]
  4. Maican, M.-A.; Cocoradá, E. Online Foreign Language Learning in Higher Education and Its Correlates during the COVID-19 Pandemic. Sustainability 2021, 13, 781. [Google Scholar] [CrossRef]
  5. Joia, L.A.; Lorenzo, M. Zoom In, Zoom Out: The Impact of the COVID-19 Pandemic in the Classroom. Sustainability 2021, 13, 2531. [Google Scholar] [CrossRef]
  6. Madden, S. From Databases to Big Data. IEEE Internet Comput. 2012, 16, 4–6. [Google Scholar] [CrossRef]
  7. Fu, W.; Liu, S.; Srivastava, G. Optimization of Big Data Scheduling in Social Networks. Entropy 2019, 21, 902. [Google Scholar] [CrossRef] [Green Version]
  8. Obar, J.A.; Oeldorf-Hirsch, A. The biggest lie on the internet: Ignoring the privacy policies and terms of service policies of social networking services. Inf. Commun. Soc. 2020, 23, 1, 128–147. [Google Scholar] [CrossRef]
  9. Chawla, R. Deepfakes: How a pervert shook the world. Int. J. Adv. Res. Dev. 2019, 4, 4–8. [Google Scholar]
  10. LeCun, Y.; Bengio, Y.; Hinton, G. Deep learning. Nature 2015, 521, 436–444. [Google Scholar] [CrossRef]
  11. Westerlund, M. The Emergence of Deepfake Technology: A Review. Technol. Innov. Manag. Rev. 2019, 9, 40–53. [Google Scholar] [CrossRef]
  12. Yadav, D.; Salmani, S. Deepfake: A Survey on Facial Forgery Technique Using Generative Adversarial Network. In Proceedings of the 2019 International Conference on Intelligent Computing and Control Systems (ICCS), Madurai, India, 15–17 May 2019; pp. 852–857. [Google Scholar]
  13. Yu, P.; Xia, Z.; Fei, J.; Lu, Y. A Survey on Deepfake Video Detection. IET Biom. 2021, 10, 607–624. [Google Scholar] [CrossRef]
  14. Kietzmann, J.; Lee, L.W.; McCarthy, I.P.; Kietzmann, T.C. Deepfakes: Trick or treat? Bus. Horizons 2020, 63, 135–146. [Google Scholar] [CrossRef]
  15. Xu, L.D.; Xu, E.L.; Li, L. Industry 4.0: State of the art and future trends. Int. J. Prod. Res. 2018, 56, 2941–2962. [Google Scholar] [CrossRef] [Green Version]
  16. Shpak, N.; Kuzmin, O.; Dvulit, Z.; Onysenko, T.; Sroka, W. Digitalization of the Marketing Activities of Enterprises: Case Study. Information 2020, 11, 109. [Google Scholar] [CrossRef] [Green Version]
  17. Cioffi, R.; Travaglioni, M.; Piscitelli, G.; Petrillo, A.; De Felice, F. Artificial Intelligence and Machine Learning Applications in Smart Production: Progress, Trends, and Directions. Sustainability 2020, 12, 492. [Google Scholar] [CrossRef] [Green Version]
  18. Royakkers, L.; Timmer, J.; Kool, L.; van Est, R. Societal and ethical issues of digitization. Ethics Inf. Technol. 2018, 20, 127–142. [Google Scholar] [CrossRef] [Green Version]
  19. McEwen, B.; Bowles, N.; Gray, J.; Hill, M.N.; Hunter, R.G.; Karatsoreos, I.N.; Nasca, C. Mechanisms of stress in the brain. Nat. Neurosci. 2015, 18, 1353–1363. [Google Scholar] [CrossRef]
  20. Meaney, J.M. Maternal Care, Gene Expression, and the Transmission of Individual Differences in Stress Reactivity Across Generations. Annu. Rev. Neurosci. 2001, 24, 1161–1192. [Google Scholar] [CrossRef]
  21. Piardi, L.N.; Pagliusi, M.; Bonet, I.J.M.; Brandáo, A.F.; Magalháes, S.F.; Zanelatto, F.B.; Tambeli, C.H.; Parada, C.A.; Sartori, C.R. Social stress as a trigger for depressive-like behavior and persistent hyperalgesia in mice: Study of the comorbidity between depression and chronic pain. J. Affect. Disord. 2020, 274, 759–767. [Google Scholar] [CrossRef]
  22. Tsutsumi, A.; Shimazu, A.; Eguchi, H.; Inoue, A.; Kawakami, N. A Japanese Stress Check Program screening tool predicts employee long-term sickness absence: A prospective study. J. Occup. Health 2018, 60, 55–63. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  23. Katada, E.; Sato, K.; Madokoro, H.; Kadowaki, S. Visualization and Meanings of Mental Health State Based on Mutual Interactions Between Eye-gaze and Lip Movements. In Proceedings of the International Symposium on Intelligent Signal Processing and Communication Systems (ISPACS), Ishigaki, Japan, 27–30 November 2018; pp. 18–23. [Google Scholar]
  24. Manga, M.; Beresnev, I.; Brodsky, E.E.; Elkhoury, J.E.; Elsworth, D.; Ingebritsen, S.E.; Mays, D.C.; Wang, C.Y. Changes in permeability caused by transient stresses: Field observations, experiments, and mechanisms. Rev. Geophys. 2012, 50, 2. [Google Scholar] [CrossRef]
  25. Pozos-Radillo, B.E.; Preciado-Serrano, M.L.; Acosta-Fernández, M.; Aguilera-Velasco, M.Á.; Delgado-Garcia, D.D. Academic stress as a predictor of chronic stress in university students. Psicol. Educ. 2014, 20, 47–52. [Google Scholar] [CrossRef] [Green Version]
  26. Kohonen, T. Self-organized formation of topologically correct feature maps. Biol. Cybern. 1982, 43, 59–69. [Google Scholar] [CrossRef]
  27. Flexer, A. On the use of self-organizing maps for clustering and visualization. Intell. Data Anal. 2001, 5, 373–384. [Google Scholar] [CrossRef]
  28. Galvan, D.; Effting, L.; Cremasco, H.; Adam Conte-Junior, C. Can Socioeconomic, Health, and Safety Data Explain the Spread of COVID-19 Outbreak on Brazilian Federative Units? Int. J. Environ. Res. Public Health 2020, 17, 8921. [Google Scholar] [CrossRef]
  29. Galvan, D.; Effting, L.; Cremasco, H.; Conte-Junior, C.A. The Spread of the COVID-19 Outbreak in Brazil: An Overview by Kohonen Self-Organizing Map Networks. Medicina 2021, 57, 235. [Google Scholar] [CrossRef]
  30. Feng, S.; Li, H.; Ma, L.; Xu, Z. An EEG Feature Extraction Method Based on Sparse Dictionary Self-Organizing Map for Event-Related Potential Recognition. Algorithms 2020, 13, 259. [Google Scholar] [CrossRef]
  31. Locati, L.D.; Serafini, M.S.; Iannó, M.F.; Carenzo, A.; Orlandi, E.; Resteghini, C.; Cavalieri, S.; Bossi, P.; Canevari, S.; Licitra, L.; et al. Mining of Self-Organizing Map Gene-Expression Portraits Reveals Prognostic Stratification of HPV-Positive Head and Neck Squamous Cell Carcinoma. Cancers 2019, 11, 1057. [Google Scholar] [CrossRef] [Green Version]
  32. Melin, P.; Castillo, O. Spatial and Temporal Spread of the COVID-19 Pandemic Using Self Organizing Neural Networks and a Fuzzy Fractal Approach. Sustainability 2021, 13, 8295. [Google Scholar] [CrossRef]
  33. Andrades, I.S.; Castillo Aguilar, J.J.; García, J.M.V.; Carrillo, J.A.C.; Lozano, M.S. Low-Cost Road-Surface Classification System Based on Self-Organizing Maps. Sensors 2020, 20, 6009. [Google Scholar] [CrossRef] [PubMed]
  34. Mayaud, J.R.; Anderson, S.; Tran, M.; Radić, V. Insights from Self-Organizing Maps for Predicting Accessibility Demand for Healthcare Infrastructure. Urban Sci. 2019, 3, 33. [Google Scholar] [CrossRef] [Green Version]
  35. Lee, J.; Kim, J.; Ko, W. Day-Ahead Electric Load Forecasting for the Residential Building with a Small-Size Dataset Based on a Self-Organizing Map and a Stacking Ensemble Learning Method. Appl. Sci. 2019, 9, 1231. [Google Scholar] [CrossRef] [Green Version]
  36. Blanco-M., A.; Gibert, K.; Marti-Puig, P.; Cusidó, J.; Solé-Casals, J. Identifying Health Status of Wind Turbines by Using Self Organizing Maps and Interpretation-Oriented Post-Processing Tools. Energies 2018, 11, 723. [Google Scholar] [CrossRef] [Green Version]
  37. Angulo-Saucedo, G.A.; Leon-Medina, J.X.; Pineda-Muńoz, W.A.; Torres-Arredondo, M.A.; Tibaduiza, D.A. Damage Classification Using Supervised Self-Organizing Maps in Structural Health Monitoring. Sensors 2022, 22, 1484. [Google Scholar] [CrossRef]
  38. Bui, X.K.; Marlim, M.S.; Kang, D. Optimal Design of District Metered Areas in a Water Distribution Network Using Coupled Self-Organizing Map and Community Structure Algorithm. Water 2021, 13, 836. [Google Scholar] [CrossRef]
  39. Ying, Y.; Li, Z.; Yang, M.; Du, W. Multimode Operating Performance Visualization and Nonoptimal Cause Identification. Processes 2020, 8, 123. [Google Scholar] [CrossRef] [Green Version]
  40. Lu, D.; Tian, Y.; Liu, V.Y.; Zhang, Y. The Performance of the Smart Cities in China—A Comparative Study by Means of Self-Organizing Maps and Social Networks Analysis. Sustainability 2015, 7, 7604–7621. [Google Scholar] [CrossRef] [Green Version]
  41. Taczanowska, K.; González, L.-M.; García-Massó, X.; Zieba, A.; Brandenburg, C.; Muhar, A.; Pellicer-Chenoll, M.; Toca-Herrera, J.-L. Nature-based Tourism or Mass Tourism in Nature? Segmentation of Mountain Protected Area Visitors Using Self-Organizing Maps (SOM). Sustainability 2019, 11, 1314. [Google Scholar] [CrossRef] [Green Version]
  42. Molina-García, J.; García-Massó, X.; Estevan, I.; Queralt, A. Built Environment, Psychosocial Factors and Active Commuting to School in Adolescents: Clustering a Self-Organizing Map Analysis. Int. J. Environ. Res. Public Health 2019, 16, 83. [Google Scholar] [CrossRef] [Green Version]
  43. Chun, E.; Jun, S.; Lee, C. Identification of Promising Smart Farm Technologies and Development of Technology Roadmap Using Patent Map Analysis. Sustainability 2021, 13, 10709. [Google Scholar] [CrossRef]
  44. Penfound, E.; Vaz, E. Analysis of Wetland Landcover Change in Great Lakes Urban Areas Using Self-Organizing Maps. Remote Sens. 2021, 13, 4960. [Google Scholar] [CrossRef]
  45. Anda, A.; Simon-Gáspár, B.; Soós, G. The Application of a Self-Organizing Model for the Estimation of Crop Water Stress Index (CWSI) in Soybean with Different Watering Levels. Water 2021, 13, 3306. [Google Scholar] [CrossRef]
  46. Li, B.; Watanabe, K.; Kim, D.-H.; Lee, S.-B.; Heo, M.; Kim, H.-S.; Chon, T.-S. Identification of Outlier Loci Responding to Anthropogenic and Natural Selection Pressure in Stream Insects Based on a Self-Organizing Map. Water 2016, 8, 188. [Google Scholar] [CrossRef] [Green Version]
  47. Salas, I.; Cifrian, E.; Andres, A.; Viguri, J.R. Self-Organizing Maps to Assess the Recycling of Waste in Ceramic Construction Materials. Appl. Sci. 2021, 11, 10010. [Google Scholar] [CrossRef]
  48. Wang, L.; Li, Z.; Wang, D.; Hu, X.; Ning, K. Self-Organizing Map Network-Based Soil and Water Conservation Partitioning for Small Watersheds: Case Study Conducted in Xiaoyang Watershed, China. Sustainability 2020, 12, 2126. [Google Scholar] [CrossRef] [Green Version]
  49. Dixit, N.; McColgan, P.; Kusler, K. Machine Learning-Based Probabilistic Lithofacies Prediction from Conventional Well Logs: A Case from the Umiat Oil Field of Alaska. Energies 2020, 13, 4862. [Google Scholar] [CrossRef]
  50. Riese, F.M.; Keller, S.; Hinz, S. Supervised and Semi-Supervised Self-Organizing Maps for Regression and Classification Focusing on Hyperspectral Data. Remote Sens. 2020, 12, 7. [Google Scholar] [CrossRef] [Green Version]
  51. Thurn, N.; Williams, M.R.; Sigman, M.E. Application of Self-Organizing Maps to the Analysis of Ignitable Liquid and Substrate Pyrolysis Samples. Separations 2018, 5, 52. [Google Scholar] [CrossRef] [Green Version]
  52. Li, J.; Shi, Z.; Wang, G.; Liu, F. Evaluating Spatiotemporal Variations of Groundwater Quality in Northeast Beijing by Self-Organizing Map. Water 2020, 12, 1382. [Google Scholar] [CrossRef]
  53. Tsuchihara, T.; Shirahata, K.; Ishida, S.; Yoshimoto, S. Application of a Self-Organizing Map of Isotopic and Chemical Data for the Identification of Groundwater Recharge Sources in Nasunogahara Alluvial Fan, Japan. Water 2020, 12, 278. [Google Scholar] [CrossRef] [Green Version]
  54. Inaba, A. Marital Status and Psychological Distress in Japan. Jpn. Sociol. Rev. 2002, 53, 69–84. [Google Scholar] [CrossRef]
  55. Araki, S.; Kawakami, N. Health Care of Work Stress: A review. J. Occup. Health 1993, 35, 88–97. [Google Scholar]
  56. Kakamu, T.; Tsuji, M.; Hidaka, T.; Kumagai, T.; Hayakawa, T.; Fukushima, T. The Relationship between Fatigue Recovery after Late-night shifts and Stress Relief Awareness. J. Occup. Health 2014, 56, 116–120. [Google Scholar]
  57. Hernando, D.; Roca, S.; Sancho, J.; Alesanco, Á.E.; Bailón, R. Validation of the Apple Watch for Heart Rate Variability Measurements during Relax and Mental Stress in Healthy Subjects. Sensors 2018, 18, 2619. [Google Scholar] [CrossRef] [Green Version]
  58. De Pero, R.; Minganti, C.; Cibelli, G.; Cortis, C.; Piacentini, M.F. The Stress of Competing: Cortisol and Amylase Response to Training and Competition. J. Funct. Morphol. Kinesiol. 2021, 6, 5. [Google Scholar] [CrossRef]
  59. Takatsu, H.; Munakata, M.; Ozaki, O.; Tokoyama, K.; Watanabe, Y.; Takata, K. An Evaluation of the Quantitative Relationship between the Subjective Stress Value and Heart Rate Variability. IEEJ Trans. Electron. Inf. Syst. 2000, 120, 104–110. [Google Scholar]
  60. Matsumoto, Y.; Mori, N.; Mitajiri, R.; Jiang, Z. Study of Mental Stress Evaluation based on analysis of Heart Rate Variability. J. Life Support Eng. 2010, 22, 105–111. [Google Scholar] [CrossRef] [Green Version]
  61. Daudelin-Peltier, C.; Forget, H.; Blais, C.; Deschénes, A.; Fiset, D. The effect of acute social stress on the recognition of facial expression of emotions. Sci. Rep. 2017, 7, 1036. [Google Scholar] [CrossRef]
  62. Zhang, H.; Feng, L.; Li, N.; Jin, Z.; Cao, L. Video-Based Stress Detection through Deep Learning. Sensors 2020, 20, 5552. [Google Scholar] [CrossRef]
  63. Gavrilescu, M.; Vizireanu, N. Predicting Depression, Anxiety, and Stress Levels from Videos Using the Facial Action Coding System. Sensors 2019, 19, 3693. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  64. Wei, H.; Hauer, R.J.; Chen, X.; He, X. Facial Expressions of Visitors in Forests along the Urbanization Gradient: What Can We Learn from Selfies on Social Networking Services? Forests 2019, 10, 1049. [Google Scholar] [CrossRef] [Green Version]
  65. Firth, J.; Torous, J.; Nicholas, J.; Carney, R.; Rosenbaum, S.; Sarris, J. Can smartphone mental health interventions reduce symptoms of anxiety? A meta-analysis of randomized controlled trials. J. Affect. Disord. 2017, 218, 15–22. [Google Scholar] [CrossRef]
  66. Terry, N.P.; Gunter, T.D. Regulating mobile mental health apps. Behav. Sci. 2018, 36, 136–144. [Google Scholar] [CrossRef] [Green Version]
  67. Alreshidi, A.; Ullah, M. Facial Emotion Recognition Using Hybrid Features. Informatics 2020, 7, 6. [Google Scholar] [CrossRef] [Green Version]
  68. Berge, J.M.; Tate, A.; Trofholz, A.; Fertig, A.; Crow, S.; Neumark-Sztainer, D.; Miner, M. Examining within- and across-day relationships between transient and chronic stress and parent food-related parenting practices in a raciallyethnically diverse and immigrant population. Int. J. Behav. Nutr. Phys. Acta 2018, 15, 7. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  69. Hutmacher, F. Why Is There So Much More Research on Vision Than on Any Other Sensory Modality? Front. Psychol. 2019, 10, 2246. [Google Scholar] [CrossRef]
  70. Deubel, H. Localization of targets across saccades: Role of landmark objects. Vis. Cogn. 2004, 11, 173–202. [Google Scholar] [CrossRef]
  71. Mizushina, H.; Sakamoto, K.; Kaneko, H. The Relationship between Psychological Stress Induced by Task Workload and Dynamic Characteristics of Saccadic Eye Movements. IEICE Trans. Inf. Syst. 2011, 94, 1640–1651. [Google Scholar]
  72. Iizuka, Y. On the relationship of gaze to emotional expression. Jpn. J. Exp. Soc. Psychol. 1991, 31, 147–154. [Google Scholar] [CrossRef] [Green Version]
  73. Ekman, P.; Oster, H. Facial Expressions of Emotion. Annu. Rev. Psychol. 1979, 30, 527–554. [Google Scholar] [CrossRef]
  74. Pantic, M.; Rothkrantz, L.J.M. Automatic analysis of facial expressions: The state of the art. IEEE Trans. Pattern Anal. Mach. Intell. 2000, 22, 1424–1445. [Google Scholar] [CrossRef] [Green Version]
  75. Akamatsu, S. Recognition of Facial Expressions by Human and Computer. I: Facial Expressions in Communications and Their Automatic Analysis by Computer. J. Inst. Electron. Inf. Commun. Eng. 2002, 85, 680–685. [Google Scholar]
  76. Sato, K.; Otsu, H.; Madokoro, H.; Kadowaki, S. Analysis of Psychological Stress Factors and Facial Parts Effect on Intentional Facial Expressions. In Proceedings of the Third International Conference on Ambient Computing, Applications, Services and Technologies, Porto, Portugal, 29 September–4 October 2013; pp. 7–16. [Google Scholar]
  77. Madokoro, H.; Sato, K.; Kawasumi, A.; Kadowaki, S. Facial Expression Spatial Charts for Representing of Dynamic Diversity of Facial Expressions. In Proceedings of the IEEE International Conference on Systems, Man and Cybernetics (SMC), San Antonio, TX, USA, 11–14 October 2009; pp. 2884–2889. [Google Scholar]
  78. Sato, K.; Otsu, H.; Madokoro, H.; Kadowaki, S. Analysis of psychological stress factors by using a Bayesian network. In Proceedings of the IEEE International Conference on Mechatronics and Automation, Takamatsu, Japan, 4–7 August 2013; pp. 811–818. [Google Scholar]
  79. Hamada, M.; Fukuzoe, T.; Watanabe, M. The Research of Estimating Human’s Internal Status by Behavior and Expression Analysis. IPSJ SIG Tech. Rep. 2007, 158, 77–84. [Google Scholar]
  80. Arita, M.; Ochi, H.; Matsufuji, T.; Sakamoto, H.; Fukushima, S. An Estimation Method of Emotion from Facial Expression and Physiological Indices by CCA Using Fuzzy Membership Degree. In Proceedings of the 27th Fuzzy System Symposium, Fukui, Japan, 12–14 September 2011. [Google Scholar]
  81. Ueda, S.; Suga, T. The Effect of Facial Individuality on Facial Impression. Trans. Jpn. Acad. Facial Stud. 2006, 6, 17–24. [Google Scholar]
  82. Heuchert, J.P.; McNair, D.M. POMS 2: Profile of Mood States, 2nd ed.; Multi-Health Systems Inc.: Tonawanda, NY, USA, 2012. [Google Scholar]
  83. Konuma, H.; Hirose, H.; Yokoyama, K. Relationship of the Japanese Translation of the Profile of Mood States Second Edition (POMS 2) to the First Edition (POMS). Juntendo Med. J. 2015, 61, 517–519. [Google Scholar] [CrossRef] [Green Version]
  84. Koskela, T.; Varsta, M.; Heikkonen, J.; Kaski, K. Temporal sequence processing using recurrent SOM. In Proceedings of the Second International Conference, Knowledge-Based Intelligent Electronic Systems, Adelaide, Australia, 21–23 April 1998; pp. 290–297. [Google Scholar]
  85. Dittenbach, M.; Merkl, D.; Rauber, A. The growing hierarchical self-organizing map. In Proceedings of the IEEE-INNS-ENNS International Joint Conference on Neural Networks, Como, Italy, 27–27 July 2000; pp. 15–19. [Google Scholar]
  86. Ultsch, A. Clustering with SOM: U*C. In Proceedings of the Workshop on Self-Organizing Maps, Paris, France, 5–8 September 2005; pp. 75–82. [Google Scholar]
  87. Varsta, M.; Heikkonen, J.; Millan, J.R. Context learning with the self organizing map. In Proceedings of the WSOM’97, Workshop on Self-Organizing Maps, Helsinki, Finland, 4–6 June 1997. [Google Scholar]
  88. Yeloglu, O.; Zincir-Heywood, A.N.; Heywood, M.I. Growing recurrent self organizing map. In Proceedings of the 2007 IEEE International Conference on Systems, Man and Cybernetics, Montreal, QC, Canada, 7–10 October 2007; pp. 290–295. [Google Scholar]
  89. Itoi, Y.; Ogawa, E.; Shimizu, T.; Furuyama, A.; Kaneko, J.; Watanabe, M.; Takagi, K.; Okada, T. Development of a Questionnaire for Predicting Adolescent Women with an Unstable Autonomic Function. Jpn. J. Psychosom. Med. 2017, 57, 452–460. [Google Scholar]
  90. Viola, P.; Jones, M.J. Robust Real-Time Face Detection. Int. J. Comput. Vis. 2004, 57, 137–154. [Google Scholar] [CrossRef]
  91. Lee, T. Image representation using 2D Gabor wavelets. IEEE Trans. Pattern Anal. Mach. Intell. 1996, 18, 959–971. [Google Scholar]
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Article Metrics

Citations

Article Access Statistics

Multiple requests from the same IP address are counted as one view.