Next Article in Journal
Bidirectional Drive with Inhibited Hysteresis for Piezoelectric Actuators
Next Article in Special Issue
A Deep Learning-Based Chinese Semantic Parser for the Almond Virtual Assistant
Previous Article in Journal
Beep4Me: Automatic Ticket Validation to Support Fare Clearing and Service Planning
Previous Article in Special Issue
Cross-Modal Object Detection Based on a Knowledge Update
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Graph Representation Learning-Based Early Depression Detection Framework in Smart Home Environments

Department of Industrial Engineering, Sungkyunkwan University, Suwon 16419, Korea
*
Author to whom correspondence should be addressed.
Sensors 2022, 22(4), 1545; https://doi.org/10.3390/s22041545
Submission received: 31 December 2021 / Revised: 14 February 2022 / Accepted: 15 February 2022 / Published: 17 February 2022
(This article belongs to the Special Issue Internet of Things for Smart Homes Ⅲ)

Abstract

:
Although the diagnosis and treatment of depression is a medical field, ICTs and AI technologies are used widely to detect depression earlier in the elderly. These technologies are used to identify behavioral changes in the physical world or sentiment changes in cyberspace, known as symptoms of depression. However, although sentiment and physical changes, which are signs of depression in the elderly, are usually revealed simultaneously, there is no research on them at the same time. To solve the problem, this paper proposes knowledge graph-based cyber–physical view (CPV)-based activity pattern recognition for the early detection of depression, also known as KARE. In the KARE framework, the knowledge graph (KG) plays key roles in providing cross-domain knowledge as well as resolving issues of grammatical and semantic heterogeneity required in order to integrate cyberspace and the physical world. In addition, it can flexibly express the patterns of different activities for each elderly. To achieve this, the KARE framework implements a set of new machine learning techniques. The first is 1D-CNN for attribute representation in relation to learning to connect the attributes of physical and cyber worlds and the KG. The second is the entity alignment with embedding vectors extracted by the CNN and GNN. The third is a graph extraction method to construct the CPV from KG with the graph representation learning and wrapper-based feature selection in the unsupervised manner. The last one is a method of activity-pattern graph representation based on a Gaussian Mixture Model and KL divergence for training the GAT model to detect depression early. To demonstrate the superiority of the KARE framework, we performed the experiments using real-world datasets with five state-of-the-art models in knowledge graph entity alignment.

1. Introduction

1.1. Background

Depression is the most common psychiatric disorder in adults, accompanied by symptoms such as a depressed mood, lack of motivation, changes in appetite, trouble sleeping, decreased energy, fatigue, fatigue, and anxiety [1,2]. According to the World Health Organization (WHO), in 2021, 3.8% of the world’s population suffered from depression, and the rate in the elderly over the age of 60 was 5.7%, 1.5 times the overall level [3]. It should pay particular attention to depression of the elderly because it is difficult for the elderly to recognize that they are depressed on their own, so they often complain of physical symptoms rather than complaining of changes in their emotions [4,5].
Although diagnosing and treating depression is a medical field, rapidly developing information and communication technologies (ICTs) and artificial intelligence (AI) technologies contribute to timely diagnosis and treatment through its early detection [6,7,8,9,10,11,12]. Early Detection of Depression (EDD) research using ICTs and AI technology is divided into two aspects. The first one detects depression early by capturing behavioral changes while continuously monitoring the behavior of the elderly in a smart home environment in which various sensors and their network are installed. Researchers in this field detect depression by recognizing changes in body shape such as a patient’s gait [13,14], head position [15], and thoracic kyphosis [16]. The method of detecting depression using the sensors is again divided into obtrusive and unobtrusive; the former has the problem of inconvenience and the latter suffers from low accuracy [5,17]. The other one detects depression early by analyzing behavior in cyberspace using various AI techniques [18]. However, unlike the young, the elderly in cyberspace are mostly passive users who enjoy surfing, watching, and enjoying, rather than active users who actively express their opinions and participate [19]. For this reason, it may be difficult to detect depression in the elderly at an early stage with a language model based on text mining and deep learning used in existing sentiment analysis. In addition, as mentioned previously, although sentiment changes and physical changes, which are signs of depression in the elderly, are highly correlated [16,20], there is very little research on them at the same time.

1.2. Motivation

This paper proposes a method for the earlier detection of depression in the elderly by considering both sentiment changes in cyberspace and physical changes. For this, the following assumptions are made: the elderly subject lives in a smart home that can monitor their daily activities, and his/her Internet activity logs and service usage records are all collectible. The first assumption is for sensor data related to physical changes in the elderly. The second assumption is for collecting sentiment change-related data required for early detection of depression. Even if sufficient data have been collected through this process, the following problems need to be addressed for the detection of depression in the elderly while considering both sentiment and physical changes. The first is data heterogeneity. Generally, sensor data are mostly signal data in numerical form, whereas log or service usage history data are a mixture of categorical and numerical data. Therefore, it is imperative to address the syntactic heterogeneity of data that are difficult to incorporate in matrix or tabular form due to the differences in the structures of data elements, as well as differences in terms of format and temporal characteristics. The second is the integration of changes in cyberspace and physical space related to the symptoms of depression in the elderly. To achieve this, it is necessary to use a modeling method for the cross-domain knowledge that connects them while maintaining the unique characteristics of the spaces. At this time, the cross-domain knowledge acts as a share point that can detect physical and sentiment-based changes. The third is the accurate capturing of minute changes in the physical activity and sentiment of elderly subjects. This requires a new modeling method to represent activity patterns and their minute changes.

1.3. Main Idea

To address the above issues, this paper proposes knowledge graph-based cyber–physical view-based activity pattern recognition for the early detection of depression, also known as KARE. At this time, a knowledge graph (KG) is defined as a multi-relational graph composed of entities as nodes and relations with labels to represent multidisciplinary domain knowledge. In the KARE framework, the KG plays key roles in providing cross-domain knowledge as well as resolving issues of grammatical and semantic heterogeneity in order to integrate the cyber and the physical world. In addition, it can flexibly express the patterns of different activities for each elderly subject. The KARE framework consists of two modules: a knowledge graph-based cyber–physical view representation (CPVR) module and a personalized activity pattern recognition (PAPR) module for detecting depression, the latter of which is based on anomaly detection.
CPVR module: The CPVR module is used to integrate views represented by cyberspace and the physical world as a graph structure using the KG. At this time, a view is defined as a set of essential information needed to describe an entity in cyberspace and the physical world, and can include concepts, terms, schema, features, and/or relationships.
PAPR module: This is a module that represents the activity patterns of the elderly in a graph and performs prediction model-based learning focused on the recognition of behavioral patterns using a Graph Neural Network (GNN) model. It undertakes the learning of normal activity patterns and detects depression earlier based on the learned model using the graphs of the activity patterns and the Graph Attention Network (GAT) model.
The rest of this paper is organized as follows. Section 2 summarizes the related works. Section 3 presents an illustrative example highlighting that the detection of depression should be carried out through the integration of cyberspace and the physical world. Section 4 describes the overall architecture of the KARE framework and its components. Section 5 demonstrates the performance of the proposed framework through several experiments. Finally, Section 6 presents the conclusions and future research.

2. Related Works

2.1. Knowledge Graph Alignment

Even though a Knowledge Graph (KG) contains a huge amount of information from various domains, it has several limitations that impede its widespread use [21,22]. First, the complexity and high dimensionality of the KG, which hinder the rapid information retrieval, query answering, and data integration, are caused by the knowledge representation of various domains with different schemas, vocabularies, and graph structures [23,24]. Second, the incompleteness and inconsistency of the KG is a very primitive phenomenon that can occur as the KG’s scale grows and the domain coverage of the KG widens. It may prevent the performance of reasoning using acquired knowledge as well as the discovery of new knowledge in the KG [25]. Lastly, the large-scale issue of the KG occurs due to the huge volume of the RDF data as well as the great diversity of the sources over multiple domains [26].
To resolve the limitations of the KG, knowledge graph alignment that interprets and predicts the relationship between two heterogeneous nodes is in the spotlight as a core technology. However, since the KG is represented as a triple form, it is difficult to align the KG with its original properties. Thus, knowledge graph embedding was usually adopted for the KG alignment methods. The KG embedding techniques can be divided into two categories [27,28]. One is translational distance models, and the other one is semantic matching models. The former use distance-based score functions. TransE [29], TransH [30], and TransD [27] are typical translational distance models. Since these models recognize graph patterns in KGs unidirectionally, the graph patterns that can be captured are limited. The latter model uses similarity-based scoring functions. RESCAL [31] and Semantic Matching Energy (SME) [32] are typical semantic matching models. Additionally, there is research focused on the embedding of RDF graphs into a continuous vector named RDF2Vec [33]. This method uses neural language models to embed RDF graphs into the vector. Before training the neural language model, the RDF graph should be transformed into sequences; graph walks and Weisfeiler–Lehman subtree RDF graph kernels [34] are mostly used for this translation.
Recently, with the development of the Graph Neural Network (GNN) model, KG alignment methods based on GNN representation learning have been proposed to recognize patterns in complex graph structures [35]. GNN representation learning is a method of representing KG nodes or graphs as low-dimension vectors that can effectively discriminate components using the predictive performance of the GNN model. At this time, the types of the GNN model utilized are the Graph Convolutional Network (GCN), GraphSAGE, and Graph Attention Network (GAT). The GCN is a model that is inspired by the Convolutional Neural Network (CNN); it receives a subset of the neighboring nodes of a node as an input and discovers low and dense dimensions that can differentiate nodes, and it is usually used in cross-lingual KG alignment [36,37,38,39]. GraphSAGE minimizes information loss by concatenating vectors of neighbors rather than summing them into a single value in the process of neighbor aggregation [40,41]. GAT utilizes the concept of attention to individually deal with the importance of neighbor nodes or relations [21,42,43,44,45,46,47]. Since each model has different characteristics and advantages, suitable models for KG alignment differ depending on the components and the topological structure of the KG.

2.2. Graph-Based Anomaly Detection

Since the methods of anomaly detection based on the distribution or pattern of existing tabular data have fundamental limitations in terms of capturing the complex relational information of objects, graph-based anomaly detection (GAD) research has recently received great attention in relation to overcoming the limitations of anomaly detection in tabular data [48,49]. These GAD studies can be classified according to whether they focus on the complexities or dynamics of graphs. The GAD studies on the complexity of graphs attempt to obtain embedding vectors with low dimensions from complex-dimensional graphs (in particular, attribute networks) and use embedding vectors to perform anomaly detection. The GCN, which is relatively simple and has low computational complexity, is often used as the main model because it has to deal with graphs of complex dimensions [50,51]. However, since the GCN has structural limitations in the interpretation of complex and dynamic relationships in graphs, some anomaly detection studies that hybridize GCN and GAT using the concept of attention have been proposed [52,53]. Since the GAD for the attributes network assumes one complex graph in which graphs of various domains are mixed, they focus on finding a dimension that can interpret the complex dimension of the graph by hybridizing various models instead of using one GCN model [54]. In other words, it aims to train a comprehensive and generalized model that can detect various types of anomalies from various data sources and graphs.
The GAD studies in dynamic graphs focus on dealing with anomalies of time dependence. In particular, some anomalies that can only be found in long-term graphs are receiving great attention because they show a very latent pattern to be distinguished from normal patterns, but also impose a fatal risk to the system. The core process of these studies can be divided into two components. The first one involves the creation of a temporal graph that can represent the time dependency in the graph structure, and the second one relates to the design of a model that can capture temporal patterns. NetWalk proposed clique embedding and reservoir sampling to quickly capture dynamically changing graph dimensions [55]. AnomRank classifies the types of anomalies into structural and node-relational anomalies and proposes a method to detect anomalies in two ways [56]. In addition, GAD approaches that incorporate graph clustering or community detection methods have been proposed for dynamic graphs [57,58]. However, since these methods concentrate on the time features related to the anomaly, the features that distinguish the nodes are somewhat overlooked. Therefore, there are cases where the detection of anomalies largely depends on the topological structure of the graph, and it is difficult to capture the change of the node itself.

3. Illustrative Scenario

A smart home is a space where devices for the physical world, such as sensors and devices, are deployed and operated, and a space where services related to the cyber world such as social media services and over-the-top (OTT) services are executed. Mrs. Jane, a resident of a smart home in her 70 s, is an active senior who shares her opinions, emotions, and status with acquaintances through her SNS activities. She recently went to the hospital for insomnia and loss of appetite, but her symptoms did not improve because she did not recognize by herself that there were symptoms of depression. Generally, the elderly have difficulty recognizing signs that they are depressed, so they often complain of physical symptoms such as insomnia and loss of appetite rather than complaining of emotional changes [5]. As a result, it is hard to detect depression in the elderly in the early stages, so treatment time is often missed, leading to serious social problems such as suicide in the elderly. To solve the problem, it is necessary to recognize changes in the elderly in cyberspace as well as in the physical world. Suppose a smart home system can detect keywords such as “death” and “loneliness” from a person’s Facebook while also finding that the lights in their house are darker and more filthy than usual. Then, it could contribute to the early detection of depression. However, regarding the detection of depression in the elderly at the early stages, although the objects of the physical space are tightly connected to services within cyberspace, they are unfortunately implemented and operated separately because one type is sensor-oriented whereas the other is service-oriented (Figure 1 left pane).
A challenging issue concerns the connecting of depression-related objects between physical space and cyberspace while maintaining interoperability between them. This can be tackled using methods such as data schema, ontologies, and data exchange models [59]. However, these methods have the following limitations: inability to flexibly manage the schema of newly added or changing objects and difficulty in the interpretation of relationships between the objects that have different schema. We adopted a knowledge graph that contains various cross-domain knowledge to overcome the limitations. Since the entire KG is huge, we propose an intermediary knowledge graph called the cyber–physical view (CPV) that extracts and represents the core elements (called view) of the KG required for the early detection of depression (Figure 1 right pane). In addition, we propose a method to develop CPV.

4. Overall Architecture

As depicted in Figure 2, the KARE framework consists of two modules, namely a cyber–physical view representation (CPVR) module and personalized activity pattern Recognition (PAPR) module. The CPVR module integrates the views represented by cyberspace and the physical world as a graph structure using the KG. The PAPR module delivers the activity patterns to detect depression earlier. Lastly, the KARE framework detects depression earlier using the trained Graph Attention Network (GAT) model.

4.1. Cyber–Physical View Representation Module

The CPVR module develops intermediary KG that can integrate the elements of cyberspace and the physical world that are related to the detection of depression into a graph structure. The intermediary KG consists of elements of the KG that can connect cyberspace and the physical world. To identify the elements of the KG, the CPVR module performs entity alignment with the KG and the physical world and cyberspace, respectively. At this time, the entity alignment is performed between attributes of the data schema and vertices of the KG. Prior to discussing these issues in detail, the attribute-related data used in the entire KARE framework are defined as follows.
Definition 1.
Attribute-associated data ( A i ) comprise the set of all data associated with the i t h attribute, such as its name, its relationships to other entities, and its own instances, in the data schema. They are simply represented as follows.
A i = { e i , n a i ,   I i } ,     2 i N  
where e i is an entity name having A i as an attribute, n a i is a name of A i , and I i is an instance vector for the i t h attribute.
Definition 2.
Triplets of Vertex ( V j ) is a set of triplets of the j t h vertex in the KG. It is simply represented as follows.
V j = { v j , P j , O j } ,     2 j  
where v j is a URI of the j t h vertex, P j is a predicate vector corresponding to O j , O j is an object vector ( | P j | = | O j | ), P j has predicate labels as elements, and O j has vertices or Literals as elements ( O j { v j , L i t e r a l s   |   j , j j ,   j j } ).

4.1.1. Attribute Representation Learning with 1D-CNN Model

Since the vertices of the KG are graph-natured data and the attributes of the data schema are tabular data, it is difficult to find the equivalent relationship between the attributes and the vertices with the conventional entity alignment methods using the topological structure of the graph. To overcome this difficulty, we propose a novel entity alignment method to discover the shared space of the attributes and the vertices using the CNN model. The critical issue of the novel entity alignment method relates to the means of generating the attributes and the vertices of the fixed dimension as inputs to the CNN model. Since the vertices can be extracted as vertices of embedding vectors by applying GCN-based node representation learning [60], we focused only on generating the attributes of embedding vectors. As with word representation, one-hot encoding must be performed for each attribute to create an embedding vector for the attributes. However, as mentioned in Definition 1, the attributes do not exist independently but are related to several elements, such as other entities, their instances, etc. Thus, the process is required to generate one-hot encoding vectors for the attributes. In this study, we devised attribute representation learning with 1D-CNN by borrowing the approach of word representation learning. In other words, just as word representation learning extracts embedding vectors by capturing the relationship between the center word and neighbor words, the proposed attribute representation learning is a method of extracting embedding vectors by identifying the relationship between the center attribute and its related neighbor attributes. In the proposed method, the most important thing is to resolve the structural differences between the words and the attributes. Hereinafter, the solution will be described in detail.
Step 1: Generation of the attribute sequence
Generally, 1D-CNN sequentially receives a set of center words and neighbor words as input data according to a specific word sequence. However, the attribute-associated data are just a chunk of information, and there is no sequence between them. To generate the sequences of the neighbor attribute, two-phased sub-sampling is performed. The first-phase sub-sampling is performed to randomly select a set of the instance vectors of neighbor attributes of a specific attribute called the center attribute. The sampling procedure is as follows. Let one of A i be the central attribute A c . At this time, A c has l neighbor attributes A c r   that have I c r as an instance vector ( r = 1 , 2 , ,   l ) . In addition, A c r is a randomly selected k 1 size subset of A c r   ( r = 1 , 2 , ,   k 1 ,   k 1 < l ) used to find the relationship between the central attribute and the neighbor attributes. For all r , I c r is randomly sorted, and an index p indicating the order of I c r is added according to the sorted order ( p = 1 , , k 1 )   . Finally, the sequence vector of A c r is created as follows.
n A c = { [ , I c ( r , p ) , ]   | I c ( r , p ) A c r ,   r , r < l }
where k 1 is the sampling size of first-phase sub-sampling.
The second-phase sub-sampling is performed for the selection of k 2 elements from I c ( r , p ) (   p ). Let X c ( r , p ) be a part of I c ( r , p )   with size k 2 . Thus, the result of the second-phase sub-sampling—in other words, a sequence of the attribute A c —is represented as follows.
A S c = { , X c ( r , p ) ,   | r , r k 1 }
where k 2 is the sampling size ( | X c ( r , p ) | = k 2 ).
For each i , multiple random sampling is performed as the A i central attribute and the results are used to generate a series of sequences A S i s   ( s 2 ) .
Step 2: Vector transformation
Finally, to represent the attribute’s information as an initial input vector, one-hot-encoding is performed using the instance vector containing the most attribute information. As a result, k 2 × N can be obtained as the input matrix even though only 1 × N initial input vector is needed. Since the matrix cannot be used as input data for the 1D-CNN, vector transformation is performed as follows. First, 1 × N empty vector of k 2 is created, similar to the approach used in one-hot encoding. At this time, N   is the number of attributes. Using A S i s and its instance vector X i ( r , p ) s , the related neighbor r is identified and X i ( r , p ) s is assigned into the r t h column of the empty vector as a value. All other columns of the empty vector except r are assigned dummy vectors as inputs. This process is repeated in sequence according to the p value. Finally, a matrix of encoded attribute sequence e A S i s ( k 1 × N ) is obtained.
However, 1D-CNN uses only a single value as an input, not a matrix. Therefore, the matrix is converted to a single value through flattening and concatenation. As a result, matrix e A S i s is converted into f A S i s ( k 1 × ( N × k 2 )), which has a single value as an entry. The f A S i s is an input sequence for the 1D-CNN training and the filter used for the training process is as follows.
f i l t e r f l h = f ( w · f A S i s [ f l : f l + h ] + b )
where f l is the filter index, f is the activation function (relu), ,   is the dot product, f A S i s [ f l : f l + h ] is a sub-matrix of f A S i s , h is the filter size, and b is the bias.
In addition, the loss function based on cross-entropy is as follows.
l o s s = i ι y ι s   log ( p ι s )
where y ι s is the ι t h true label of A i ( y ι s I i ), and p ι s is the ι t h prediction value.
The overall procedure for attribute representation learning with the 1D-CNN model is depicted in Figure 3. Finally, the attributes of the embedding vectors ( E e b i ) are obtained for all attributes.

4.1.2. Entity Alignment with Embedding Vectors and CNN

In the next step, using the attributes and vertices of embedding vectors, entity alignment between the attributes and the vertices is performed. To achieve this, descriptive information such as the name or title of the entities expressed in the ‘rdf:label’ is absolutely necessary. However, the vertices of the embedding vectors contain descriptive information, whereas the attributes of the embedding vector contain only information on the relationship between the attributes based on the distribution of instances. Thus, to determine whether two vectors are equivalent or not, the attributes of the embedding vectors ( E e b i ) are first concatenated based on the name of the entity ( e i )   and the attributes ( n a i ) . Next, the embedding vectors of the pair of the attribute and the vertex that has an equivalence relationship are concatenated and annotated with the label ‘1.’ This process is repeated for all attributes and vertex pairs that have a relationship of equivalence. Furthermore, the pairs of the attributes and the vertices that are not equivalent are randomly selected. After concatenating their embedding vectors, they are annotated with ‘0.’ Finally, CNN is performed using the concatenated and annotated vectors as inputs. The overall procedure of the entity alignment is summarized in Figure 4.
The trained CNN model is used to predict the equivalence of all pairs of the attribute and the vertex. When a pair of attributes and vertices that have an equivalence relationship is found, a triplet with the predicate ‘owl:sameAs’ is created and added to KG. In this case, it is assumed that all attributes can be identified by URI. Integrated knowledge graphs on which entity alignment has been performed are represented as follows.
Definition 3.
An integrated knowledge graph ( I K G ) is a set of triplets (subject, predicate, and object) including inferred equivalence relations.
I K G = { ( s q , p q , o q ) } ,     2 q
where s q is a URI of the attribute or the vertex on the subject of the triplet, p q is a predicate label on the predicate of the triplet ( o w l : s a m e A S { p p | p } ), and o q is a URI of the attribute or vertex or L i t e r a l s on the object of the triplet.

4.1.3. View Selection Based on Feature Selection Method

IKG is a vast graph containing various types of information such as concepts, terms, observations, and relations such as the equivalence of the attributes and the vertices. However, as already known, the GNN model restricts the structure and size of the input graph, i.e., the number of hops or nodes, due to inherent limitations. Thus, the large number of concepts and relationships that IKG has can be non-informative and noisy in terms of recognizing the graph patterns needed for the detection of depression. To remove the noise, we propose a method for extracting a sub-graph of the IKG, known as the cyber–physical view (CPV), composed of informative and representative nodes and relations that are needed to recognize the graph patterns that link cyberspace and the physical world. The CPV is represented in triplets as follows:
C P V = { ( s q , p q , o q ) } , C P V I K G , 2 q
To devise a means of extracting the CPV from the IKG, we borrowed the feature selection method that finds the most informative and smallest subset of the original feature set. Applying the existing feature selection method requires prior knowledge such as label information or specific distance metrics. However, it is not possible to assume prior knowledge of IKG’s views. Therefore, this paper proposes two methods: unsupervised pseudo-label generation using GNN representation learning and wrapper-based view selection. At this time, the first one is used to generate embedding vectors of a full-size sub-graph connected to an arbitrary node and its partial sub-graphs based on GNN representation learning. After calculating the distance between the full-size embedding vector and the embedding vectors of the partial sub-graphs, the partial sub-graphs with the smallest difference to the full-size embedding vector in terms of size are selected as a view. The result is a CPV, which is a combination of the graphs made up of only informative triplets.

4.2. Personalized Activity Pattern Recognition

The behavioral patterns of elderly subjects living in smart homes are very different depending on their individual interests, habits, and health status. In addition, since the behavioral patterns are time-dependent, even the same behavior can have different time intervals. For the early detection of depression based on recognizing minute changes in the normal activity patterns, an informative and flexible graph of time features as well as the GNN model are needed. However, since the CPV contains all kinds of informative and representative nodes and relations that link cyberspace and the physical world, it is necessary to identify those related to activity patterns. Three steps are involved in the discovery of the activity patterns: finding activities of attributes via data-centric approaches, generating activity graphs using CPV triplet augmentation associated with the activities of attributes, and arranging activity graphs sequentially. As a result of the discovery of the activity patterns, activity pattern graphs are acquired and used as inputs to learn the GAT that can capture the normal activity patterns. Finally, early detection of depression is performed with the trained GAT performs.

4.2.1. Generation of the Activity Pattern Graphs

It is difficult to find the activities directly in CPV due to the complexity of the graph structures. Thus, for this stage in the process, we decided to focus on finding the activities of attributes using data; in other words, instances. At this time, it is important to find the timestamp and the time intervals of each activity, but it is assumed that they are given because these aspects are beyond the scope of this paper. The finding of the activities of attributes via a data-centric approach is performed with the time-window drift and Gaussian Mixture Model (GMM). Prior to a detailed step-by-step description that follows, Figure 5 schematically illustrates the generation procedure of activity pattern graphs.
Step 1: Finding the activities of the attributes
As shown in Figure 5, there are instances for every attribute (e.g., A t r a f f i c and A r o t a t i o n ) along the timeline. However, since not all instances of the attributes are related to the activities, it is necessary to find only instances that are related to the activities. To achieve this, the instances are sliced to the size of the time window with a given hyperparameter. As a result, an attribute-specific dataset consisting of a series of instances with a time interval w is created. For all instances I i (for all i ) with time interval w , their start time and end time are assigned and then sorted in chronological order.
I i t i = { x i v t i , s i t i , e i t i | x i t i I i , v = 1 , 2 , } ,   1 t i
where x i v t i is the v t h instance of I i in the t i t h time interval [ s t i , e t i ] , and s t i and e t i are the start time and the end time of the t i t h time interval ( s t i < e t i ,   | s t i e t i | = w ), respectively.
Equation (10) shows the probability distribution of the I i t i value estimated using GMM and the expectation-maximization (EM) algorithm.
p ( x i t i ) = g = 1 G w g × N ( x i t i ; μ g , σ g 2   )
where w g is an EM-based estimated probability that the g t h Gaussian distribution is selected, and μ g and σ g 2 are the EM-based estimated mean and standard deviation, respectively. G is the number of Gaussian distributions.
To determine whether the instances, which are the elements of the attribute-specific dataset, contains information on the activities, p ( x i t i ) is compared with the distribution of instances in the sleep state called p ( x i s l e e p ) , which does not carry any activities. At this time, it is assumed that p ( x i s l e e p ) has already been generated using the sleep dataset and Equation (10). Finally, the distance between the two probability distributions is calculated using the Kullback–Leibler distances Equation (11).
K L i t l = p ( x i n o n ) l o g p ( x i n o n ) p ( x i t i ) d x i
If K L i t l   is greater than the threshold δ, the t i t h   activity of A i is assigned as an element of A A i t i * ( | t i * | < | t i | ). For all t i and i , the above process is repeated and A A i = { A A i t i * | t i * } and A A = { A A i | i } are generated.
Step 2: Generation of activity graphs using the CPV
A A r is randomly selected from the set A A ( r i ), and subsequently, the node of A A r sharing the index with A A r is discovered. A sub-graph composed of A r and its neighbor triplets is extracted from the CPV. By using the extracted sub-graphs and the A A i t i * , the activity graphs A G i t i *   of the i t h attribute are created as follows:
A G i t i * = { A A i t i * , { , ( s i t i * , p i t i * , o i t i * ) , } | s i t i * , p i t i * , o i t i * C P V }
At this time, although A G i t i * generated from A i is a graph of the same structure, it is possible to classify it according to A A i t i * . The above process is repeated for all t i * and i values to obtain the sets A G i = { A G i t i * | t i * } and A G = { A G i | i } .
Step 3: Arrangement of activity graphs to represent the activity pattern graphs
To represent the activity patterns in the graphs, the relationship between A G i t i * is identified and serialized. At this time, the relationship is of two types: a time series relationship and a path relationship. The former is the back-and-forth relationship between the A G i t i * elements of the set A G i . This relationship is established if the time difference between arbitrary values of A G i t i * = a and A G i t i * = b is less than the size w ( | e t l * = a s t l * = b | < w ). The latter establishes a pair of activity graphs ( A G i = a t i * = c , A G i = b t i * = c ) where A G i = a and A G i = b   share t l * but are elements of different sets. It is created by extracting triplets between A i = a and A i = b in the CPV. Finally, the activity patterns in the graphs of A P G i t i * are created: ( | A P G i t i * | | A G i t i * | ).

4.2.2. GAT-Based Activity Pattern Recognition

The Graph Convolutional Network (GCN), GraphSAGE, and Graph Attention Network (GAT) are mainly used as GNN models to capture graph patterns [61]. Each method has the following pros and cons. The GCN has the advantage of being able to preserve the node or subgraph information of very complex graph topologies, but it is difficult to capture the relationship between them. GraphSAGE can preserve information on the relationships between the nodes and their neighbors more effectively than the GCN, but its generalization performance requires a large amount of data due to the concatenation process. Finally, the GAT has an advantage in terms of its generalization performance because it effectively preserves relational information between nodes and their neighbors with the concept of attention but has a high computational complexity. As mentioned previously, the input data A P G i t i * to be used for training is a small graph related to the activities and a temporal graph focusing on the relationship between the nodes along the timeline of the activities. Since the GAT is suitable for such A P G i t i * data, the GAT and A P G i t i * are used to train a GAT model so that it can recognize normal activity patterns. By applying this model, minute changes that are different from normal activity patterns, that is, symptoms of depression, can be detected earlier.

5. Evaluation

In order to evaluate the superiority of the proposed KARE method, we conducted experiments and evaluations using the state-of-the-art comparative models with experimental graph datasets covering multidisciplinary domains. We used a system with an Intel(R) Xeon(R) CPU @ 2.30 GHz processor, a Teslar P100 graphics card, and 12 GB memory

5.1. Experimental Datasets

Experimental graph datasets collected from multiple domains such as the KG were required, and information such as the label text needed to be attached to the components of the graph. In addition, label information for anomalies was required to evaluate the anomaly detection performance. As the most suitable dataset for these requirements, we used the redditLinks dataset that graphs the relationship between the body of posts and the topics in multi-topic discussions on the Reddit platform [62].
The redditLinks dataset is the most suitable graph for the KG entity alignment and anomaly detection that we chose to experiment with, but the structure of the graph, especially the entity-to-entity connection relationship, is somewhat different. In general, the relationship between entities in the graph is established directly, but in the redditLinks dataset, the relationship between entities is composed via posts. We used the posts as a predicate to connect entities and generate triples. Additionally, on the redditLinks dataset, the title of the entity is attached, not the ID of the individual entity. After replacing all of these with IDs, we created relations of equivalence using the entity’s titles. The created equivalence relations were used as a training and test set for the KG alignment. Finally, the features or attributes of each entity on the redditLinks dataset were represented as sets of vectors of features rather than attached in a graph form. We disassembled vectors and reconstructed them in the form of a graph centered on all entities. Figure 6 shows the structure of the graph transformed from the redditLinks’s graph for our experiment. The statistics of the transformed dataset are summarized in Table 1.

5.2. Knowledge Graph Alignment

We proposed a method of KG entity alignment based on attribute representation learning. To evaluate the performance of the proposed method, five state-of-the-art models of KG entity alignment were selected. MtarnsE, TransD, RotatE, and ConvE do not utilize the relationships of attributes among the entities when embedding the entity. On the other hand, GCn-align uses the relationships of attributes to embed entities, similar to the KARE framework. MtarnsE, TransD, RotatE, and ConvE do not utilize the relationships of attributes among the entities’ relationships for the embedding of entities. On the other hand, GCN-align uses the relationship of attributes to find the dimensions of entities, similar to the KARE framework. Each model was trained by limiting it to the epoch 50, and the default parameters were adopted for the training models. In addition, since the scale of the experimental dataset was very large, it was sampled and tested. Table 2 shows the experimental results of the full dataset with the metrics of Hits@k, mean rank (MR), and mean reciprocal rank (MRR). Hits@k is the percentage of corrected alignment in the top k. MR and MRR are the averages of ranks and inverse ranks for test sets, respectively. Figure 7 shows the experimental results of the sampling datasets obtained for the five metrics.
  • TransH [30]: Translation-based model for a relation as a hyperplane.
  • TransD [27]: Translation-based model to improve TransR/CTransR with representation of a named symbol object.
  • RotatE [63]: Relational pattern-based KG embedding model including relationships of symmetry/antisymmetry, inversion, and composition.
  • ConvE [36]: A multi-layer convolutional network model used to improve the performance of extracted feature sets from KG for link prediction.
  • GCN-align [64]: A graph convolutional network (GCN)-based KG entity alignment model based on both the structural and attribute information of entities.
As a result of embedding in the full dataset, the proposed KARE model was found to be superior in all metrics. The MRR metric showed the best performance, and it can be seen that the KARE model did not perform erroneous alignment for unrelated equivalence relationships, that is, for relationships in which priority was considered. In other words, the accuracy and precision of the alignment of the KARE model showed the best performance compared to the state-of-the-art methods. KARE showed a rather low value for MR, but the other models fitted the majority equivalence relationship well; these values were found to be relatively high compared to that obtained for the KARE model.
As the sampling ratio changed, we conducted an experiment to compare the performance using five metrics and comparative models. The results of the experiment are shown in Table 3 and Figure 7.
As a result of the experiment, the proposed KARE model showed high accuracy at many of the sampling ratios. The Hit@k metric decreased as the sampling ratio increased, because the complexity of the equivalence relationship to be matched increased. On the other hand, the MR and MRR values moved in opposite directions, because in the case of MR, they increased when the same equivalence relationship was continuously met, and MRR fell because the wrong equivalence relationship must have been met. In the interval where the sampling ratio was low, the performance of most models deteriorated because the complexity of the sampling set increased faster than the information required to find the generalized embedding space. The model that showed the most similar performance to our method was the GCN-align model. Most of the KARE methods showed higher values than GCN-align, but in the last 50% of the sampling process, GCN-align showed significantly higher values. This was because the information about the hyperplane to be discovered by the GCN-align model exceeded the threshold. In addition, we found a rather low value in the first 50% of the sampling process because the information loss in the embedding process was not obtained in this 50% sampling process. However, taking into account the process of the entire dataset, as shown in Table 2, the proposed KARE method showed the best performance for the full dataset.

6. Conclusions and Further Research

In this paper, we proposed the use of the KARE framework, which enables the early detection of the depression in the elderly by integrating data from the physical world and the cyber world. To achieve this, we developed the KG and the attribute representation learning method for the alignment of the KG and attributes of the data schema. Since the integrated KG, which is obtained by alignment, contains many duplicate entities and unnecessary graph structures for the detection of depression, we utilized graph representation learning and wrapper-based feature selection to extract CPV from the graph, which was integrated using the unsupervised method. Lastly, we proposed a method to generate activity pattern graphs using GMM and KL divergence. Using these activity pattern graphs, the GAT model was trained for the detection of normal activity patterns, and the early detection of depression was performed. Since the proposed KARE framework integrates physical space and cyberspace to detect observable anomalies based on human behavior, it can be applied in various scenarios in the e-health care sector. In addition, if events are detected instead of anomalies, the KARE framework can be extended to recommender systems or decision support systems. When detecting anomalies in system components rather than humans, it can be applied to various applications within the context of intrusion detection and cyberattack defense, such as DDoS attacks.
However, this paper has several limitations. To perform alignment of the KG and the attributes of the data schema, we assumed the existence of relationships of equivalence between the KG and the data schema. In real world applications, it is hard to assume that the information on relationships of equivalence exists. Therefore, a method that detects relationships of equivalence using unsupervised methods is needed to align heterogeneous graphs. Moreover, our method depends on the numerical space of the instances to generate the activity graphs. Even though categorical data can be projected into numerical space, some non-time series data cannot be detected by embedding numerical space, which preserves the original information. The KNN or decision tree can be appropriate models to handle categorical and non-linear data and transform original data to the continuous space to find the best embedding space as a preprocessing model. Therefore, a more robust and general method is needed to recognize activity graphs.
Further research will be performed to overcome the limitations mentioned above. First, using natural language processing methods and graph representation learning, the GNN model will be developed to detect equivalence between heterogeneous graphs, not between KGs. Secondly, to generate activity graphs from attributes of the data schema, a multi-view CNN model will be developed and applied to deal with a wide variety of data. We will continue to conduct research on assistive AI technology that can detect depression in the elderly at an early stage and lead a safe life within smart homes.

Author Contributions

Conceptualization, J.K. and M.S.; Data curation, J.K.; Formal analysis, M.S.; Investigation, J.K.; Methodology, J.K.; Project administration, M.S.; Software, J.K.; Supervision, M.S.; Writing—original draft, J.K.; Writing—review & editing, M.S. All authors have read and agreed to the published version of the manuscript.

Funding

This research was funded by the National Research Foundation of Korea (NRF) grant number (NRF-2019R1A2C1004102).

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Acknowledgments

This research was supported by the Basic Science Research Program through the National Research Foundation of Korea (NRF) funded by the Ministry of Education, Science and Technology (NRF-2019R1A2C1004102).

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Su, D.; Zhang, X.; He, K.; Chen, Y. Use of machine learning approach to predict depression in the elderly in China: A longitudinal study. J. Affect. Disord. 2021, 282, 289–298. [Google Scholar] [CrossRef]
  2. Ye, J.; Yu, Y.; Wang, Q.; Li, W.; Liang, H.; Zheng, Y.; Fu, G. Multi-modal depression detection based on emotional audio and evaluation text. J. Affect. Disord. 2021, 295, 904–913. [Google Scholar] [CrossRef]
  3. World Health Organization. Depression. 2021. Available online: https://www.who.int/news-room/fact-sheets/detail/depression (accessed on 30 November 2021).
  4. Eyers, K.; Parker, G.; Brodaty, H. Managing Depression, Growing Older: A Guide for Professionals and Carers; Routledge: London, UK, 2012. [Google Scholar]
  5. Kim, J.-Y.; Liu, N.; Tan, H.-X.; Chu, C.-H. Unobtrusive monitoring to detect depression for elderly with chronic illnesses. IEEE Sens. J. 2017, 17, 5694–5704. [Google Scholar] [CrossRef]
  6. Liu, H.; Li, D.; Wang, X.; Liu, L.; Zhang, Z.; Subramanian, S. Precise head pose estimation on HPD5A database for attention recognition based on convolutional neural network in human-computer interaction. Infrared Phys. Technol. 2021, 116, 103740. [Google Scholar] [CrossRef]
  7. Liu, T.; Wang, J.; Yang, B.; Wang, X. NGDNet: Nonuniform Gaussian-label distribution learning for infrared head pose estimation and on-task behavior understanding in the classroom. Neurocomputing 2021, 436, 210–220. [Google Scholar] [CrossRef]
  8. Trotzek, M.; Koitka, S.; Friedrich, C.M. Utilizing neural networks and linguistic metadata for early detection of depression indications in text sequences. IEEE Trans. Knowl. Data Eng. 2018, 32, 588–601. [Google Scholar] [CrossRef] [Green Version]
  9. Burdisso, S.G.; Errecalde, M.; Montes-Y-Gómez, M. A text classification framework for simple and effective early depression detection over social media streams. Expert Syst. Appl. 2019, 133, 182–197. [Google Scholar] [CrossRef] [Green Version]
  10. Hickey, B.; Chalmers, T.; Newton, P.; Lin, C.-T.; Sibbritt, D.; McLachlan, C.; Clifton-Bligh, R.; Morley, J.; Lal, S. Smart Devices and Wearable Technologies to Detect and Monitor Mental Health Conditions and Stress: A Systematic Review. Sensors 2021, 21, 3461. [Google Scholar] [CrossRef]
  11. Zaitseva, E.; Levashenko, V. Construction of a reliability structure function based on uncertain data. IEEE Trans. Reliab. 2016, 65, 1710–1723. [Google Scholar] [CrossRef]
  12. Baraldi, P.; Podofillini, L.; Mkrtchyan, L.; Zio, E.; Dang, V. Comparing the treatment of uncertainty in Bayesian networks and fuzzy expert systems used for a human reliability analysis application. Reliab. Eng. Syst. Saf. 2015, 138, 176–193. [Google Scholar] [CrossRef] [Green Version]
  13. Wang, T.; Li, C.; Wu, C.; Zhao, C.; Sun, J.; Peng, H.; Hu, X.; Hu, B. A gait assessment framework for depression detection using kinect sensors. IEEE Sens. J. 2020, 21, 3260–3270. [Google Scholar] [CrossRef]
  14. Michalak, J.; Troje, N.F.; Fischer, J.; Vollmar, P.; Heidenreich, T.; Schulte, D. Embodiment of sadness and depression—gait patterns associated with dysphoric mood. Psychosom. Med. 2009, 71, 580–587. [Google Scholar] [CrossRef] [Green Version]
  15. Alghowinem, S.; Goecke, R.; Wagner, M.; Parkerx, G.; Breakspear, M. Head pose and movement analysis as an indicator of depression. In Proceedings of the 2013 Humaine Association Conference on Affective Computing and Intelligent Interaction, Geneva, Switzerland, 2–5 September 2013. [Google Scholar]
  16. Canales, J.Z.; Cordás, T.A.; Fiquer, J.T.; Cavalcante, A.F.; Moreno, R.A. Posture and body image in individuals with major depressive disorder: A controlled study. Braz. J. Psychiatry 2010, 32, 375–380. [Google Scholar] [CrossRef] [Green Version]
  17. Chalmers, T.; Hickey, B.A.; Newton, P.; Lin, C.-T.; Sibbritt, D.; McLachlan, C.S.; Clifton-Bligh, R.; Morley, J.; Lal, S. Stress Watch: The Use of Heart Rate and Heart Rate Variability to Detect Stress: A Pilot Study Using Smart Watch Wearables. Sensors 2021, 22, 151. [Google Scholar] [CrossRef]
  18. Chiong, R.; Budhi, G.S.; Dhakal, S.; Chiong, F. A textual-based featuring approach for depression detection using machine learning classifiers and social media texts. Comput. Biol. Med. 2021, 135, 104499. [Google Scholar] [CrossRef] [PubMed]
  19. Aalbers, G.; McNally, R.J.; Heeren, A.; de Wit, S.; Fried, E.I. Social media and depression symptoms: A network perspective. J. Exp. Psychol. 2019, 148, 1454. [Google Scholar] [CrossRef] [Green Version]
  20. Zhang, B.; Zaman, A.; Acharyya, R.; Hoque, E.; Silenzio, V.; Kautz, H. Detecting Individuals with Depressive Disorder fromPersonal Google Search and YouTube History Logs. arXiv 2020, arXiv:2010.15670. [Google Scholar]
  21. Zhang, B.; Zaman, A.; Acharyya, R.; Hoque, E.; Silenzio, V.; Kautz, H. Learning knowledge graph embedding with heterogeneous relation attention networks. arXiv 2010, arXiv:2010.15670. [Google Scholar]
  22. Zhang, Z.; Li, Z.; Liu, H.; Xiong, N.N. Multi-scale dynamic convolutional network for knowledge graph embedding. In IEEE Transactions on Knowledge and Data Engineering; IEEE: Piscataway, NJ, USA, 2020. [Google Scholar]
  23. Zheng, L.; Qu, Y.; Qian, X.; Cheng, G. A hierarchical co-clustering approach for entity exploration over Linked Data. Knowl. Based Syst. 2018, 141, 200–210. [Google Scholar] [CrossRef]
  24. Eddamiri, S.; Zemmouri, E.M.; Benghabrit, A. An improved RDF data Clustering Algorithm. Procedia Comput. Sci. 2019, 148, 208–217. [Google Scholar] [CrossRef]
  25. Xiao, H.; Huang, M.; Zhu, X. From one point to a manifold: Knowledge graph embedding for precise link prediction. arXiv 2015, arXiv:1512.04792. [Google Scholar]
  26. Hernández-Illera, A.; Martínez-Prieto, M.A.; Fernández, J.D. RDF-TR: Exploiting structural redundancies to boost RDF compression. Inf. Sci. 2020, 508, 234–259. [Google Scholar] [CrossRef]
  27. Ji, G.; He, S.; Xu, L.; Liu, K.; Zhao, J. Knowledge graph embedding via dynamic mapping matrix. In Proceedings of the 53rd Annual Meeting of the Association for Computational Linguistics and the 7th International Joint Conference on Natural Language Processing, Beijing, China, 26 August 2015; Volume 1. [Google Scholar]
  28. Nie, B.; Sun, S. Knowledge graph embedding via reasoning over entities, relations, and text. Future Gener. Comput. Syst. 2019, 91, 426–433. [Google Scholar] [CrossRef]
  29. Bordes, A.; Usunier, N.; Garcia-Duran, A.; Weston, J.; Yakhnenko, O. Translating embeddings for modeling multi-relational data. In Advances in Neural Information Processing Systems; Curran Associates Inc.: Red Hook, NY, USA, 2013. [Google Scholar]
  30. Wang, Z.; Zhang, J.; Feng, J.; Chen, Z. Knowledge graph embedding by translating on hyperplanes. In Proceedings of the AAAI Conference on Artificial Intelligence, Quebec City, QC, Canada, 27–31 July 2014. [Google Scholar]
  31. Nickel, M.; Tresp, V.; Kriegel, H.P. A three-way model for collective learning on multi-relational data. In Proceedings of the International Conference on Machine Learning, Bellevue, WA, USA, 28 June–2 July 2011. [Google Scholar]
  32. Bordes, A.; Glorot, X.; Weston, J.; Bengio, Y. A semantic matching energy function for learning with multi-relational data. Mach. Learn. 2014, 94, 233–259. [Google Scholar] [CrossRef] [Green Version]
  33. Ristoski, P.; Paulheim, H. Rdf2vec: Rdf graph embeddings for data mining. In Proceedings of the International Semantic Web Conference, Kobe, Japan, 17–21 October 2016. [Google Scholar]
  34. De Vries, G.K. A fast approximation of the Weisfeiler-Lehman graph kernel for RDF data. In Proceedings of the Joint European Conference on Machine Learning and Knowledge Discovery in Databases, Prague, Czech Republic, 23–27 September 2013. [Google Scholar]
  35. Li, Z.; Liu, H.; Zhang, Z.; Liu, T.; Shu, J. Recalibration convolutional networks for learning interaction knowledge graph embedding. Neurocomputing 2021, 427, 118–130. [Google Scholar] [CrossRef]
  36. Dettmers, T.; Minervini, P.; Stenetorp, P.; Riedel, S. Convolutional 2d knowledge graph embeddings. In Proceedings of the 32nd AAAI Conference on Artificial Intelligence, New Orleans, LA, USA, 2–7 February 2018. [Google Scholar]
  37. Sun, Z.; Wang, C.; Hu, W.; Chen, M.; Dai, J.; Zhang, W.; Qu, Y. Knowledge graph alignment network with gated multi-hop neighborhood aggregation. In Proceedings of the AAAI Conference on Artificial Intelligence, New York, NY, USA, 7–12 February 2020. [Google Scholar]
  38. Wang, Z.; Lv, Q.; Lan, X.; Zhang, Y. Cross-lingual knowledge graph alignment via graph convolutional networks. In Proceedings of the 2018 Conference on Empirical Methods in Natural Language Processing, Brussels, Belgium, 31 October–4 November 2018. [Google Scholar]
  39. Yan, Y.; Liu, L.; Ban, Y.; Jing, B.; Tong, H. Dynamic Knowledge Graph Alignment. In Proceedings of the AAAI Conference on Artificial Intelligence, Vancouver, BC, Canada, 2–9 February 2021. [Google Scholar]
  40. Hamilton, W.; Ying, Z.; Leskovec, J. Inductive representation learning on large graphs. In Proceedings of the 31st International Conference on Neural Information Processing Systems, Long Beach, CA, USA, 4–9 December 2017. [Google Scholar]
  41. Zhu, Q.; Wei, H.; Sisman, B.; Zheng, D.; Faloutsos, C.; Dong, X.L.; Han, J. Collective multi-type entity alignment between knowledge graphs. In Proceedings of the Web Conference 2020, Taipei, Taiwan, 20–24 April 2020. [Google Scholar]
  42. Wang, R.; Li, B.; Hu, S.; Du, W.; Zhang, M. Knowledge graph embedding via graph attenuated attention networks. IEEE Access 2019, 8, 5212–5224. [Google Scholar] [CrossRef]
  43. Liu, X.; Tan, H.; Chen, Q.; Lin, G. RAGAT: Relation Aware Graph Attention Network for Knowledge Graph Completion. IEEE Access 2021, 9, 20840–20849. [Google Scholar] [CrossRef]
  44. Li, C.; Peng, X.; Niu, Y.; Zhang, S.; Peng, H.; Zhou, C.; Li, J. Learning graph attention-aware knowledge graph embedding. Neurocomputing 2021, 461, 516–529. [Google Scholar] [CrossRef]
  45. Jiang, D.; Wang, R.; Yang, J.; Xue, L. Kernel multi-attention neural network for knowledge graph embedding. Knowl. Based Syst. 2021, 227, 107188. [Google Scholar] [CrossRef]
  46. Han, X.; Liu, Z.; Sun, M. Neural knowledge acquisition via mutual attention between knowledge graph and text. In Proceedings of the Thirty-Second AAAI Conference on Artificial Intelligence, New Orleans, LA, USA, 2–7 February 2018. [Google Scholar]
  47. Wang, P.; Han, J.; Li, C.; Pan, R. Logic attention based neighborhood aggregation for inductive knowledge graph embedding. In Proceedings of the AAAI Conference on Artificial Intelligence, Honolulu, HI, USA, 27 January–1 February 2019. [Google Scholar]
  48. Ma, X.; Wu, J.; Xue, S.; Yang, J.; Zhou, C.; Sheng, Q.Z.; Xiong, H.; Akoglu, L. A comprehensive survey on graph anomaly detection with deep learning. In IEEE Transactions on Knowledge and Data Engineering; IEEE: Piscataway, NJ, USA, 2021. [Google Scholar]
  49. Pourhabibi, T.; Ong, K.-L.; Kam, B.H.; Boo, Y.L. Fraud detection: A systematic literature review of graph-based anomaly detection approaches. Decis. Support Syst. 2020, 133, 113303. [Google Scholar] [CrossRef]
  50. Kumagai, A.; Iwata, T.; Fujiwara, Y. Semi-supervised anomaly detection on attributed graphs. In Proceedings of the 2021 International Joint Conference on Neural Networks (IJCNN), Shenzhen, China, 18–22 July 2021. [Google Scholar]
  51. Zhu, D.; Ma, Y.; Liu, Y. DeepAD: A joint embedding approach for anomaly detection on attributed networks. In Proceedings of the International Conference on Computational Science, Faro, Portugal, 12–14 June 2020. [Google Scholar]
  52. Zheng, L.; Li, Z.; Li, J.; Li, Z.; Gao, J. AddGraph: Anomaly Detection in Dynamic Graph Using Attention-based Temporal GCN. In Proceedings of the Twenty-Eighth International Joint Conference on Artificial Intelligence, Macao, China, 10–16 August 2019. [Google Scholar]
  53. Pei, Y.; Huang, T.; van Ipenburg, W.; Pechenizkiy, M. ResGCN: Attention-based deep residual modeling for anomaly detection on attributed networks. arXiv 2020, arXiv:2009.14738. [Google Scholar]
  54. Peng, Z.; Luo, M.; Li, J.; Xue, L.; Zheng, Q. A Deep Multi-View Framework for Anomaly Detection on Attributed Networks. In IEEE Transactions on Knowledge and Data Engineering; IEEE: Piscataway, NJ, USA, 2020. [Google Scholar]
  55. Yu, W.; Cheng, W.; Aggarwal, C.C.; Zhang, K.; Chen, H.; Wang, W. Netwalk: A flexible deep embedding approach for anomaly detection in dynamic networks. In Proceedings of the 24th ACM SIGKDD International Conference on Knowledge Discovery & Data Mining, London, UK, 19–23 August 2018. [Google Scholar]
  56. Yoon, M.; Hooi, B.; Shin, K.; Faloutsos, C. Fast and accurate anomaly detection in dynamic graphs with a two-pronged approach. In Proceedings of the 25th ACM SIGKDD International Conference on Knowledge Discovery & Data Mining, Anchorage, AK, USA, 4–8 August 2019. [Google Scholar]
  57. Kiouche, A.E.; Lagraa, S.; Amrouche, K.; Seba, H. A simple graph embedding for anomaly detection in a stream of heterogeneous labeled graphs. Pattern Recognit. 2021, 112, 107746. [Google Scholar] [CrossRef]
  58. Meng, X.; Wang, S.; Liang, Z.; Yao, D.; Zhou, J.; Zhang, Y. Semi-supervised anomaly detection in dynamic communication networks. Inf. Sci. 2021, 571, 527–542. [Google Scholar] [CrossRef]
  59. Muppavarapu, V.; Ramesh, G.; Gyrard, A.; Noura, M. Knowledge extraction using semantic similarity of concepts from Web of Things knowledge bases. Data Knowl. Eng. 2021, 135, 101923. [Google Scholar] [CrossRef]
  60. Hu, F.; Zhu, Y.; Wu, S.; Huang, W.; Wang, L.; Tan, T. Graphair: Graph representation learning with neighborhood aggregation and interaction. Pattern Recognit. 2021, 112, 107745. [Google Scholar] [CrossRef]
  61. Zhou, J.; Cui, G.; Hu, S.; Zhang, Z.; Yang, C.; Liu, Z.; Wang, L.; Li, C.; Sun, M. Graph neural networks: A review of methods and applications. AI Open 2020, 1, 57–81. [Google Scholar] [CrossRef]
  62. Kumar, S.; Hamilton, W.L.; Leskovec, J.; Jurafsky, D. Community interaction and conflict on the web. In Proceedings of the 2018 World Wide Web Conference, Lyon, France, 23–27 April 2018. [Google Scholar]
  63. Sun, Z.; Deng, Z.H.; Nie, J.Y.; Tang, J. Rotate: Knowledge graph embedding by relational rotation in complex space. arXiv 2019, arXiv:1902.10197. [Google Scholar]
  64. Sun, Z.; Hu, W.; Zhang, Q.; Qu, Y. Bootstrapping Entity Alignment with Knowledge Graph Embedding. In Proceedings of the International Joint Conference on Artificial Intelligence, Stockholm, Sweden, 13–19 July 2018. [Google Scholar]
Figure 1. Illustrative example of a knowledge graph-based cyber–physical view (CPV).
Figure 1. Illustrative example of a knowledge graph-based cyber–physical view (CPV).
Sensors 22 01545 g001
Figure 2. Overall structure of the KARE framework.
Figure 2. Overall structure of the KARE framework.
Sensors 22 01545 g002
Figure 3. The proposed attribute representation learning with the 1D-CNN.
Figure 3. The proposed attribute representation learning with the 1D-CNN.
Sensors 22 01545 g003
Figure 4. The entity alignment between attributes and vertices with the CNN model.
Figure 4. The entity alignment between attributes and vertices with the CNN model.
Sensors 22 01545 g004
Figure 5. The generation procedure of activity patterns graphs.
Figure 5. The generation procedure of activity patterns graphs.
Sensors 22 01545 g005
Figure 6. Graph transformation process to convert redditLinks graph to KG-like structure.
Figure 6. Graph transformation process to convert redditLinks graph to KG-like structure.
Sensors 22 01545 g006
Figure 7. Comparison of knowledge graph alignment methods according to sampling ratio.
Figure 7. Comparison of knowledge graph alignment methods according to sampling ratio.
Sensors 22 01545 g007
Table 1. The summary of the transformed dataset from redditLinks.
Table 1. The summary of the transformed dataset from redditLinks.
#Entities#Entity Titles#Entity Relations#Predicates#Attributes#Attributes Relations#Anomalies (%)
573,12235,776286,561259,092863,438,73221,070 (7.93%)
Table 2. Results of knowledge graph alignment with the full dataset with 5 metrics.
Table 2. Results of knowledge graph alignment with the full dataset with 5 metrics.
MethodsHits@1Hits@5Hits@10MRMRR
TransH00.110.2012744.1060.001517
TransD0.0180.110.2012759.8990.001858
RotatE00.0730.1462724.4650.001502
ConvE0.0180.0730.1462759.120.001571
GCN-align0.0730.2380.3842763.4670.002591
KARE0.0910.3280.4922790.590.003123
Table 3. Results of knowledge graph alignment with the sampling datasets.
Table 3. Results of knowledge graph alignment with the sampling datasets.
MethodsHits@1
5101520253035404550
TransH0.09100.052000.0360.097000
TransD0.0910.0650.1030.0440.077000.030.0540.065
RotatE00.19400.0440.0770.0360.032000.065
ConvE00.0650.20700.0770.0360.0320.05900.129
GCN-align8.7193.3551.8080.7090.3480.3580.2260.3550.1363.355
KARE10.4043.951.8240.8540.4360.4840.1290.3550.1770.158
MethodsHits@5
TransH0.3630.4520.0520.1770.1160.1430.1940.030.0270.581
TransD0.1820.5160.2070.0890.1540.0720.0650.0890.0810.516
RotatE0.1820.5160.1030.2220.2320.2150.1940.0890.1090.452
ConvE0.6360.1290.3620.3550.3090.1430.2260.1780.1090.194
GCN-align12.6255.9353.5121.1971.0431.0020.8060.80.2986.129
KARE13.9016.273.1051.5371.1881.0030.5811.0660.6480.606
MethodsHits@10
TransH1.1810.9680.1030.3990.2320.3220.290.2370.190.645
TransD0.8170.8390.4650.2220.3090.3940.1290.2370.2710.645
RotatE0.6360.8390.2580.5760.270.3220.4190.2370.2170.968
ConvE0.9990.1291.0850.6650.6570.2150.3870.3260.2710.839
GCN-align14.0787.0974.1841.7731.5841.3241.0651.0660.4617.29
KARE15.7857.3983.6472.3061.3471.6260.711.3920.8040.843
MethodsMR
TransH563.444768.733981.9091121.0721296.7271384.3051548.7241674.0151860.349778.442
TransD551.669774.206953.9771130.2021265.1881362.2311560.1291686.6971837.794774.777
RotatE548.125789.419964.5991163.4511270.3661407.1591559.1621653.5891836.923773.315
ConvE547.085759.823978.211140.2271271.1511403.5231544.4621656.7731865.935765.801
GCN-align458.357697.176926.0831094.4911272.6671399.5271564.6271683.8271858.828685.275
KARE453.374718.921960.9471133.8461243.7911434.6391537.6971688.8021716.4021912.765
MethodsMRR
TransH0.0067110.0053990.0032580.0033170.3578370.0030110.0031050.0022250.0018160.005183
TransD0.0060990.0061530.0046540.0032950.0036430.0026980.0021580.0022550.0025840.005378
RotatE0.0051740.0064870.0033080.0036370.0034680.0031380.0029550.0022270.0020780.005829
ConvE0.0065390.004470.0063950.0039540.0039250.0027850.0032820.0030820.0022270.005568
GCN-align0.1087020.0496450.0286280.0128190.0091160.0082190.0064390.0073010.0036220.049873
KARE0.1257740.0536970.0271730.0145840.0093980.0094340.0052260.0084080.0052830.004791
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Share and Cite

MDPI and ACS Style

Kim, J.; Sohn, M. Graph Representation Learning-Based Early Depression Detection Framework in Smart Home Environments. Sensors 2022, 22, 1545. https://doi.org/10.3390/s22041545

AMA Style

Kim J, Sohn M. Graph Representation Learning-Based Early Depression Detection Framework in Smart Home Environments. Sensors. 2022; 22(4):1545. https://doi.org/10.3390/s22041545

Chicago/Turabian Style

Kim, Jongmo, and Mye Sohn. 2022. "Graph Representation Learning-Based Early Depression Detection Framework in Smart Home Environments" Sensors 22, no. 4: 1545. https://doi.org/10.3390/s22041545

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop