Next Article in Journal / Special Issue
Exploring the Spatial Association Between Spatial Categorical Data Using a Fuzzy Geographically Weighted Colocation Quotient Method
Previous Article in Journal
Assessing Spatial Digital Twins for Oil and Gas Projects: An Informed Argument Approach Using ISO/IEC 25010 Model
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

A Spatio-Temporal Evolutionary Embedding Approach for Geographic Knowledge Graph Question Answering

1
College of Civil Engineering, Hefei University of Technology, Hefei 230009, China
2
State Key Laboratory of Geographic Information Science and Technology, Institute of Geographic Sciences and Natural Resources Research, Chinese Academy of Sciences, Beijing 100101, China
3
University of Chinese Academy of Sciences, Beijing 100049, China
4
Jiangsu Center for Collaborative Innovation in Geographical Information Resource Development and Application, Nanjing 210046, China
*
Author to whom correspondence should be addressed.
ISPRS Int. J. Geo-Inf. 2025, 14(8), 295; https://doi.org/10.3390/ijgi14080295
Submission received: 30 May 2025 / Revised: 11 July 2025 / Accepted: 24 July 2025 / Published: 28 July 2025
(This article belongs to the Special Issue Spatial Data Science and Knowledge Discovery)

Abstract

In recent years, geographic knowledge graphs (GeoKGs) have shown great promise in representing spatio-temporal and event-driven knowledge. However, existing knowledge graph embedding approaches mainly focus on structural patterns and often overlook the dynamic evolution of entities in both time and space, which limits their effectiveness in downstream reasoning tasks. To address this, we propose a spatio-temporal evolutionary knowledge embedding approach (ST-EKA) that enhances entity representations by modeling their evolution through type-aware encoding, temporal and spatial decay mechanisms, and context aggregation. ST-EKA integrates four core components, including an entity encoder constrained by relational type consistency, a temporal encoder capable of handling both time points and intervals through unified sampling and feedforward encoding, a multi-scale spatial encoder that combines geometric coordinates with semantic attributes, and an evolutionary knowledge encoder that employs attention-based spatio-temporal weighting to capture contextual dynamics. We evaluate ST-EKA on three representative GeoKG datasets—GDELT, ICEWS, and HAD. The results demonstrate that ST-EKA achieves an average improvement of 6.5774% in AUC and 5.0992% in APR on representation learning tasks. In question answering tasks, it yields a maximum average increase of 1.7907% in AUC and 0.5843% in APR. Notably, it exhibits superior performance in chain queries and complex spatio-temporal reasoning, validating its strong robustness, good interpretability, and practical application value.

1. Introduction

The concept of the knowledge graph (KG) was introduced by Google in 2012 with the objective of creating a more intelligent search engine [1,2]. Since 2013, KGs have gained significant traction in both academic and industrial circles. A knowledge graph abstracts diverse knowledge into a complex structure comprising innumerable triples (s, r, o), where s and o represent the head and tail entities, respectively, and r denotes the relationship between them, its essence is a semantic network. The advent and application of KGs have significantly enhanced the level of intelligence in domains such as question answering, information dissemination, and natural language processing [3,4]. The fundamental method for the application of knowledge graphs lies in knowledge graph embedding [5]. Specifically, knowledge graph embedding methods map entities and relationships into a low dimensional vector space to obtain a low dimensional representation [6], which are then input into a scoring function to measure the validity of the triples. Translation embedding (TransE) [7] and translation on hyperplanes (TransH) [8] assume that entities and relations exist within the same vector space, treating relations as translational and rotational transformations between head and tail entities, respectively. However, since relationships and entities are fundamentally different types of objects, they should not share the same vector space. To address this limitation, Translation on relation-specific spaces (TransR) was proposed to construct entity and relation embeddings in separate spaces [9]. To better model symmetrical and anti-symmetrical relations, rotational embedding (RotatE) [10] defined each relation as a rotation in the complex vector space from the source entity to the target entity. Furthermore, translation embedding with rotational representation (TransERR) employs two learnable unit quaternion vectors to rotate the head and tail entities, encoding knowledge graphs in a hyper-complex space [11]. This approach improves knowledge graph completion by minimizing the translational distance between head and tail entities. These models are primarily based on distance variation, while others rely on matrix factorization methods [12], such as relational scalable embedding (RESCAL) models pairwise interactions between latent semantics vectors and latent relation matrix [13]. DistMult simplifies the relation matrices to diagonal matrices [14]. Due to these models focus on the structural information of triples and lack of a comprehensive design for encoding entities and relationships, their accuracy is inherently limited.
Some methods aim to enhance the vector representation of entities by refining entity encoding techniques or incorporating external information to improve performance [15]. For instance, type-aware relation prediction (TaRP) highlights that entity encoding is heavily dependent on entity types, with type-based encoding proving to be more accurate [16]. So it combines type information and ontology information for prediction tasks. Similarly, knowledge graph completion using structural and textual embeddings integrates structural information of nodes, extracted from paths, with textual information represented by pre-trained language models, further enhancing the quality of entity representation [17]. Description-embodied knowledge representation learning (DKRL) also leverages entity description information to enhance entity representations [18]. By incorporating textual descriptions, these methods enrich entity embeddings, enabling them to capture semantic nuances and enhance overall representation quality. However, these approaches primarily focus on the static information of triples. In contrast, real-world knowledge contains rich temporal dynamics. Triples and entities constantly evolve over time, with triples only holding true at specific points or during particular time periods.
To address this shortcoming, temporal knowledge graphs (TKGs) were proposed [19]. TKGs incorporate temporal information into triples, extending them to form quadruples (s, r, t, o), where t represents the time. By introducing temporal constraints to triples, TKGs can effectively capture time-sensitive knowledge. The embedding representation learning methods for TKGs are enhancements of traditional KG methods, tailored to incorporate temporal dimensions [20]. For instance, temporal TransE (TTransE) extends TransE by incorporating temporal information, treating time and relationships as translation transformations between head and tail entities [21]. Similarly, temporal rotation (TeRo) introduces a temporal rotation mechanism, where head and tail entities are rotated over time. It further employs conjugation operations to derive new head and tail entities based on the rotation results of the tail entities [22]. Hyperplane-based temporal embedding (HyTE) incorporates time into the entity-relation space by associating each timestamp with a corresponding hyperplane [23]. These methods primarily focus on the structure of triples or quadruples, or on the entities themselves and their attribute description information. However, in knowledge graphs, entities do not exist in isolation; each entity is interconnected with several others, forming a network influenced by the interactions of multiple entities. To address this, some approaches build on existing representation learning methods by incorporating contextual information. They achieve this by integrating information from adjacent nodes into entity embeddings, enriching the representation of entities within their relational context. For example, a multi-relationship graph neural network (GNN) model calculates the contribution value by comparing neighboring nodes with their corresponding relationships and the target entity. It then sums the contribution representations of all neighboring nodes to derive the final representation of the target entity [24]. Similarly, neighborhood knowledge graph embedding (NKGE) employs an attention mechanism to aggregate the neighboring nodes of an entity, extracting neighborhood information, which is subsequently fused with the entity’s intrinsic features as input [25]. Graph sample and aggregation (GraphSAGE), on the other hand, uses a common aggregation function to combine information from neighboring nodes, thereby generating entity embeddings that capture the contextual information of entities [26]. These methods integrate neighboring entities either by assigning uniform weights to all neighboring nodes or applying varying weights based on relationships. However, these aggregation methods fail to consider the temporal dimension, leading to poor performance when applied to knowledge graphs with rich spatio-temporal information or constantly evolving entities, such as geographic knowledge graph (GeoKG) [27,28,29,30]. Furthermore, these aggregation approaches do not take into account the time decay effect [31] and lack interpretability, limiting their effectiveness in applications requiring a clear understanding of the reasoning process.
Building on this, recent advancements in temporal and spatio-temporal knowledge graph representation learning have introduced a variety of novel modeling strategies. For instance, time-aware fuzzy logic embedding for complex temporal queries (TFLEX) [32] integrates logical reasoning patterns into temporal KGs via time-aware fuzzy logic encoding, significantly improving generalization in multi-hop temporal queries. Relational evolutionary graph convolutional network (RE-GCN) [33] enhances relational modeling by combining graph convolutional operations with recurrent temporal aggregation, enabling more effective capture of time-evolving relationships. Meanwhile, temporal event-aware spatio-temporal transformer (TeAST) [34] introduces an attention-guided time encoder and multi-scale structural learning to improve performance in event-based temporal reasoning. These cutting-edge methods reflect a growing trend toward integrating fine-grained temporal structure, relation-aware propagation, and explainable modeling into temporal KG embedding frameworks. However, most existing models fail to explicitly combine temporal evolution with type-aware semantics—an integration that forms the core innovation of this work. Moreover, their spatial modeling capabilities remain limited, hindering their application in tasks such as regional event analysis, spatial-temporal prediction, and location-based question answering [35]. This indicates that current advancements, while effective in temporal encoding, still fall short in modeling the joint semantics of type-aware entity evolution—particularly within the context of geographic knowledge.
To overcome this challenge, we propose a spatio-temporal evolutionary knowledge embedding approach (ST-EKA). This method explicitly models the spatio-temporal evolution of entities to improve their representation. For each entity, its feature representation is divided into two components: the entity’s intrinsic features and its spatio-temporal evolutionary knowledge. The intrinsic feature representation is enhanced by introducing entity type constraints through relationships. Meanwhile, the evolutionary knowledge of the entity is derived from its neighborhood information. We first classify adjacent quadruples based on the role of the target entity. A quadruple where the target entity serves as the tail entity is considered to represent the formation process of the target entity, while a quadruple where the target entity serves as the head entity is deemed to represent the outcomes of its destruction or impact. By combining these two perspectives, we achieve a comprehensive understanding of the entity’s evolutionary process. Incorporating the time decay effect, we hypothesize that neighboring nodes closer in time exert a greater influence on the target entity. Similarly, we hypothesize that geographically closer entities should also be assigned higher influence. Thus, ST-EKA integrates both time and spatial decay mechanisms to dynamically weight neighboring events [36,37]. To model this process, we adopt an attention mechanism to encode both temporal and spatial embeddings, yielding corresponding weights that are then fused to obtain a spatio-temporal attention score. This score quantifies the influence of neighboring nodes on the target entity. Based on this, we aggregate information from the neighbors to capture the spatio-temporal evolution of the entity. Finally, we combine the entity’s intrinsic features with its evolution-aware neighborhood representation to produce an enhanced entity embedding. We have evaluated the performance of ST-EKA on multiple datasets and applied it to the question answering (Q&A) task [38,39,40], demonstrating its effectiveness.

2. Materials and Methods

2.1. Experimental Data

GeoKGs, central to GeoAI, represent spatio-temporal and event-centric information in the geographic domain. Such GeoKGs support diverse applications such as regional event tracking, disaster response, urban planning, and location-based services [41,42,43]. In this study, we select three typical GeoKG datasets of the Global Database of Events, Language, and Tone (GDELT), Historical Administrative Division (HAD), and Integrated Crisis Early Warning System (ICEWS) for comparative analysis.
Supported by Google Jigsaw, GDELT continuously monitors global broadcasts, print media, and online news in over 100 languages from nearly every region of the world. It extracts key entities—such as people, locations, organizations, themes, sources, emotions, counts, quotes, images, and events—that shape global society, thereby forming a comprehensive and open platform for modeling global dynamics. GDELT is regarded as the largest, most comprehensive, and highest-resolution open database of human society, with historical records spanning more than 215 years. For this study, due to the large data volume, we utilized the GDELT 1.0 simplified event database, which includes all events from 1 December 2009 to 1 January 2010. This subset contains 336,170 triples, including 6385 entities, 238 relations, and 336,170 spatio-temporal records.
This study conducts a spatial analysis of the GDELT dataset by plotting all event coordinates and aggregating event counts at the national level, as shown in Figure 1.
Figure 1a illustrates the global distribution of event locations. Events are densely concentrated in highly urbanized and populous regions—such as Europe, East and South Asia, and the coastal areas of North America—forming clear spatial hotspots. In contrast, sparsely populated regions like Central Asia, the Sahara, and Antarctica record minimal events, reflecting the correlation between human activity intensity and event observability. Figure 1b visualizes the number of events per country, revealing marked regional disparities. Russia, India, the United States, and China rank highest in frequency due to their large populations and advanced media infrastructures. Middle Eastern countries such as Iraq and Syria also show high counts, driven by ongoing conflicts. Meanwhile, many countries in Africa, Central America, and the Pacific report very few events, potentially due to limited media coverage rather than low activity. Together, these figures reveal uneven global event visibility shaped by socio-political, technological, and observational factors.
The HAD dataset is constructed based on geographic divisions defined by China’s administrative hierarchy. It spans the administrative divisions from 1949 to 2019 and primarily consists of three levels: province, city, and county. It captures the structural evolution and jurisdictional changes of administrative divisions over the decades. HAD contains 216,695 triples, including 6454 geographic entities and an equal number of spatio-temporal records.
ICEWS is a comprehensive, integrated, automated, generalizable, and validated system designed to monitor, assess, and forecast national, sub-national, and internal crises. It consists of coded interactions between socio-political actors, including cooperative or hostile actions between individuals, groups, sectors, and nation-states. Events are automatically identified and extracted from news articles using the BBN ACCENT event coder. These events are represented as triples, consisting of a source actor, an event type (according to the CAMEO taxonomy of events), and a target actor. Geographical and temporal metadata are also extracted and associated with the relevant events within a news article. For this study, we used data from the ICEWS Coded Event Data from 2014. The dataset contains a total of 181,460 triples, including 7128 entities, 230 relations, and 181,460 spatio-temporal records.
A partial view of each dataset is illustrated in Figure 2. Then the dataset is divided into training, validation, and testing sets in a ratio of 90:1:9. The dataset partitioning details are shown in Table 1.

2.2. ST-EKA

ST-EKA is designed to enhance entity representations by capturing their spatio-temporal evolution. The framework comprises four core components, as illustrated in Figure 3. The first component is the Entity Encoder. This module leverages the observation that entities connected by the same relation—either as head or tail—tend to share the same type. To model this property, we impose relation-specific similarity constraints and assign semantic types to each entity accordingly. We then employ entity-type matrices to generate embeddings, which are subsequently refined using relational constraints.
The second component is the Temporal Encoder. This module classifies temporal information in the knowledge graph into three categories: time points, time periods, and no time. Each category is processed separately to generate time-aware embeddings, thereby enabling precise temporal representation of entities. In temporal geographic knowledge graphs, however, the influence of neighboring nodes on a target entity is inherently time dependent. As entities and their relationships evolve over time, temporally proximate neighbors typically exert a greater influence, a phenomenon commonly referred to as the time decay effect. Traditional aggregation methods, which assign uniform or static weights to all neighbors regardless of their temporal distance, are inadequate for capturing such temporal dynamics. Consequently, these approaches often yield suboptimal representations for entities with evolving temporal contexts.
The third component is the Spatial Encoder. This module encodes spatial locations across multiple scales and takes into account the displacement relationships between different latitude–longitude coordinates. It then integrates these with the spatial attribute encodings associated with each coordinate to generate a comprehensive spatial representation. This enables the model to capture both fine-grained local spatial details and broader global spatial patterns, thereby enhancing its capacity to represent complex environmental contexts. Moreover, the influence of neighboring nodes on the target entity is spatially variant, exhibiting a distance-based attenuation effect—that is, the greater the spatial distance, the weaker the influence. This spatial decay mechanism reflects the intuition that proximity often correlates with stronger spatial relevance in real-world geospatial systems.
The fourth component is the Evolutionary Knowledge Encoder. To explicitly model the spatio-temporal decay effect, this module assigns corresponding weights to neighboring nodes based on their spatio-temporal proximity. Spatio-temporally closer neighbors receive greater weights, reflecting their stronger influence on the target entity’s current state. The encoder then aggregates the weighted neighboring quadruples to produce a neighborhood-level embedding. This captures the entity’s evolutionary characteristics, thereby enhancing its overall representation.

2.2.1. Entity Encoder

Typically, an entity encoder extracts key features of an entity and encodes them as high-dimensional vectors for subsequent processing in neural network modules. In traditional models, this step is commonly implemented using a randomly initialized embedding matrix. However, TaRP reveals that a substantial portion of information encoded in entity embeddings is inherently type-dependent. Based on this insight, we first classify entities according to the principle that head or tail entities connected by the same relation typically belong to the same type. We then employ an entity-type embedding matrix to encode entities of different types, thereby implicitly incorporating relational constraints. Each column of this matrix corresponds to a distinct entity type, as illustrated in Figure 4. We define the Entity Encoder Enc() as follows: Given any entity e E , where the type of e is l e L , and L is the set of all entity types, the entity encoder proceeds as follows: First represents e as a vector w . Next, perform an embedding lookup operation in the entity type embedding matrix M , where each column corresponds to a specific entity type. This retrieves the embedding of the entity’s type M l e . Multiply the entity embedding w and the entity type embedding M l e to combine the entity’s features with its type-specific features. Finally, normalize the result using the L2 norm to ensure consistency and scale invariance:
ε = E n c ( e ) = M l e w   M l e w L 2
where ε denotes the final entity feature embedding, M l e is the entity type embedding matrix, w is the entity’s encoding vector, and · L 2 represents the L2 norm, which is the square root of the sum of the squared elements of vector.

2.2.2. Time Encoder

The time encoder is designed to express time as a high-dimensional vector, aiming to preserve both the scale information and temporal context, ensuring that no temporal information is lost. Most existing time encoders focus on a single time scalar, and here we also unify the representation of time. All times are unified into a 6-dimensional vector: t = [ y y y y , M M , d d , H H , m m , s s ] representing year, month, day, hour, minute, and second, respectively. Due to the multiple recording forms of time in knowledge graphs, it is crucial for time encoders to be able to handle various types of time. Here we divide the time in the knowledge graph into the following three categories:
Time points [ t 1 , t 2 , , t i ] : Some quadruples may possess multiple time points, such as (First District, is part of, Beijing City, [1949-10-01 00:00:00, 1950-10-01 00:00:00]). In such cases, we select one time point via uniform sampling. This approach ensures that all possible temporal information is eventually leveraged over multiple training steps.
Time period [ t m i n , t m a x ] : For example, (1949-10-01 00:00:00, 1969-10-01 00:00:00). In the model, time intervals are converted to time points for computation. There are various conversion methods, such as using the start, end, or midpoint of the interval as the representative point. Methods that can convert time intervals to time points are generally applicable. However, to retain the scale information inherent in time intervals, simplistic methods like substituting intervals with only their start or end times are inadequate. Thus, similar to the handling of multiple time points, we use uniform sampling within the interval to select a time point.
No time t = : For some quadruples, the time may be missing. In such cases, we assign the time range of the dataset to these quadruples. For instance, if the dataset represents a historical administrative division knowledge graph spanning from 1949-10-01 00:00:00 to 2023-10-01 00:00:00, then we use the interval [1949-10-01 00:00:00, 2023-10-01 00:00:00] as the time for these quadruples. This time interval is then processed in the same manner as other time intervals.
Through time classification and uniform sampling, the three types of time are converted into a single time point for processing. Then we use a feedforward neural network (FFN) encoder to encode the selected time point, generating a time embedding, as shown in Figure 5. Furthermore, through iterative training and dimensionality reduction sampling, the encoder is ensured to accurately encode all three types of time while preserving essential information.
τ = T ( t ) = F F N ( U ( [ t 1 , t 2 , t 3 , , t i ] ) ) i f   t = [ t 1 , t 2 , t 3 , , t i ] F F N ( U ( [ t min , t max ] ) ) i f   t = [ t min , t max ] F F N ( U ( [ t s t a r t , t e n d ] ) ) i f   t =
where τ represents time embedding, U ( ) represents uniform sampling, t s t a r t , t e n d represent the start and end times of the dataset, FFN represents feedforward neural network encoding.

2.2.3. Spatial Encoder

This study designs a spatial encoder that combines Spatial Location and Spatial Attributes (LA). The spatial location is represented by geographic coordinates in the form of a two-dimensional vector x = ( l a t , l o n ) , where lat denotes latitude and lon denotes longitude. Each data point or entity corresponds to a two-dimensional (2D) spatial point x . The spatial attributes include address, city, country, sea area, point of interest (POI), land use type, etc.
(1)
Spatial Location Encoder
This study proposes a multi-scale spatial location encoder. For any given 2D spatial point x , sine and cosine functions with different frequencies are used to encode the coordinate value:
ρ l = E n c p o s ( x ) = N N ( P E ( t ) ( x ) )
where PE ( t ) ( x ) = [ PE 0 ( t ) ( x ) ; ; PE s ( t ) ( x ) ; ; PE S 1 ( t ) ( x ) ] represents the concatenation of multi-scale encodings. The dimensionality of P E ( x ) is 6 S , where S denotes the total number of scales, with s = 0 , 1 , 2 , , S 1 . The function N N ( · ) denotes a fully connected ReLU layer.
Then, three two-dimensional unit vectors are defined:
α 1 = 1 0 , α 2 = 1 2 3 2 , α 3 = 1 2 3 2 2
where α 1 , α 2 , α 3 are three unit vectors with equal angular spacing of 2 π 3 .
At each scale s , the positional encoding P E s ( t ) ( x ) is defined as:
P E s ( t ) ( x ) = [ P E s , 1 ( t ) ( x ) ; P E s , 2 ( t ) ( x ) ; P E s , 3 ( t ) ( x ) ]
where
P E s , j ( t ) ( x ) = cos ( < x , α j > λ min · g s / ( S 1 ) ) sin ( < x , α j > λ min · g s / ( S 1 ) ) , j = 1 , 2 , 3
Here, λ min and λ max denote the minimum and maximum wavelength scales, and g = λ max / λ min controls the scale range. A larger value of g indicates a broader range of scale encoding, allowing the model to perform encoding at more extensive spatial levels. Through this encoding mechanism, the model can simultaneously capture fine-grained local spatial structures and broader global spatial information, thereby enhancing its capability to represent complex spatial distributions.
To ensure translation invariance, the following condition must be satisfied:
E n c p o s ( x + Δ x ) = M ( Δ x ) E n c p o s ( x )
where Δ x represents the displacement in the 2D spatial domain, and the matrix M ( Δ x ) performs rotation and translation operations to achieve position updates. C C = I , where I is the identity matrix.
(2)
Spatial Attribute Encoder
After completing the spatial location embedding, the corresponding spatial attributes such as address, city, country, sea area, point of interest (POI), and land use type are encoded using Word2Vec. These spatial attribute embeddings are concatenated with the position encoding and passed through a linear transformation to obtain the final spatial feature vector, as shown in Figure 6.
ρ = L i n e a r ( ρ l W 2 V ( x ) W 2 V ( y ) W 2 V ( z ) multivariable :   x , y , , z )
Here, W 2 V denotes the Word2Vec embedding of the spatial attribute values x, y, …, z; represents vector concatenation; and L i n e a r ( ) is a fully connected linear transformation layer.

2.2.4. Evolutionary Knowledge Encoder

Considering the spatio-temporal decay effect, we hypothesize that the spatio-temporal proximity between adjacent entities significantly affects the influence exerted by its neighboring nodes. Specifically, the closer the spatio-temporal proximity, the greater the influence of neighboring nodes on the entity [44,45,46,47].
Firstly, adjacent quadruples are categorized into two types based on the position of the target entity within the quadruple:
(1)
When the target entity appears as the tail entity, the corresponding quadruple is regarded as part of the formation process of the entity. For example, in the quadruple (Continuous heavy rainfall, caused, 18 March 2025, 4:50 a.m., Fugong landslide), “Fugong landslide” is the target entity, and “continuous heavy rainfall” is the inducing factor.
(2)
When the target entity appears as the head entity, the corresponding quadruple reflects the consequences or impact exerted by the entity on other entities. For instance, in (Fugong landslide, caused, 18 March 2025, 4:50 a.m., 1 dead and 4 missing), “Fugong landslide” is the target entity, and “1 dead and 4 missing” represents the resulting impact.
These two types of quadruples jointly reveal the spatio-temporal evolution trajectory of the target entity within the knowledge graph. However, directly modeling both types of relationships would significantly increase the computational and modeling complexity.
To simplify the algorithm, reverse relations are added for all existing relations, and the corresponding quadruples are inverted. This transformation simplifies the problem by focusing only on quadruples in which the entity serves as the tail. The attention mechanism [48] is then applied to calculate the weights for the time embedding and spatial embedding.
w = S o f t m a x ( Q · K T d k ) · V
where Q, K, and V are the query, key, and value matrices generated by linear layers for each adjacent quadruple time embedding, and dk is the dimension.
Subsequently, the spatial and temporal weights are aggregated to obtain an integrated spatio-temporal weight, as illustrated in Figure 7. Then, for each adjacent quadruple, the embedding is computed based on the embedding of the neighboring node, the corresponding relation, and the spatio-temporal weight assigned to the group to which the quadruple belongs:
ε = ε R r ( w t w l )
where ε denotes the neighboring entity, and R r denotes the relation corresponding to the neighboring entity.
After performing the above calculation on all adjacent quadruples, the encodings of all adjacent quadruples are aggregated to obtain the neighborhood information of the entity, as shown in Figure 8:
n e i = i = 1 m ε i R r i w i
where ε i represents the encoding of adjacent node, R r i represents the relationship between adjacent node and target node, w i is the corresponding spatio-temporal weight, and m is the number of adjacent nodes.
Finally, we fuse neighborhood information with entity embedding to obtain a spatio-temporal evolutionary knowledge aware entity enhanced representation, and calculate it with relationship and time to obtain the predict tail entity embedding, denoted as σ :
σ = ( n e i + ε h ) R r + τ
where n e i represents the entity spatio-temporal evolution embedding, ε h represents the head entity embedding, τ represents the time embedding.
We then use the cosine similarity function to calculate the loss values between σ and both the positive and negative samples, and obtain the final loss value:
c o s e 1 , e 2 = e 1 e 2 e 1 e 2
l o s s = 1 ( cos ( σ , ε o ) cos ( σ , ε n e g ) )
where ε o represents the positive tail entity embedding, and ε n e g represents the negative tail entity embedding.
The training objective is to minimize the distance between the predicted embedding and the correct entity embedding, while maximizing the gap with the negative samples. In other words, the goal is to make cos ( σ , ε o ) as close to 1 as possible, cos ( σ , ε n e g ) as close to 0 as possible, and ultimately minimize loss to approach 0.

2.3. Question-Answering Model

We further evaluate the performance of ST-EKA on the question answering (Q&A) task. Graph query embedding (GQE) is an end-to-end logical query answering model designed to handle connected graph queries [49]. Contextual graph attention (CGA) extends GQE by integrating a self-attention-based intersection operator [50]. In this study, we adopt these two models as baselines and integrate ST-EKA into their architectures.
In Q&A tasks, any complex query graph (QG) can be decomposed into a combination of chain and intersection structures. The chain structure is depicted in Figure 9a, whereas the intersection structure is illustrated in Figure 9b. As illustrated, the chain structure involves the processing of simple quadruples, for which ST-EKA is applied. In contrast, the intersection structure aggregates the outputs of several sub-query graphs to produce a combined result. To perform this operation, we retain the original intersection operator from the baseline models. We designed six types of QGs and conducted Q&A experiments based on these structures, incorporating both ST-EKA and the baseline intersection operators. A detailed introduction to the query graphs is provided in Section 3.3.
The intersection operator is used to consolidate multiple vectors into a single vector. Typically, any method that preserves the original sequence of input vectors, such as taking the minimum, maximum, or average, is applicable. The GQE model utilizes a linear feedforward network combined with a minimum function as the intersection operator, as shown in Figure 10a. The CGA model improves this intersection operator by incorporating a self-attention mechanism, as shown in Figure 10b.
When equipped with these two basic neural network modules, any complex query graph can be encoded according to its structure to obtain a unique result vector. This vector is then used to retrieve the most probable answer through nearest neighbor search in the entity space.

2.4. Model Training

2.4.1. Experimental Platform Configuration

The experiments in this study were conducted on a system equipped with an 11th Gen Intel (R) Core (TM) i9-11900K CPU @ 3.50 GHz, an NVIDIA GeForce RTX 3080 Ti GPU with 28 GB of memory, and 32 GB of RAM. The entire experiment was performed in a Windows environment. The implementation was based on Python 3.9 and PyTorch 1.12.1.

2.4.2. Model Parameter Settings

To mitigate overfitting, both the representation learning and question-answering tasks employ an early stopping strategy based on a predefined tolerance threshold. When the performance improvement falls below this threshold, the model is considered to have converged, and training is terminated early. The learning rate, which controls the step size for updating model parameters, has a critical impact on model performance. An excessively large learning rate may result in unstable loss fluctuations, whereas an overly small rate may lead to slow convergence. In our experiments, we tested multiple learning rates and observed that the optimal performance was achieved with a learning rate of 0.001 in combination with the Adam optimizer. This training configuration ensures both computational efficiency and high model performance. The detailed hyper parameter settings for each model are summarized in Table 2.

2.4.3. Training Phase

The training of the Q&A model consists of the intrinsic structure of the KG and a series of query-answer pairs. The final loss function is defined as follows:
L = L K G + L Q A
(1)
KG Training Phase
During the KG training phase, we treat the quadruples in the KG as the basis chain query graph, predicting the tail entity using ST-EKA. The purpose of this step is to pre-train the model and obtain embeddings for entities, relationships, and time:
L K G = i = 1 n max ( 0 , 1 c o s e i p , e i o c o s e i p , e i n )
where e i p represents the predicted result, e i o represents the tail entity, and e i n represents the negative sample.
(2)
Question-Answer Pair Training Phase
We design six types of query graphs, and sample each type to obtain question-answer pairs in the form of ( Q , A , N e g ) , where Q represents the question with a different query graph type, A represents the correct answer, and N e g represents the incorrect answer, i.e., the negative sample. The sampling method first sorts the nodes within the query graph according to their topological sequence and then performs the sampling based on this order, selecting one basic graph structure at a time. The training objective is to minimize the distance between the predicted embedding and the correct entity embedding while maximizing the gap with the negative samples. Cosine similarity is chosen as the method to compute the similarity:
L Q A = q i , a i χ i = 1 X max ( 0 , 1 ω ( ψ ( Q i ) , A i ) ω ( ψ ( Q i ) , N e g i ) )
where ψ ( ) represents the function that calculates the predicted result based on the query graph structure, and a n signifies the negative sample.

3. Results and Discussion

Model performance is evaluated using two metrics: area under the receiver operating characteristic curve (AUC) and average percentile rank (APR). AUC quantifies the model’s ability to distinguish correct answers from negative samples by computing the area under the ROC curve, which reflects the trade-off between true positive and false positive rates. A higher AUC value (closer to 1) indicates stronger discriminative capability. APR measures the average percentile position of the correct answer among all predicted candidates. It reflects how effectively the model ranks correct answers relative to negative ones—higher APR values indicate better ranking performance.

3.1. Comparison Between ST-EKA and Other Representation Learning Methods

We selected five state-of-the-art (SOTA) methods for comparison: TransE, TTransE, TeRo, HyTE, and TeAST. ST-EKA aggregates adjacent node information, but in a knowledge graph, the number of neighbors of a node can vary significantly—some nodes have only one neighbor, while others may have thousands. When there is a large number of neighboring nodes, considering all of them becomes inefficient. Therefore, it is necessary to sample the neighboring nodes; we base our sampling strategy on time. Specifically, we first sort the quadruples with the target entity as the tail entity by time and select the three quadruples with the largest time as the causal quadruples. For quadruples with the target entity as the head entity, we select the three quadruples with the smallest time as result quadruples. These six quadruples are then used as neighboring quadruples for training. After sampling, ST-EKA is applied for training. The experimental results are shown in Table 3.
From Table 3, it can be observed that ST-EKA achieves the highest AUC and APR scores across all datasets, with the most notable improvement on the GDELT dataset. Compared with baseline methods, ST-EKA improves AUC by 4.7315% over TransE, 0.7603% over TTransE, 2.0409% over TeRo, 0.5709% over HyTE, and 2.7965% over TeAST. Unlike entity-centric knowledge graphs, GDELT focuses on dynamic international events, where temporal progression and evolving relationships among geopolitical actors are essential for understanding event causality and influence chains. ST-EKA effectively models these dynamics by incorporating both the temporal decay of neighboring quadruples and type-aware entity constraints. This results in more accurate and context-rich event embeddings, which are critical for downstream geographic question-answering tasks that require fine-grained temporal inference.
On HAD, ST-EKA improves AUC by 14.0258% over TransE, 11.0853% over TTransE, 17.4138% over TeRo, 11.2204% over HyTE, and 10.9222% over TeAST. On ICEWS, ST-EKA outperforms TransE by 11.7145%, as well as achieving respective gains of 11.7145%, 4.5006%, 2.3338%, 1.4468%, and 3.0977% over other methods. These results confirm that the spatio-temporal evolutionary embedding mechanism introduced in ST-EKA can be effectively generalized to different domains, including structured administrative knowledge (HAD) and socio-political event streams (ICEWS).
Based on the results from all three datasets, ST-EKA effectively improves representation learning by assigning types to entities through relationships, enhancing entity representation with relationship constraints, and incorporating weights based on the temporal dynamics of quadruples. These strategies allow ST-EKA to capture entity evolution knowledge and further refine the entity representation. Overall, ST-EKA is shown to be both effective and efficient.

3.2. Ablation Study

To evaluate the contribution of each component in ST-EKA, we conducted ablation experiments on the HAD, GDELT, and ICEWS datasets. The following model variants were assessed:
(1)
w/o TimeDecay: This variant removes the temporal weighting module from the evolutionary knowledge encoder and uses only spatial weight for aggregation.
(2)
w/o SpatialDecay: This variant removes the spatial weighting module from the evolutionary knowledge encoder and uses only temporal weight for aggregation.
(3)
w/o Evolutionary Knowledge Encoder: This variant eliminates the entire evolutionary knowledge encoder module.
(4)
Number of adjacent quadruples: This setting tests the effect of varying the number of adjacent quadruples used in the evolutionary knowledge encoder. The sampling strategy for adjacent quadruples is described in Section 3.1.
It is important to note that, unless otherwise specified, the number of adjacent quadruples used in other experiments, including the w/o TimeDecay variant and the w/o SpatialDecay, is fixed at 6. The experimental results are summarized in Table 4.
As shown in Table 4, removing the temporal weight (w/o TimeDecay) leads to performance degradation across all datasets. On the GDELT dataset, the AUC drops from 0.949833 to 0.931227, suggesting that using only spatial weight to aggregate neighboring quadruples fails to capture their dynamic temporal correlations, thereby neglecting the time decay effect. In addition, the ablation of the spatial weight (w/o SpatialDecay) also leads to a performance drop, with the AUC decreasing to 0.940203, demonstrating that neglecting temporal proximity similarly weakens the model’s ability to capture contextual influence—especially in event-centric graphs where temporal plays a critical role. The performance decline is even more pronounced on the HAD dataset. Removing the time decay mechanism results in an AUC drop from 0.782160 to 0.601407, and eliminating the spatial decay module reduces it to 0.607226, reflecting relative decreases of 18.0753% and 17.4934%, respectively. Moreover, removing the evolutionary knowledge encoder results in the most significant performance decline—specifically, a drop of 2.0709% in AUC on the GDELT dataset, 32.6965% on the HAD dataset, and 3.9654% on the ICEWS dataset. This highlights the indispensable role of entity evolutionary knowledge in handling diverse and heterogeneous datasets.
Meanwhile, as the number of adjacent quadruples used by the evolutionary knowledge encoder increases, the model performance steadily improves. On GDELT, the AUC grows from 0.937942 (2 quadruples) to 0.949833 (6 quadruples) and remains high at 0.947735 with 8 quadruples. Similar patterns are observed on HAD and ICEWS, affirming that richer spatio-temporal context enhances representation learning. These results collectively confirm the effectiveness of the proposed temporal and spatial weighting mechanisms and the evolutionary knowledge encoder in capturing spatio-temporal entity evolution for event-based reasoning tasks.

3.3. Application of ST-EKA in Geographic Knowledge Graph Question-Answering Tasks

After confirming ST-EKA’s strong performance in representation learning, we focused on its application to downstream query question answering tasks. In this experiment, we designed six types of query graphs for the Q&A task, as shown in Figure 11:
2-inter: Intersection structure formed by two quadruples whose tail entities are connected.
2-chain: A chain structure composed sequentially of two quadruples’ head and tail entities.
3-chain_inter: Formed by connecting the tail node of a 2-inter structure to another quadruple.
3-inter_chain: Formed by connecting the tail node of a 2-chain structure to another quadruple.
3-inter: Intersection structure formed by three quadruples whose tail entities are connected.
3-chain: A chain structure composed sequentially of three quadruples’ head and tail entities.
We first conducted question–answer sampling based on these six query graphs. The training QAs must have answers derived from the training set, meaning the training QAs are sampled based on the training quadruple set. In contrast, validation and test QAs are sampled from the entire dataset. Specifically, each QA in the validation and test sets contains at least one quadruple not present in the training set. The number of QAs is shown in Table 5.
We selected two baseline models, CGA and GQE, and conducted comparative experiments by replacing the chain structure encoder in the original models with ST-EKA. This allowed us to assess how ST-EKA’s entity evolution modeling could improve the performance of these existing query answering frameworks. The experimental results are shown in Table 6, Table 7 and Table 8.
As shown in Table 6, Table 7 and Table 8, the proposed ST-EKA model significantly improves the performance of geographic question-answering tasks. Among the three datasets, the GDELT dataset serves as the most representative benchmark due to its high event density, global scope, and fine-grained temporal granularity. In this context, ST-EKA demonstrates substantial improvements across all QG types. Taking the GDELT dataset as the main case, when integrated with the GQE model, ST-EKA improves the macro-averaged AUC by 1.5295% and APR by 0.1819. Based on the CGA model, the AUC increases by 3.4850% and the APR by 0.6480. These results confirm that the addition of ST-EKA effectively improves the overall query accuracy of the model.
Notably, ST-EKA explicitly models entity evolution by aggregating neighboring quadruples based on both temporal and spatial distances. Unlike prior methods that treat adjacent entities equally or rely solely on time-aware aggregation, ST-EKA introduces a dual-decay mechanism that assigns dynamic attention weights according to both time proximity and spatial proximity. This enables the model to capture not only temporal dependencies but also geographic co-occurrence patterns, leading to higher-quality event embeddings. Such fine-grained representation is particularly beneficial in geographic reasoning tasks, where the influence of nearby events is often subject to both spatial and temporal attenuation.
For chain-based query structures such as 2-chain and 3-chain, the improvements are even more significant—demonstrating that ST-EKA excels at sequential reasoning by capturing the progressive evolution of entities across space and time. For instance, in the GDELT dataset, the AUC improvement for the 3-chain QG under GQE reaches 4.7434%.
As shown in Figure 12, although ST-EKA also improves intersection-based QGs, the margin is relatively smaller—especially in the CGA backbone—where existing attention-based intersection operators already handle complex queries effectively. In fact, as observed in both the GDELT and HAD datasets, when the query structure contains intersection components, the intersection processors could not effectively integrate the evolutionary knowledge of entities, leading to less improvement or even a slight decrease in query accuracy. This suggests the need to redesign and optimize the intersection operator to better leverage the spatio-temporal embeddings produced by ST-EKA.
In HAD, which contains structured historical administrative changes, ST-EKA improves the GQE model’s macro AUC and APR by 0.5913% and 0.1592, respectively. In the ICEWS dataset—a temporal event dataset with clear geopolitical framing—macro AUC and APR improvements reach 1.2502% and 0.7830 under GQE, and 1.3633% and 0.7303 under CGA, respectively. These results underscore the adaptability of ST-EKA to various temporal and spatial resolutions and its strong capacity for generalization across different event types and spatial scales.
In summary, the integration of temporal and spatial decay mechanisms empowers ST-EKA to capture both temporal correlations and spatial influences, providing a more realistic and nuanced view of entity evolution. This design is especially valuable in geographic knowledge graphs, where entities often exhibit strong spatio-temporal clustering. By enhancing the quality of entity embeddings, ST-EKA not only boosts question answering accuracy but also sets a foundation for more interpretable and context-aware reasoning in spatio-temporal domains.

4. Discussion

In this study, we propose ST-EKA, a spatio-temporal evolutionary knowledge embedding approach, as a foundational representation learning model designed for geographic knowledge graph Q&A. Unlike traditional methods that primarily focus on structural information from static triples, ST-EKA captures the dynamic nature of entities by aggregating temporally and spatially adjacent quadruples. Through this fusion of temporal and spatial contexts, ST-EKA forms a structured representation of entity evolution, enhancing the expressiveness of embeddings for downstream tasks.
ST-EKA first categorizes temporal information in the GeoKG into distinct types and applies corresponding encoding strategies. Then, to overcome the neglect of spatio-temporal decay in neighbor aggregation, ST-EKA introduces an evolutionary knowledge encoder that divides neighboring quadruples into process-oriented and result-oriented types based on the head/tail role of the target entity. By assigning attention weights that decay with both temporal and spatial distance, the model emphasizes contextually relevant events and attenuates distant noise. Meanwhile, ontological features are enhanced by entity-type constraints. By integrating both spatio-temporal evolution and type-aware semantic information, the model produces enriched embeddings capable of capturing complex real-world transformations.
To verify the effectiveness of ST-EKA, we conducted comparative experiments against five state-of-the-art representation models on three representative datasets. Results in Table 3 confirm that ST-EKA performs particularly well in dynamic environments with rich spatio-temporal correlations, such as the GDELT and ICEWS datasets. On these datasets, where events exhibit both frequent temporal changes and broad geographic diversity, ST-EKA consistently outperforms the baseline models. In contrast, on the simpler HAD dataset, where variability is lower, models tend to overfit more easily, resulting in overall lower accuracy. However, the simplicity of HAD also forces the models to rely more heavily on entity features, thereby amplifying the impact of entity evolution knowledge. Consequently, ST-EKA achieves its most pronounced performance improvements on this dataset. Additionally, since administrative changes in HAD often occur abruptly, the ability of ST-EKA to encode contextual dependencies offers unique advantages in capturing such irregular patterns.
The performance of ST-EKA was further validated through its application in geographic question answering tasks. By replacing the original chain structure encoder in the QA framework with ST-EKA, we evaluated its impact across various query graph structures, with results shown in Table 6, Table 7 and Table 8. In these tasks, accurate contextual reasoning is crucial, and ST-EKA excels by embedding evolution-based knowledge derived from spatial and temporal proximities. For instance, in the GDELT dataset—characterized by dense global events—ST-EKA improves both AUC and APR significantly under both GQE and CGA backbones. However, for intersection-type queries, especially under CGA, the improvement is less stable. This is due to existing intersection processors’ limited ability to effectively absorb evolutionary embeddings. Thus, future work should explore adaptive intersection operators tailored to leverage spatio-temporal context. Nonetheless, ST-EKA consistently improves query accuracy across all datasets and structures, highlighting its robustness and versatility.
Although the integration of temporal attention and spatio-temporal aggregation introduces moderate computational overhead, this is offset by its performance gains. The model remains tractable due to lightweight neural modules and efficient sampling strategies. In practice, runtime is slightly higher than simpler baselines, but acceptable. In future work, we plan to explore optimization techniques such as sparse temporal attention and neighbor pruning to further reduce complexity without sacrificing model performance.
Finally, as a general-purpose embedding method, ST-EKA shows strong potential for application in other spatio-temporal domains. Its precise modeling of entity evolution makes it particularly suitable for knowledge reasoning in geographic event monitoring, geographic decision support, and public safety systems.

5. Conclusions and Future Work

This article explores the limitations of existing representation learning methods when applied to knowledge graphs with rich spatio-temporal information. We take GeoKG as the research focus and propose ST-EKA, which aims to capture the rich evolutionary knowledge of entities and enhance their representations. The goal is to improve the representation accuracy of knowledge graphs that contain complex, constantly evolving entities and rich spatio-temporal information. The main contributions of this study are as follows: For the entity itself, we assign types based on its associated relations and encode the entity using a type matrix to enhance its representation. In addition, we design separate temporal and spatial encoders to comprehensively capture temporal and spatial characteristics. Furthermore, we develop an evolutionary knowledge encoder, which treats quadruples where the entity appears as the tail as part of its formation process and those where the entity appears as the head as the result of its influence. Based on spatio-temporal distance, we assign weights to these neighboring pieces of information and integrate the evolutionary knowledge to further strengthen the entity’s representation.
This method can be applied to various downstream tasks, such as the completion, inference, and prediction of spatio-temporal knowledge graphs. Currently, spatial positions are represented as two-dimensional coordinate points. However, geographic entities include multiple geometric types—such as points, lines, and polygons—whose coordinates are not limited to single latitude-longitude pairs. In future work, we plan to incorporate knowledge graphs containing complex geographic entities and refine the spatial encoding mechanism to enable accurate encoding of spatial positions across different object types, thereby facilitating precise spatial feature extraction. Moreover, because of dataset limitations, the current implementation enhances entity representations solely through type information. However, entities—particularly geographic ones—often contain diverse and complex attributes, and relying only on types may be insufficient to capture this complexity. These limitations may pose challenges in more multifaceted application scenarios.
Future research will focus on integrating spatial dimensions and multivariate entity attributes to further enhance the semantic expressiveness of the model. At the same time, improving the generalizability and scalability of the proposed approach represents a key direction for future work. On the one hand, the method can be extended to a broader range of application domains; on the other hand, experiments conducted across diverse regional datasets will help evaluate its robustness and adaptability. For instance, in Geographic Information System applications, the proposed method can improve system intelligence by enhancing entity disambiguation and user-oriented service accuracy. In the domain of disaster management, it enables precise spatio-temporal tracking of disaster events, thereby improving the performance of disaster modeling and impact assessment.

Author Contributions

Conceptualization, Chunju Zhang; data curation, Kang Zhou; formal analysis, Chaoqun Chu; funding acquisition, Shu Wang and Yunqiang Zhu; investigation, Chaoqun Chu; methodology, Chaoqun Chu; project administration, Chunju Zhang, Shu Wang and Yunqiang Zhu; software, Chaoqun Chu and Kang Zhou; supervision, Chunju Zhang, Shu Wang and Yunqiang Zhu; validation, Chaoqun Chu; writing—original draft, Chaoqun Chu; writing—review and editing, Chaoqun Chu, Jianwei Huang, Zhaofu Wu and Fei Gao. All authors have read and agreed to the published version of the manuscript.

Funding

This work was partially funded by the National Key Research and Development Program of China (Grant No. 2022YFB3904200), the Fundamental Research Funds for the Central Universities (Grant No. JZ2024HGTG0288), and the National Natural Science Foundation of China (Grant No. 42171453).

Data Availability Statement

The ICEWS dataset is available at https://www.lockheedmartin.com/en-us/capabilities/research-labs/advanced-technology-labs/icews.html (accessed on 25 November 2023). The GDELT dataset is available at https://www.gdeltproject.org/data.html (accessed on 26 January 2024).

Acknowledgments

We acknowledge the data support from the ICEWS and the GDELT project.

Conflicts of Interest

The authors declare no conflicts of interest.

References

  1. Song, Q.; Wu, Y.; Lin, P.; Dong, L.; Sun, H. Mining Summaries for Knowledge Graph Search. In Proceedings of the 2016 IEEE 16th International Conference on Data Mining (ICDM), Barcelona, Spain, 12–15 December 2016; pp. 1215–1220. [Google Scholar]
  2. Chen, X.; Jia, S.; Xiang, Y. A Review: Knowledge Reasoning over Knowledge Graph. Expert Syst. Appl. 2020, 141, 112948. [Google Scholar] [CrossRef]
  3. Liu, Y.; Xu, Z.; Zhao, J.; Song, C.; Wang, D. Hierarchical Fault Propagation Path Recognition Method Based on Knowledge-Driven Graph Attention Autoencoder with Bilayer Pooling for Large-Scale Industrial System. Adv. Eng. Inform. 2025, 63, 102930. [Google Scholar] [CrossRef]
  4. Nickel, M.; Murphy, K.; Tresp, V.; Gabrilovich, E. A Review of Relational Machine Learning for Knowledge Graphs. Proc. IEEE 2016, 104, 11–33. [Google Scholar] [CrossRef]
  5. Ji, G.; Liu, K.; He, S.; Zhao, J. Knowledge Graph Completion with Adaptive Sparse Transfer Matrix. In Proceedings of the AAAI Conference on Artificial Intelligence, Phoenix, AZ, USA, 12–17 February 2016; Volume 30. [Google Scholar]
  6. Socher, R.; Chen, D.; Manning, C.D.; Ng, A. Reasoning with Neural Tensor Networks for Knowledge Base Completion. In Proceedings of the Advances in Neural Information Processing Systems 26: 27th Annual Conference on Neural Information Processing Systems 2013, Lake Tahoe, NV, USA, 5–8 December 2013. [Google Scholar]
  7. Bordes, A.; Usunier, N.; Garcia-Durán, A.; Weston, J.; Yakhnenko, O. Translating Embeddings for Modeling Multi-Relational Data. In Proceedings of the 26th International Conference on Neural Information Processing Systems, Vancouver, BC, Canada, 8–14 December 2019; Curran Associates Inc.: Red Hook, NY, USA, 2013; Volume 2, pp. 2787–2795. [Google Scholar]
  8. Wang, Z.; Zhang, J.; Feng, J.; Chen, Z. Knowledge Graph Embedding by Translating on Hyperplanes. In Proceedings of the Twenty-Eighth AAAI Conference on Artificial Intelligence, Québec City, QC, Canada, 27–31 July 2014; Volume 28. [Google Scholar] [CrossRef]
  9. Lin, Y.; Liu, Z.; Sun, M.; Liu, Y.; Zhu, X. Learning Entity and Relation Embeddings for Knowledge Graph Completion. In Proceedings of the AAAI Conference on Artificial Intelligence, Austin, TX, USA, 25–30 January 2015; Volume 29. [Google Scholar] [CrossRef]
  10. Sun, Z.; Deng, Z.; Nie, J.-Y.; Tang, J. RotatE: Knowledge Graph Embedding by Relational Rotation in Complex Space. arXiv 2018, arXiv:1902.10197. [Google Scholar]
  11. Li, J.; Su, X. TransERR: Translation-Based Knowledge Graph Embedding via Efficient Relation Rotation. In Proceedings of the 2024 Joint International Conference on Computational Linguistics, Torino, Italy, 26 June 2023. [Google Scholar]
  12. Wang, Q.; Mao, Z.; Wang, B.; Guo, L. Knowledge Graph Embedding: A Survey of Approaches and Applications. IEEE Trans. Knowl. Data Eng. 2017, 29, 2724–2743. [Google Scholar] [CrossRef]
  13. Nickel, M.; Tresp, V.; Kriegel, H.-P. A Three-Way Model for Collective Learning on Multi-Relational Data. In Proceedings of the 28th International Conference on International Conference on Machine Learning, Bellevue, WA, USA, 28 June–2 July 2011; Omnipress: Madison, WI, USA, 2011; pp. 809–816. [Google Scholar]
  14. Yang, B.; Yih, W.; He, X.; Gao, J.; Deng, L. Embedding Entities and Relations for Learning and Inference in Knowledge Bases. In Proceedings of the International Conference on Learning Representations, Banff, AB, Canada, 19 December 2014. [Google Scholar]
  15. Pujara, J.; Augustine, E.; Getoor, L. Sparsity and Noise: Where Knowledge Graph Embeddings Fall Short. In Proceedings of the 2017 Conference on Empirical Methods in Natural Language Processing, Copenhagen, Denmark, 7–11 September 2017; Association for Computational Linguistics: Copenhagen, Denmark, 2017; pp. 1751–1756. [Google Scholar]
  16. Cui, Z.; Kapanipathi, P.; Talamadupula, K.; Gao, T.; Ji, Q. Type-Augmented Relation Prediction in Knowledge Graphs. In Proceedings of the The Thirty-Fifth AAAI Conference on Artificial Intelligence, Vancouver, BC, Canada, 2–9 February 2021; Volume 35, pp. 7151–7159. [Google Scholar] [CrossRef]
  17. Alqaaidi, S.K.; Kochut, K.J. Knowledge Graph Completion Using Structural and Textual Embeddings; Maglogiannis, I., Iliadis, L., Macintyre, J., Avlonitis, M., Papaleonidas, A., Eds.; Springer Nature: Cham, Switzerland, 2024; Volume 713, pp. 235–247. [Google Scholar]
  18. Xie, R.; Liu, Z.; Jia, J.; Luan, H.; Sun, M. Representation Learning of Knowledge Graphs with Entity Descriptions. In Proceedings of the AAAI Conference on Artificial Intelligence, Phoenix, AZ, USA, 5 March 2016; Volume 30. [Google Scholar]
  19. Ji, S.; Pan, S.; Cambria, E.; Marttinen, P.; Yu, P.S. A Survey on Knowledge Graphs: Representation, Acquisition, and Applications. IEEE Trans. Neural Netw. Learn. Syst. 2022, 33, 494–514. [Google Scholar] [CrossRef]
  20. Cai, B.; Xiang, Y.; Gao, L.; Zhang, H.; Li, Y.; Li, J. Temporal Knowledge Graph Completion: A Survey. In Proceedings of the Thirty-Second International Joint Conference on Artificial Intelligence, Macau, China, 19–25 August 2023; International Joint Conferences on Artificial Intelligence Organization: Macau, China, 2023; pp. 6545–6553. [Google Scholar]
  21. Leblay, J.; Chekol, M.W. Deriving Validity Time in Knowledge Graph. In Proceedings of the Companion of the The Web Conference 2018 on The Web Conference 2018—WWW ’18, Lyon, France, 23–27 April 2018; Association for Computing Machinery: New York, NY, USA, 2018; pp. 1771–1776. [Google Scholar] [CrossRef]
  22. Xu, C.; Nayyeri, M.; Alkhoury, F.; Shariat Yazdi, H.; Lehmann, J. TeRo: A Time-Aware Knowledge Graph Embedding via Temporal Rotation. In Proceedings of the 28th International Conference on Computational Linguistics, Barcelona, Spain, 8–13 December 2020; pp. 1583–1593. [Google Scholar]
  23. Dasgupta, S.S.; Ray, S.N.; Talukdar, P. HyTE: Hyperplane-Based Temporally Aware Knowledge Graph Embedding. In Proceedings of the 2018 Conference on Empirical Methods in Natural Language Processing, Brussels, Belgium, 31 October–4 November 2018; pp. 2001–2011. [Google Scholar]
  24. Vashishth, S.; Sanyal, S.; Nitin, V.; Talukdar, P. Composition-Based Multi-Relational Graph Convolutional Networks. arXiv 2019, arXiv:1911.03082. [Google Scholar]
  25. Wang, K.; Liu, Y.; Xu, X.; Lin, D. Knowledge Graph Embedding with Entity Neighbors and Deep Memory Network. arXiv 2018, arXiv:1808.03752. [Google Scholar]
  26. Hamilton, W.L.; Ying, Z.; Leskovec, J. Inductive Representation Learning on Large Graphs. arXiv 2017, arXiv:1706.02216. [Google Scholar]
  27. Wang, S.; Zhang, X.; Ye, P.; Du, M.; Lu, Y.; Xue, H. Geographic Knowledge Graph (GeoKG): A Formalized Geographic Knowledge Representation. ISPRS Int. J. Geo-Inf. 2019, 8, 184. [Google Scholar] [CrossRef]
  28. Wang, J.; Liu, K.; Li, H. LSTM-Based Graph Attention Network for Vehicle Trajectory Prediction. Comput. Netw. 2024, 248, 110477. [Google Scholar] [CrossRef]
  29. Jiang, B.; Tan, L.; Ren, Y.; Li, F. Intelligent Interaction with Virtual Geographical Environments Based on Geographic Knowledge Graph. ISPRS Int. J. Geo-Inf. 2019, 8, 428. [Google Scholar] [CrossRef]
  30. Yan, B.; Janowicz, K.; Mai, G.; Zhu, R. A Spatially Explicit Reinforcement Learning Model for Geographic Knowledge Graph Summarization. Trans. GIS 2019, 23, 620–640. [Google Scholar] [CrossRef]
  31. Chen, Y.-C.; Hui, L.; Thaipisutikul, T. A Collaborative Filtering Recommendation System with Dynamic Time Decay. J. Supercomput. 2021, 77, 244–262. [Google Scholar] [CrossRef]
  32. Lin, X.; Xu, C.; E, H.; Su, F.; Zhou, G.; Hu, T.; Li, N.; Sun, M.; Luo, H. TFLEX: Temporal Feature-Logic Embedding Framework for Complex Reasoning over Temporal Knowledge Graph. arXiv 2022. [Google Scholar] [CrossRef]
  33. Zhang, S.; Liang, X.; Niu, S.; Niu, Z.; Wu, B.; Hua, G.; Wang, L.; Guan, Z.; Wang, H.; Zhang, X.; et al. Integrating Large Language Models and Möbius Group Transformations for Temporal Knowledge Graph Embedding on the Riemann Sphere. In Proceedings of the AAAI Conference on Artificial Intelligence, Philadelphia, PA, USA, 11 April 2025; Volume 39, pp. 13277–13285. [Google Scholar]
  34. Li, J.; Su, X.; Gao, G. TeAST: Temporal Knowledge Graph Embedding via Archimedean Spiral Timeline. In Proceedings of the 61st Annual Meeting of the Association for Computational Linguistics, Toronto, ON, Canada, 9–14 July 2023; Association for Computational Linguistics: Toronto, ON, Canada, 2023; pp. 15460–15474. [Google Scholar]
  35. Böckling, M.; Paulheim, H.; Detzler, S. Comparing Spatial-Temporal Knowledge Graph on Spatial Downstream Tasks. In Proceedings of the 32nd ACM International Conference on Advances in Geographic Information Systems, New York, NY, USA, 22 November 2024; pp. 581–584. [Google Scholar]
  36. Ji, X.; Cao, Y.; Zhang, J.; Zhao, X. STSE: Spatio-Temporal State Embedding for Knowledge Graph Completion. Knowl.-Based Syst. 2025, 317, 113469. [Google Scholar] [CrossRef]
  37. Hu, L.; Li, W.; Zhu, Y. Geometric Feature Enhanced Knowledge Graph Embedding and Spatial Reasoning 2024. In Proceedings of the 7th ACM SIGSPATIAL International Workshop on AI for Geographic Knowledge Discovery, New York, NY, USA, 18 November 2024. [Google Scholar]
  38. Saxena, A.; Chakrabarti, S.; Talukdar, P. Question Answering Over Temporal Knowledge Graphs. In Proceedings of the 59th Annual Meeting of the Association for Computational Linguistics and the 11th International Joint Conference on Natural Language Processing (Volume 1: Long Papers), Bangkok, Thailand, 1–6 August 2021; Association for Computational Linguistics: Online, 2021; pp. 6663–6676. [Google Scholar]
  39. Long, S.; Liao, J.; Yang, S.; Zhao, X.; Lin, X. Complex Question Answering Over Temporal Knowledge Graphs; Chbeir, R., Huang, H., Silvestri, F., Manolopoulos, Y., Zhang, Y., Eds.; Springer International Publishing: Cham, Switzerland, 2022; Volume 13724, pp. 65–80. [Google Scholar]
  40. Kumar, S.; Zhang, X.; Leskovec, J. Predicting Dynamic Embedding Trajectory in Temporal Interaction Networks. In Proceedings of the 25th ACM SIGKDD International Conference on Knowledge Discovery & Data Mining, Anchorage, AK, USA, 4–8 August 2019; pp. 1269–1278. [Google Scholar] [CrossRef]
  41. Han, B.; Qu, T.; Tong, X.; Wang, H.; Liu, H.; Huo, Y.; Cheng, C. AugGKG: A Grid-Augmented Geographic Knowledge Graph Representation and Spatio-Temporal Query Model. Int. J. Digit. Earth 2023, 16, 4934–4957. [Google Scholar] [CrossRef]
  42. Wang, S.; Qiu, P.; Zhu, Y.; Yang, J.; Peng, P.; Bai, Y.; Li, G.; Dai, X.; Qi, Y. Review, Framework, and Future Perspectives of Geographic Knowledge Graph (GeoKG) Quality Assessment. Geo-Spat. Inf. Sci. 2024, 27, 1–21. [Google Scholar] [CrossRef]
  43. Zheng, K.; Xie, M.H.; Zhang, J.B.; Xie, J.; Xia, S.H. A Knowledge Representation Model Based on the Geographic Spatio-temporal Process. Int. J. Geogr. Inf. Sci. 2022, 36, 674–691. [Google Scholar] [CrossRef]
  44. He, Z.; Zhao, P.; Xiao, Z.; Huang, X.; Li, Z.; Kang, T. Exploring the Distance Decay in Port Hinterlands under Port Regionalization Using Truck GPS Data. Transp. Res. Part E Logist. Transp. Rev. 2024, 181, 103390. [Google Scholar] [CrossRef]
  45. Guillaume, G.; Gauvreau, B.; L’Hermite, P. Numerical Study of the Impact of Vegetation Coverings on Sound Levels and Time Decays in a Canyon Street Model. Sci. Total Environ. 2015, 502, 22–30. [Google Scholar] [CrossRef]
  46. Jain, G.; Mahara, T.; Sharma, S.C.; Agarwal, S.; Kim, H. TD-DNN: A Time Decay-Based Deep Neural Network for Recommendation System. Appl. Sci. 2022, 12, 6398. [Google Scholar] [CrossRef]
  47. Verma, R.; Ukkusuri, S.V. What Determines Travel Time and Distance Decay in Spatial Interaction and Accessibility? J. Transp. Geogr. 2025, 122, 104061. [Google Scholar] [CrossRef]
  48. Vaswani, A.; Shazeer, N.M.; Parmar, N.; Uszkoreit, J.; Jones, L.; Gomez, A.N.; Kaiser, L.; Polosukhin, I. Attention Is All You Need. In Proceedings of the 31st Conference on Neural Information Processing Systems (NIPS 2017), Long Beach, CA, USA, 12 June 2017. [Google Scholar]
  49. Hamilton, W.L.; Bajaj, P.; Zitnik, M.; Jurafsky, D.; Leskovec, J. Embedding Logical Queries on Knowledge Graphs. In Proceedings of the 32nd International Conference on Neural Information Processing Systems, Red Hook, NY, USA, 1 June 2018. [Google Scholar]
  50. Mai, G.; Janowicz, K.; Yan, B.; Zhu, R.; Cai, L.; Lao, N. Contextual Graph Attention for Answering Logical Queries over Incomplete Knowledge Graphs. In Proceedings of the 10th International Conference on Knowledge Capture, Marina del Rey, CA, USA, 19–21 November 2019; ACM: Marina Del Rey, CA, USA, 2019; pp. 171–178. [Google Scholar]
Figure 1. Spatial analysis of the GDELT dataset. (a) Global distribution of event locations. (b) Global distribution of event frequencies by country.
Figure 1. Spatial analysis of the GDELT dataset. (a) Global distribution of event locations. (b) Global distribution of event frequencies by country.
Ijgi 14 00295 g001
Figure 2. Datasets example. (a) HAD example. (b) GDELT example. (c) ICEWS example.
Figure 2. Datasets example. (a) HAD example. (b) GDELT example. (c) ICEWS example.
Ijgi 14 00295 g002
Figure 3. ST-EKA structure. represents matrix multiplication, represents matrix addition.
Figure 3. ST-EKA structure. represents matrix multiplication, represents matrix addition.
Ijgi 14 00295 g003
Figure 4. Structure of the entity encoder.
Figure 4. Structure of the entity encoder.
Ijgi 14 00295 g004
Figure 5. Structure of the time encoder.
Figure 5. Structure of the time encoder.
Ijgi 14 00295 g005
Figure 6. Structure of the spatial encoder.
Figure 6. Structure of the spatial encoder.
Ijgi 14 00295 g006
Figure 7. Structure of the spatio-temporal weight aggregation module.
Figure 7. Structure of the spatio-temporal weight aggregation module.
Ijgi 14 00295 g007
Figure 8. Structure of the evolutionary knowledge encoder.
Figure 8. Structure of the evolutionary knowledge encoder.
Ijgi 14 00295 g008
Figure 9. Basic query graphs. (a) is a chain structure, and (b) is an intersection structure.
Figure 9. Basic query graphs. (a) is a chain structure, and (b) is an intersection structure.
Ijgi 14 00295 g009
Figure 10. Intersection operator. (a) The intersection operator of GQE model. (b) The intersection operator of CGA model.
Figure 10. Intersection operator. (a) The intersection operator of GQE model. (b) The intersection operator of CGA model.
Ijgi 14 00295 g010
Figure 11. Query subgraph model. Black dots represent entities in the knowledge graph, while red dots and red font represent answer entities in the question-answering system.
Figure 11. Query subgraph model. Black dots represent entities in the knowledge graph, while red dots and red font represent answer entities in the question-answering system.
Ijgi 14 00295 g011
Figure 12. Changes of AUC between ST-EKA optimized model and base model. (a) A U C using GDELT dataset. (b) A U C using HAD dataset. (c) A U C using ICEWS dataset.
Figure 12. Changes of AUC between ST-EKA optimized model and base model. (a) A U C using GDELT dataset. (b) A U C using HAD dataset. (c) A U C using ICEWS dataset.
Ijgi 14 00295 g012
Table 1. Dataset statistics of GDELT, ICEWS, and HAD.
Table 1. Dataset statistics of GDELT, ICEWS, and HAD.
HADGDELTICEWS
TrainingValidationTestingTrainingValidationTestingTrainingValidationTesting
Triple195,025 2167 19,503 302,553 3362 20,255 163,314 1815 16,331
Relation1 --238 --230 --
Entity6454 --6385 --7128 --
Table 2. Hyper parameter settings.
Table 2. Hyper parameter settings.
ParameterST-EKAGQECGA
Entity embedding dimension646464
Time encoderFFNFFNFFN
Spatial encoderLALALA
Time embedding dimension646464
Learning rate0.0010.0010.001
Activation functionLeakyReLULeakyReLULeakyReLU
Time activation functionSigmoidSigmoidSigmoid
OptimizerAdamAdamAdam
Batch size204820482048
Max training epochs100,000100,000100,000
Tolerance threshold1.00 × 10−61.00 × 10−61.00 × 10−6
Decoder-BilinearBilinear
Intersection operator-MinimumMinimum
(self-attention)
Table 3. The results of ST-EKA and SOTA methods.
Table 3. The results of ST-EKA and SOTA methods.
DatasetTransETTransETeRoHyTETeASTST-EKA
AUCAPRAUCAPRAUCAPRAUCAPRAUCAPRAUCAPR
GDELT0.90251891.47220.94223094.24450.92942492.53420.94412494.83940.92186893.87620.94983395.3718
HAD0.64190267.03040.67130772.93730.60802261.15510.66995669.63700.67293866.20460.78216076.7154
ICEWS0.83045785.39530.90259690.42740.92426491.84890.93313492.93040.91662591.94830.94760294.5066
Table 4. Ablation results on GDELT, HAD, and ICEWS datasets.
Table 4. Ablation results on GDELT, HAD, and ICEWS datasets.
VariantGDELTHADICEWS
AUCAPRAUCAPRAUCAPR
w/o TimeDecay0.93122793.92010.60140760.61200.91268290.8126
w/o SpatialDecay0.94020394.18620.60722660.19730.90250990.2744
w/o Evolutionary Knowledge Encoder0.92912492.09580.45519545.87460.90794889.7972
Number of adjacent quadruples20.93794294.11320.70687770.96040.93517693.4767
40.93823594.22960.74848372.94110.93679993.9902
60.94983395.37180.78216076.71540.94760294.5066
80.94773594.53330.80543881.18960.95026294.6585
Table 5. QA statistics of GDELT, HAD, and ICEWS.
Table 5. QA statistics of GDELT, HAD, and ICEWS.
QG
Type
GDELTHADICEWS
TrainingValidationTestingTrainingValidationTestingTrainingValidationTesting
2-inter496,6038989000427,235 726 10,000 451,184 861 9000
2-chain503,3978999000572,765 778 10,000 548,816 866 9000
3-chain_inter248,4379009000235,880 861 10,000 196,861 895 9000
3-inter_chain250,0469009000241,778 834 10,000 256,982 897 9000
3-inter249,7538999000215,893 930 10,000 256,787 890 9000
3-chain251,7648999000306,449 632 10,000 289,370 896 9000
Table 6. Experimental results on GDELT dataset.
Table 6. Experimental results on GDELT dataset.
QG TypeGQEGQE(ST-EKA)CGACGA(ST-EKA)
AUCAPRAUCAPRAUCAPRAUCAPR
2-chain0.843350 87.6403 0.882130 88.5873 0.847743 87.9664 0.89199185.5697
2-inter0.938444 93.9272 0.940005 93.5556 0.850209 82.2025 0.85542982.4082
3-chain0.804374 83.8110 0.851808 85.8270 0.811997 84.0545 0.87135783.2847
3-inter0.938924 94.3070 0.939747 94.0133 0.799320 78.1630 0.85385982.2146
3-inter_chain0.919941 90.9682 0.922091 90.0921 0.849684 81.8414 0.89236284.9798
3-chain_inter0.948497 95.0101 0.949520 94.6802 0.944780 94.7102 0.94783394.3688
Macro averaged0.898922 90.9440 0.914217 91.1259 0.850622 84.8230 0.885472 85.4710
Table 7. Experimental results on HAD dataset.
Table 7. Experimental results on HAD dataset.
QG TypeGQEGQE(ST-EKA)CGACGA(ST-EKA)
AUCAPRAUCAPRAUCAPRAUCAPR
2-chain0.96160496.89740.96893397.25180.96548996.59490.97724698.0323
2-inter0.99308799.58150.99918999.62480.99591999.72300.99841599.6852
3-chain0.95503195.83420.96315295.90830.95244595.91010.96410196.7924
3-inter0.99519399.69410.99847099.82550.99767399.85600.99814699.8247
3-inter_chain0.99200199.29880.99794399.54140.99574899.61450.99694899.5857
3-chain_inter0.99095299.18250.99565999.29190.99128299.28000.99512299.3061
Macro averaged0.98131198.41480.987224 98.5740 0.98309398.49640.988330 98.8711
Table 8. Experimental results on ICEWS dataset.
Table 8. Experimental results on ICEWS dataset.
QG TypeGQEGQE(ST-EKA)CGACGA(ST-EKA)
AUCAPRAUCAPRAUCAPRAUCAPR
2-chain0.918172 91.2901 0.93704193.19390.917873 91.5223 0.93949593.5107
2-inter0.974377 97.4357 0.98267497.85180.979914 98.3056 0.98322898.3321
3-chain0.886607 88.1777 0.91670289.83760.886998 88.0748 0.92058788.8951
3-inter0.981251 98.5451 0.98519198.59420.986658 98.7870 0.98885098.9267
3-inter_chain0.965495 96.3394 0.97573996.76990.964760 96.0015 0.97744897.3817
3-chain_inter0.962303 96.6072 0.96586896.84570.957734 96.0999 0.96612896.1264
Macro averaged0.948034 94.7325 0.960536 95.5155 0.948990 94.7985 0.962623 95.5288
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Zhang, C.; Chu, C.; Zhou, K.; Wang, S.; Zhu, Y.; Huang, J.; Wu, Z.; Gao, F. A Spatio-Temporal Evolutionary Embedding Approach for Geographic Knowledge Graph Question Answering. ISPRS Int. J. Geo-Inf. 2025, 14, 295. https://doi.org/10.3390/ijgi14080295

AMA Style

Zhang C, Chu C, Zhou K, Wang S, Zhu Y, Huang J, Wu Z, Gao F. A Spatio-Temporal Evolutionary Embedding Approach for Geographic Knowledge Graph Question Answering. ISPRS International Journal of Geo-Information. 2025; 14(8):295. https://doi.org/10.3390/ijgi14080295

Chicago/Turabian Style

Zhang, Chunju, Chaoqun Chu, Kang Zhou, Shu Wang, Yunqiang Zhu, Jianwei Huang, Zhaofu Wu, and Fei Gao. 2025. "A Spatio-Temporal Evolutionary Embedding Approach for Geographic Knowledge Graph Question Answering" ISPRS International Journal of Geo-Information 14, no. 8: 295. https://doi.org/10.3390/ijgi14080295

APA Style

Zhang, C., Chu, C., Zhou, K., Wang, S., Zhu, Y., Huang, J., Wu, Z., & Gao, F. (2025). A Spatio-Temporal Evolutionary Embedding Approach for Geographic Knowledge Graph Question Answering. ISPRS International Journal of Geo-Information, 14(8), 295. https://doi.org/10.3390/ijgi14080295

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop