Research on the Construction of a Knowledge Graph and Knowledge Reasoning Model in the Field of Urban Trafﬁc

: The integration of multi-source transportation data is complex and insufﬁcient in most of the big cities, which made it difﬁcult for researchers to conduct in-depth data mining to improve the policy or the management. In order to solve this problem, a top-down approach is used to construct a knowledge graph of urban trafﬁc system in this paper. First, the model layer of the knowledge graph was used to realize the reuse and sharing of knowledge. Furthermore, the model layer then was stored in the graph database Neo4j. Second, the representation learning based knowledge reasoning model was adopted to implement knowledge completion and improve the knowledge graph. Finally, the proposed method was validated with an urban trafﬁc data set and the results showed that the model could be used to mine the implicit relationship between trafﬁc entities and discover trafﬁc knowledge effectively.


Background
With the rapid development of information technology, such as Big Data, cloud computing, artificial intelligence, and the Internet of Things, various terminals have generated massive traffic data, such as mobile phone signaling and public traffic card swiping data containing travel records and vehicle information obtained by video surveillance equipment on the road. Traffic Big Data have the characteristics of large quantity, variety, wide coverage, fragmentation, and so on [1].
A knowledge graph describes the various concepts, entities, and relationships between entities in the objective world in a structured way, and provides a better ability to organize, manage, and understand massive amounts of information [2]. There are already some vertical knowledge graphs for specific fields, but data sparseness still exists. Therefore, the kind of reasoning method to be used to solve the problem of completeness of a knowledge graph is the current research hotspot in the field of knowledge graphs.
The manner in which to determine whether distributed storage and management of traffic data and data with different semantics are associated, to fully explore the value of data in the traffic field, and to promote the efficient use of information resources is key to promoting the construction of smart traffic, as well as the query and retrieval of traffic information resources, traffic guidance systems, intelligent scheduling management, and other applications.
It is of great significance to improve the level of intelligent management in the traffic field and to assist managers in decision-making analysis. Therefore, research on the construction of knowledge maps in the field of traffic has significant practical value. concept analysis (FCA) is used to generate an axiom hypothesis of missing ontology from the collected facts [24]. However, as the scale of a knowledge graph becomes larger and the structure becomes more complicated, theses methods are inefficient and costly. Therefore, some scholars have proposed a representation learning method to embed the entities and relationships in a knowledge graph into a low-dimensional vector space. This representation can reflect the semantic information of entities and relationships, and it can efficiently calculate complex semantic associations between entities [25]. There are many methods for representation learning when solving low-dimensional vector representations. The typical methods include structured embedding representation models, tensor neural network models, matrix decomposition methods, and translation models.
The structured embedding (SE) proposed by Bordes et al. projects the head and tail entities into the space of r through two matrices of the relationship r, and calculates the distance between the two projection vectors in space. A smaller distance between entities indicates that this relationship exists between said entities [26]. This model uses different matrices to project the head and tail entities, so the coordination is poor. Socher et al. proposed the neural tensor network (NTN), which replaces the traditional neural network layer with a bilinear tensor layer and connects the head and tail entities in different dimensions to describe a complex semantic connection [27]. However, this model has high computational complexity and poor performance on large-scale sparse knowledge graphs. The RESACL model proposed by Nickel et al. is a representative model of the matrix decomposition method, which aims to decompose the tensor value X hrt corresponding to each triplet (h, r, t) into entity and relationship representations, making it as close as possible to l h M r l t [28].
The word2vec word representation learning method proposed by Mikolov et al. found that there is a translation invariance phenomenon in the word vector space [29]. Bordes et al. [30] were inspired by this and proposed the first translation model, TransE. The entities and relationships in this model are represented by a single vector, which cannot accurately describe complex relationships such as reflexivity, 1-N, N-1, and N-N. Therefore, many models have been improved and extended on the basis of this model. For example, the TransH model [31] proposes to project the head entity h and the tail entity t into the corresponding hyperplane of the relationship r. The translation operation better solves the problem of TransE's poor effect in dealing with complex relationships, but it does not break the assumption that entities and relationships are in the same space. Therefore, the TransR model proposes to project the head entity h and the tail entity t into the corresponding relationship space of the relationship r, and to realize the translation operation from the head entity to the tail entity in the relationship space [32]. The CTransR model is based on the TransR model. First, the head and tail entity pairs are grouped by clustering, and different relationship vectors are learned for each group to make the projection expression more accurate. However, due to the introduction of the projection matrix, the model parameters have increased sharply, and the computational complexity has also greatly increased. Therefore, Ji et al. proposed a method of dynamically constructing a projection matrix-the TransD model [33], which takes into account the interaction between entities and relationships and defines the projection matrix of the head and the entities in the relationship space, respectively.
At present, most traffic knowledge graph research focuses on the detection and prediction of road congestion. However, a comparatively small amount of work focuses on discovering the implicit relationship between traffic entities. Additionally, research about knowledge reasoning methods based on representation learning mainly focus on general knowledge graphs. There are few papers that use the representation learning model to perform knowledge reasoning on constructed traffic knowledge graphs. Therefore, this paper used the representation learning model (TransD) to perform knowledge reasoning using the existing knowledge in knowledge graphs, which can discover the implicit relationship between traffic entities, such as the relationship between the points of interest and the road traffic state.

Overall Framework
The construction process of a knowledge graph is essentially a process of obtaining the required data and organizing said data into a whole in an appropriate form and method. The construction process of a knowledge graph is an iterative update process. First, it is necessary to rely on crawler technology to collect open traffic field data and to extract entities, attributes, and relationships from it. Second, the pattern layer of a knowledge graph is designed and completed, that is, the traffic field ontology is constructed. Then, the data layer of a knowledge graph is stored in the graph database. After that, knowledge reasoning is performed based on the existing data in a knowledge graph,carrying out quality evaluation of the inference results (such as contradiction and redundancy checks), thereby expanding and enriching a knowledge graph. The research framework of this paper is shown in Figure 1.

Data Collection and Preprocessing
The multi-source public traffic data obtained in this paper included the AFC card swiping data of the metro in Shenzhen and the static basic data of metro lines and stations. In addition, the urban road traffic data included road data, points of interest data, and road traffic situation data in the area within the Fifth Ring Road in Beijing.
(1) Metro static basic data This paper used Internet crawler technology to obtain Shenzhen's metro station and line data from the AutoNavi Map API [34]. Web crawlers use uniform resource locator URLs to automatically grab and download target information from websites, which improves efficiency and saves time [35].
Step 1: First, the incoming parameters need to be constructed, which mainly include the key value, city code, and city name. After the parameters are URL-encoded, a request is initiated to the target HTTP interface, that is, a request is sent.
Step 2: The response returned by the HTTP request is received, and the returned json format data is parsed. Finally, the parsed data are stored in the relational database PostgreSQL, and the stored data are shown in Table 1. The Shenzhen metro adopts a one-ticket system, and passengers must swipe their cards to enter and exit the station. The time range of the Shenzhen pass card swiping data in this paper were from 25 January 2016 (Monday) to 29 January 2016 (Friday), a total of five working days of card records. The data format is shown in Table 2. Step 1: The road data in the area within the Fifth Ring Road of Beijing are downloaded from the OpenStreetMap database.
Step 2: OSM2GMNS is used to output the road network data conforming to the General Modeling Network Specification standard from the downloaded target road network. The output file includes road network nodes (node.csv) and a road network connection link (link.csv). The main field descriptions are shown in Tables 3 and 4.  The Well-Known Text (WKT) representation of Geometry is designed to exchange geometry data in ASCII form.

(4) Points of interest data
A point of interest (POI) generally refers to a meaningful non-geographical point in the real world, such as the infrastructure closely related to people's lives [36]. This paper used the search POI interface in the AutoNavi Map API to obtain information about primary and secondary schools in the area within the Fifth Ring Road of Beijing. The specific steps were as follows: Step1: First, the query area is defined, using QGIS to draw the area boundary. Because each request returns up to 1000 POIs, it is necessary to divide the large area into multiple small grids to obtain the vertex coordinates of each small grid.
Step2: The POI information in each grid are obtained in turn through the web crawler, and then the returned json data are parsed and stored in the database. A description of the POI information fields is shown in Table 5: The time range for collecting the traffic situation in this paper was a week of online classes for students in 2020: 24 August (Monday) to 28 August (Friday); one week that students were in school: 21 September (Monday) to 25 September (Friday); a daily collection period from 6:30 a.m. to 10:30 a.m. and from 4:00 p.m. to 9:00 p.m. The method of collecting traffic situation data was the same as the method for searching for POI. Table 6 shows the fields of the road traffic situation: The collected data were preprocessed, the target data were extracted, and invalid and wrong data were eliminated, thereby improving data quality. For metro card swiping data, the card swiping records during the morning peak (7:00 a.m. to 9:00 a.m.) were filtered out, the swipe records that did not appear in pairs when entering and exiting the station were eliminated, and the data with the same entry and exit positions were deleted, with the time difference between entry and exit being greater than 5 h of data.
Web crawler technology was used to obtain the real-time traffic situation data of roads. The data on the AutoNavi Map API are updated every 2 min. However, due to the instability of the network connection, the time interval for returning traffic situation data is different. In order to facilitate subsequent analysis, the data needed to be changed. The sampling intervals was uniform every 5 min. Because the road classes returned by the traffic situation were mostly expressways and trunk roads, the road network data with traffic situation data were filtered according to the matching of road names, and then the invalid and redundant data were eliminated, thereby improving data quality.

Ontology Construction of Urban Traffic
The construction of a knowledge graph in the field of urban traffic adopted a top-down approach. First, the pattern layer of the knowledge graph was defined through the construction of domain ontology. After that, the entities, attributes, and relationships between entities were extracted from various types of data sources, and finally, the graph database Neo4j was used to store the data of the knowledge graph.
The construction of the ontology of the urban traffic field was oriented to specific urban traffic business scenarios. Moreover, the standardization of field terms and the wide applicability of concept categories, as well as the hierarchical structure of the concepts in the abstract field, were considered, and the related attributes of each concept and the relationship between concepts were defined [37]. This paper used the seven-step method published by Stanford University to manually construct the domain ontology [38], the domain ontology with the four elements of "people-vehicle-road-environment" in traffic as the core, and related to traffic subjects, travel behavior, traffic facilities, traffic tools, and other entities.
(1) Define the class and its hierarchical structure The field of urban traffic was divided into six categories, including traffic tools, traffic buildings, traffic service facilities, traffic participants, traffic state evaluation, and traffic attraction points. Each category was divided into multiple subcategories and organized in a hierarchical structure. The hierarchy of the ontology classes in the urban traffic field is shown in Figure 2.
(2) Define the attributes of the class The attributes of each class were defined; the classes had inheritance, meaning the subclasses inherited the attributes of their parent class, so the attributes were placed in the broadest class, that is, the closer to the top level, the better. Table 7 shows a list the attributes of the six main categories in this paper.

(3) Define the relationship between classes
There are some relationships in the urban traffic ontology, and two classes are connected through such a relationship. For example, the location of a certain point of interest is on a certain road, and the relationship between the point of interest and the city road is located. Figure 3 shows the relationships between all of the classes in the urban traffic ontology.
This paper used the ontology editing tool Protégé to complete the construction of the ontology in the urban traffic field, as shown in Figure 4. Protégé is an ontology editing and knowledge acquisition software developed by Stanford University based on the Java language. It provides good support in terms of visualization, query, and storage [39].    Ontology defines the mold of a knowledge graph and describes the top-level structure of said knowledge graph. Therefore, constructing ontology can analyze the system and level of domain knowledge and can realize the repeated use of domain knowledge.

Storage of an Urban Traffic Knowledge Graph
The graph database Neo4j supports massive data storage, which can solve well the problems of low value density, large quantity, and fast update speed of data in the traffic field. Additionally, the Cypher graph query language supports associated query and graph algorithms, which is more conducive to data query and value mining. Therefore, this paper chose the Neo4j graph database for knowledge storage.
Because the time and space of metro swipe card data and traffic situation data are different, the classes in the field of urban traffic ontology were divided into two: One was a public traffic knowledge graph based on swipe card data, and the other was an urban road traffic knowledge graph based on road traffic situation data.
(1) The public traffic knowledge graph According to the subway metro swipe card data and the subway static basic data in the collected urban traffic data, the entities, attributes, and relationships between entities were extracted to construct the data layer of the public traffic knowledge graph, and the data were stored in the graph database. The attributes of the entities and relationships are described in Table 8. Taking the relationship between the metro passengers and itinerary as an example, the same public traffic travel user has multiple itineraries in a week, so there is a one-to-many relationship between users and itineraries. There is a beginning or ending relationship between the itinerary and the metro station, as shown in Figure 5. (2) The urban road traffic knowledge graph According to the acquired road data, points of interest information (primary and middle schools), and traffic situation data of the area within the Fifth Ring Road in Beijing, the entities, attributes, and relationships between entities were extracted to construct a knowledge graph of urban road traffic. The entities included urban roads, points of interest, and traffic situations in the domain ontology, and entities representing spatio-temporal relationship data, such as date, time, intersection, and road section, were added. Figure 6 shows the entities included in the urban road traffic knowledge graph and the relationships between the entities.  In Neo4j graph data, both nodes and relationships can set attributes and store them in the form of key-value pairs [40]. The attributes as shown in Table 9.
The road section was connected to the road intersection and was divided according to the driving direction of the vehicle on the road. The road section had starting and ending point intersections. The relationship between the road section and the intersection is shown in Figure 7. The intersection and road sections constitute the topological structure of the urban road network.

Knowledge Reasoning Model Based on Representation Learning
The main idea of knowledge reasoning based on representation learning is to transform the semantic information of entities and relationships in a knowledge graph into dense low-dimensional vectors, to map them into the vector space, and then to participate in calculations to complete tasks such as entity linking and relationship reasoning. Knowledge representation learning can effectively improve computing efficiency and can reduce the impact of data sparseness on the model inference results.

Model Definition
This paper used the TransD model. The main idea was to use the dynamic mapping matrix constructed by the projection vector to encode the entity as a low-dimensional embedding vector in the relational space. At the same time, it was considered that entities and relationships have different types and attributes, so different types of relationships define different semantic spaces, and different attributes focus on different entities under different relationships, as shown in Figure 8. The TransD model uses two vectors to represent each entity and relationship. The first vector represents the actual meaning of the entity or relationship, and the second vector is called the projection vector and is used to construct the mapping matrix. The mapping matrix is jointly determined by the projection vector of the entity and the relationship. The entity can be mapped from the entity space to the vector space, and each mapping matrix is initialized with the identity matrix, and the vector operation is used to replace the matrix multiplication operation, which effectively reduces the amount of calculation.
The entity vector is projected into the relational space and embedded as: Inspired by word2vec's word vector translation invariance phenomenon, TransD regards the relationship vector r in the triple (h, r, t) as the translation vector of the head entity vector h and the tail entity vector t. That is, the sum of the embedding vector of the head entity h ⊥ and the embedding vector of the relationship r in the embedding space are approximately equal to the embedding vector of the tail entity t ⊥ . Therefore, a score function based on L2 Euclidean distance can be defined to measure the distance between these two vectors: The model adds the L2 norm constraint to the vector, which can make the related parameters of the model smaller, can avoid over-fitting of the model, and can improve the generalization ability of the model. We enforced constrains as: For a correct triplet, the higher the score expected, the better; meanwhile, for a wrong triple, the smaller the score, the better. Thus, a margin-based ranking loss function was defined, such as shown in Equation (7), to minimize the loss function value as the training target of the model.
where γ is a hyperparameter, which represents the maximum interval between the correct triplet and the negative triplet, [x] + = max(0, x), represents the set of correct triples, and represents the set of constructed negative triples.

Construction Method of Negative Sample
Since there are only correct triples in a knowledge graph, wrong triples need to be constructed as negative samples. The method used by the TransE model is to randomly select an entity from the set of all entities for a correct triple to replace the head or tail entity of the original triple and to obtain a new wrong triple, that is, a negative triple. However, due to the existence of the one-to-many, many-to-one, and many-to-many types of relationships, this random sampling method of constructing negative samples introduces many false-negative samples, that is, false negatives.
Taking into account the different types of relationships, when replacing the head or tail entity in the triple, the one with a smaller number is replaced with a greater probability. For example, when the relationship is one-to-many, that is, one head entity corresponds to multiple tail entities, the head entity is replaced with a greater probability. When the relationship is many-to-one, that is, multiple head entities correspond to one tail entity, the tail entity is replaced with a greater probability. For all relationship triples, the following numbers are counted: • The average number of tail entities associated with each head entity is recorded as tph; • the average number of head entities associated with each tail entity is recorded as hpt. X = 1, replace head entity 0, replace tail entity (8) The random variable X only takes the two values of 0 and 1, and the corresponding probability is: The construction of the final negative sample obeys the Bernoulli distribution with parameter p, and the distribution law of the random variable X is: For a certain correct triple of a given relationship, the probability of generating a negative triple by replacing the head entity is p, and the probability of generating a negative triple by replacing the tail entity is 1 − p.
In addition, the type of constraint of the relationship is expressed by defining the type of entity that the relationship should be associated with. Using prior knowledge of the relationship type, the relationship determines which entities to replace. For example, the definition of the relationship type marriage is only associated with a person [41]. The following variables are to be defined: • The ordered index of all entities within the domain constraint of the relationship type domain r ; • The ordered index of all entities within the range constraint of the relationship type range r .
For a triple of a given relationship r, when constructing a negative triple, the probability of replacing the head or tail entity is calculated according to the Bernoulli distribution. If the head entity is replaced, it is selected from the entity subset in the domain of the relationship type, and if the tail entity is replaced, it is selected from the entity subset in the range of the relationship type, as shown in Equation (12).
Using prior knowledge of the relationship type, the probability of extracting the correct type of entity to replace the original triple when constructing a negative sample is improved.

Model Training Process
The model uses mini-batch gradient descent (MBGD) to update the parameters and to obtain the minimum value of the loss function. In the training process, small batches of positive samples are randomly selected, and their corresponding negative samples are constructed. After a batch, the gradient is calculated and the model parameters are updated. The model updates the parameters through continuous iterations until the loss value converges or reaches the maximum number of iterations. After the training has been completed, the embedding representation of entities and relationships is obtained. Algorithm 1 is the pseudocode for the algorithm training: Algorithm 1 Reasoning algorithm training process based on representation learning. Input: Training sample set S, total number of samples N, entity set E, relationship set R, learning rate α, embedding dimension k, boundary γ, The maximum number of iterations M, the number of small batch samples batch_size Output: Vector representation of entities and relationships 1: /* initialize */ 2: for each r ∈ R do 3: r ← uni f orm(tensor, −k, k) 4: end for 5: for each e ∈ E do 6: e ← uni f orm(tensor, −k, k) 7: end for 8: while i < M do 9: while j < N batch_size do 10: S batch ← Sample(S, batch_size)

Experiment and Results Analysis
The reasoning performance of the model was evaluated through the link prediction of the evaluation task of knowledge completion. Link prediction refers to the task of predicting another entity that has a specific relationship with a given entity, that is, for a triple (h, r, t), a given relationship r and a tail entity t predict the head entity h, denoted as (?, r, t), or a given head entity h and relationship r predicts the tail entity t, denoted as (h, r, ?).
The data used in this paper were the data of a public traffic knowledge graph stored in the graph database. The triplet data were divided into a training set, a test set, and a validation set according to the ratio of 85%:10%:5%. The data volume of the entities, relationships, and triples contained in the data set is shown in Table 10. For each triplet in the test set, the prediction tail entity was taken as an example, and the tail entity was replaced with entities in the knowledge graph in turn. These replacement entities were determined by the relationship type, that is, the relationship type constraint was used to construct negative samples. Then, the score function was used to calculate the scores of these triples and to arrange them in descending order to count the ranking of the correct entities.
This paper used multiple evaluation indicators to measure the effect of link prediction tasks, including average rank (MR), average reciprocal rank (MRR), first hit rate (hits@1), top three hit rate (hits@3), and the top ten hit rate (hits@10). The average rank was the average rank of the correct triple in the test set. The top ten hit rate (hits@10) indicates the ratio of the number of triples in the top ten to the total number of triples in the test set. When the evaluation index average rank was smaller, the average reciprocal rank, hits@1, hits@3, and hits@10 were larger, indicating better reasoning performance.
Because the constructed negative triples may be the correct triples (false negatives) that exist in the training, validation, or test set, that is, knowledge that already exists in the knowledge graph but is misjudged as a wrong triple, it is necessary to delete such triples before sorting to improve the accuracy of model prediction; this operation is called filtering.
Comparing the TransD model to the classic reasoning models TransE and TransH, both filtering and relationship type constraints were used. The evaluation indicators defined in this paper were used to compare the results of the different models in the entity link prediction task. The optimal hyperparameter settings of the different models are shown in Table 11. Among them, the adadelta optimizer was used during TransD model training, which can adaptively adjust the learning rate [42]. Regarding the public traffic knowledge graph data set constructed in this paper, the results of the different models on the link prediction task were compared. Table 12 shows the results of predicting the head and tail entities. In order to facilitate a comparison, the optimal value of each evaluation index is marked in bold, and a bar chart was drawn with the MRR and hits@10 indicators as representatives, as shown in Figure 9, which can intuitively compare the performance of each model.
By comparing the prediction results of the different models, it can be found that the TransD model achieved the best effect, and it performed better than the other two reasoning models on all evaluation indicators. When predicting the head entity, the average reciprocal ranking (MRR) of the TransD model increased by 6.6% compared to the average ranking of the TransH model. When predicting the tail entity, the indicator increased by 22.4%.
For the urban road traffic travel knowledge graph, whether or not the reasoning model can complement the speed value of the road at a certain moment was verified. By comparing the average road speed before and after school starts, it was found that the average road speed decreased by approximately half an hour in the morning rush hours after school starts, as shown in Figure 10. Table 13 shows the top five entities in predicting the tail entity under the relationship between time and the traffic situation. Even if the correct triples were not always the top ones, these prediction results are consistent with common sense. At the same time, the speed value before the beginning of school was greater than the speed value after the beginning of school, which also shows that students going to school affects the traffic situation on the road.    The bold value represents the correct entitye.

Knowledge Discovery Based on the Knowledge Graph
(1) Relationship path discovery Based on the multi-depth relationship node query of the graph database Neo4J, all entities that have an association with the target entity within the specified path length range can be found, or hidden relationships between entities can be discovered. For example, there was an associated path between the point of interest and the traffic situation of the road, as shown in Figure 11. (2) Similar travelers found This paper used the Jaccard coefficient to compare the similarity of travelers, and measured the similarity of their trips by calculating the ratio of the same part of the entity to the different part of the commuter's travel chain [43]. The similarity between the travel chains of commuters A and B was calculated as per Equation (13), and the calculation results are shown in Figure 12. From the results, it can be seen that the similarity between commuters who have the same start and end points and the same departure and arrival time is the highest. (3) Shortest path query The urban road traffic knowledge graph constructed in this paper included the topological structure of the road network and the length information of the road sections. Thus, it can use the Dijkstra algorithm in the path search algorithm to find the shortest distance path between two nodes, and can provide a data basis for route planning services of traffic. As shown in Figure 13, where the intersection numbered 1 was taken as the starting point and the intersection numbered 9 as the end point, the path with the shortest distance between the start and end points was found and is marked with a red line in the figure.

Conclusions
This paper studied the construction method of a knowledge graph in the field of urban traffic. Using a top-down construction method, we first analyzed the knowledge system in the field of urban traffic, and then designed the model layer of the knowledge graph of traffic. Afterward, the extracted entities, attributes, and relationships between entities were stored in the graph database Neo4j, and finally, the construction of a public traffic knowledge graph and an urban road traffic travel knowledge graph was completed. Then, a link prediction task was performed on the traffic data set, and the effectiveness of the reasoning model used in this paper was verified by comparing the results to the evaluation indicators of other models. Finally, it was shown that knowledge discovery can be realized based on the constructed knowledge graph of the urban traffic field.
Our contributions can be summarized as following: (1) The urban traffic knowledge graph constructed in this paper aims to discover potential relationship between different traffic entities, such as discovering traffic entities related to road congestion. Compared to the congestion detection of other knowledge graphs, it can more effectively assist managers in formulating strategies to alleviate road congestion. (2) The problem that the conclusions of traditional transportation research cannot be widely promoted can be solved by the knowledge graph. The knowledge contained in the knowledge graph is universal; a set of traffic knowledge systems that can be shared and reused was formed in our paper. Moreover, with the accumulation of relevant data, the new knowledge obtained through reasoning can optimize and enrich the original knowledge graph. (3) Based on the constructed urban traffic knowledge graph, it is possible to realize traffic knowledge discovery and intelligent question answering of urban traffic services, such as similar traveler discovery and the shortest path query.
Because the construction method of a knowledge graph, such as the entities included and the relationships between the entities, affects the efficiency and accuracy of the reasoning model, in the future, we will study ways to improve the accuracy of the reasoning model by improving the construction of a knowledge graph. However, the knowledge graph we constructed herein and the research were only based on a small scene in urban traffic. In the future, we will add more traffic-related data, study the relationship between weather and traffic accidents, and explore the factors that influence the choice of the travel modes of traffic participants.  Informed Consent Statement: Informed consent was obtained from all of the subjects involved in this study.

Data Availability Statement:
The data presented in this study are available on request from the corresponding author.

Conflicts of Interest:
The authors declare no conflicts of interest.