Next Article in Journal
An Interferometric Multi-Sensor Absolute Distance Measurement System for Use in Harsh Environments
Next Article in Special Issue
Impact of Lower-Limb Muscle Fatigue on Dynamic Postural Control During Stair Descent: A Study Using Stair-Embedded Force Plates
Previous Article in Journal
MineVisual: A Battery-Free Visual Perception Scheme in Coal Mine
Previous Article in Special Issue
A Comprehensive Review of Hardware Acceleration Techniques and Convolutional Neural Networks for EEG Signals
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

QAMT: An LLM-Based Framework for Quality-Assured Medical Time-Series Data Generation

1
School of Computer Science and Technology, Beijing Institute of Technology, Beijing 100081, China
2
BNRist, DCST, RIIT, Tsinghua University, Beijing 100084, China
*
Author to whom correspondence should be addressed.
Sensors 2025, 25(17), 5482; https://doi.org/10.3390/s25175482
Submission received: 3 August 2025 / Revised: 28 August 2025 / Accepted: 31 August 2025 / Published: 3 September 2025
(This article belongs to the Special Issue Sensors Fusion in Digital Healthcare Applications)

Abstract

The extensive deployment of diverse sensors in hospitals has resulted in the collection of various medical time-series data. However, these real-world medical time-series data suffer from limited volume, poor data quality, and privacy concerns, resulting in performance degradation in downstream tasks, such as medical research and clinical decision-making. Existing studies provide generated medical data as a supplement or alternative to real-world data. However, medical time-series data are inherently complex, including temporal data such as laboratory measurements and static event data such as demographics and clinical outcomes, with each patient’s temporal data being influenced by their static event data. This intrinsic complexity makes the generation of high-quality medical time-series data particularly challenging. Traditional methods typically employ Generative Adversarial Networks (GANs) or Variational Autoencoders (VAEs), but these methods struggle to generate high-quality static event data of medical time-series data and often lack interpretability. Currently, large language models (LLMs) introduce new opportunities for medical data generation, but they face difficulties in generating temporal data and have challenges in specific domain generation tasks. In this study, we are the first to propose an LLM-based framework for modularly generating medical time-series data, QAMT, which generates quality-assured data and ensures the interpretability of the generation process. QAMT constructs a reliable health knowledge graph to provide medical expertise to the LLMs and designs dual modules to simultaneously generate static event data and temporal data, constituting high-quality medical time-series data. Moreover, QAMT introduces a quality assurance module to evaluate the generated data. Unlike existing methods, QAMT preserves the interpretability of the data generation process. Experimental results show that QAMT can generate higher-quality time-series medical data compared with existing methods.

1. Introduction

With the development of Internet-of-Things (IoT) technologies, hospitals are increasingly equipped with a wide array of sensors that monitor and capture various medical data. Since these sensors record data continuously (e.g., vital signs), most of the medical data exist in the form of time-series data. These data support medical research through effective exploration and mining [1] and offer valuable assistance in tasks such as health monitoring [2], clinical decision-making [3], disease diagnosis [4,5], and treatment recommendation [6].
Although medical time-series data have demonstrated substantial value in various downstream tasks, their acquisition remains a significant bottleneck. On one hand, although sensors collect data in real time, issues such as sensor malfunctions and the inherent complexity of medical time-series data often lead to challenges in data quality. As a result, medical time-series datasets commonly used, such as eICU [7] and MIMIC-III [8], suffer from missing values, inaccuracies, incompleteness, etc., ultimately resulting in poor data quality [9]. On the other hand, because medical time-series data contain highly sensitive information (e.g., patients’ age, gender), their access is typically subject to strict privacy regulations and governance controls [10], resulting in limited data volume and privacy concerns. Some studies attempt to mitigate privacy risks by removing personal information from medical data. However, even de-identified data remain vulnerable to re-identification [11,12]. Other studies also explore the use of federated learning to address this issue. However, these works still face challenges related to system deployment and security attacks [13]. Motivated by these limitations, generated medical time-series data offer a supplement and alternative to real-world data.
However, due to the complexity of medical time-series data, which encompass a wide variety of variables that interact in complex and nonlinear ways, generating such data is challenging. A critical aspect of medical time-series data is the presence of static event data (e.g., demographics, clinical outcomes), which can significantly influence temporal data (e.g., vital signs, laboratory measurements). For instance, an in-hospital patient’s temporal data, which would include hundreds of variables, such as heart rate and blood pressure, can be influenced by several static event variables, including age, comorbidities, etc. We define an effective medical time-series generation model as one that jointly models both static event data and temporal data. Consequently, generating medical time-series data presents several significant challenges:
  • Challenge 1: Joint Generation of Temporal Data and Static Event Data. Medical time-series data consist of diverse data types, each with unique characteristics. Among them, static event data are generally high-dimensional and discrete (e.g., demographics, clinical outcomes), whereas temporal data tend to be lower-dimensional and continuous (e.g., vital signs, laboratory measurements). Therefore, jointly generating medical time-series data that include both static and temporal components is essential for producing comprehensive and realistic datasets.
  • Challenge 2: Clinical Constraints and Variable Dependencies. Many variables in medical time-series data are governed by clinical constraints, and their values often exhibit strong interdependencies. For example, a patient’s systolic blood pressure value of zero is clinically impossible. Additionally, if a patient consistently exhibits systolic blood pressure readings above 150 mmHg, a final diagnosis would not be hypotension. Accurately modeling these constraints and dependencies is essential to ensure the clinical plausibility of generated data.
  • Challenge 3: Need for Interpretability. The generation process of medical time-series data should be interpretable. In clinical research and practice, it is crucial for stakeholders to understand how the data are produced in order to evaluate their quality, support downstream applications, and maintain trust in data-driven healthcare systems.
Table 1 shows that traditional medical time-series data generation methods based on deep learning techniques primarily rely on generative models, such as Variational Autoencoders (VAEs) [14] and Generative Adversarial Networks (GANs) [15]. However, although GANs and VAEs have demonstrated strong performance in generating long-sequence temporal data, they face significant challenges in simultaneously producing high-quality static event data. A few studies have attempted to address this by incorporating statistical summaries of the generated temporal data as inputs to guide the generation of static event data [16]. Nevertheless, GANs and VAEs still struggle to generate sparse one-hot encoded representations [17] with satisfactory fidelity. In addition, it is challenging to incorporate clinical constraints and variable dependencies into these models, and the black-box nature of deep learning models further limits the interpretability of the generation process. The emergence of large language models (LLMs) presents a promising new direction for medical time-series data generation [9,10]. LLMs have demonstrated strong capabilities in modeling complex distributions over discrete data [18]. However, since LLMs are primarily designed for text-based tasks, they struggle to jointly generate high-quality continuous temporal data [10]. Moreover, due to their limited exposure to domain-specific medical knowledge, generating clinically meaningful and high-quality medical time-series data remains a challenge for LLMs [19]. A recent study [20] proposes a pipeline framework for medical time-series data generation. However, this framework lacks clinical constraints and variable dependencies, as well as interpretability in the data generation process. Therefore, jointly generating high-quality medical time-series data that include temporal data and static event data with an interpretable generation process remains a significant challenge in current research.
In this study, we propose a modularized LLM-based framework for Qality-Assured Medical Time-series data generation, QAMT. To incorporate domain-specific medical knowledge into LLMs and guide both the generation and quality assurance of medical time-series data, we leverage a reliable Health Knowledge Graph Builder, HKGB [30], to build a health knowledge graph (HKG). During the medical time-series generating process, QAMT adopts dual modules, combining GANs for generating temporal data and health knowledge graph-based retrieval augmented generation for generating static event data. This design leverages the advantages of existing methods and jointly generates medical time-series data. To assess the quality of the generated medical time-series data, QAMT integrates a quality assurance module that evaluates the data through clinical constraints and inter-variable dependencies, using the capabilities of the HKGB, LLMs, and a chain-of-thought prompting over HKG. In contrast to existing methods, QAMT maintains interpretability throughout the medical time-series data generation process. The main contributions of this work are as follows:
  • QAMT jointly generates medical time-series data, which include both continuous temporal data and discrete static event data.
  • QAMT ensures the quality assurance of the generated data by accounting for real-world clinical constraints and variable dependencies.
  • QAMT enables interpretability in the medical time-series data generation process.
  • QAMT is evaluated on the eICU and MIMIC-III datasets, and demonstrates superior performance compared to state-of-the-art models in terms of fidelity, utility, and privacy.
In this paper, we outline the problem definition and review related work in Section 2. Section 3 introduces the overall architecture and workflow of the proposed QAMT framework. In Section 4, we delve into the design of the health knowledge graph (HKG) module and explain the motivation for incorporating HKG. Section 5 details the dual modules responsible for jointly generating medical time-series data, including the static event data generation module and temporal data generation module. We present our data quality assurance module, including how clinical constraints and variable dependencies are enforced in Section 6. In Section 7, we provide an explanation of the interpretability of QAMT. In Section 8, we provide experimental results to evaluate our proposed QAMT. Finally, Section 9 summarizes the key contributions of this work and discusses potential directions for future research.

2. Preliminaries and Related Work

In this chapter, Section 2.1 provides the definition of our medical time-series data generation task. Section 2.2, Section 2.3, Section 2.4 discuss the limitations of existing work in jointly generating high-quality and interpretable medical time-series data.

2.1. Definition

A medical time-series dataset is denoted as D = { P i } i = 1 N , where N represents the number of patients, and P i denotes the medical time-series data of the ith patient. Each patient record P i consists of a sequence of clinical visits, denoted as P i = { V i j , T i j } j = 1 | V i | , and each visit V i j at time T i j is defined as a tuple V i j = { c i j , y i j , E i j , t i j } , where: c i j is a set of covariates (e.g., demographics) of this patient, y i j is a set of clinical outcomes (e.g., expire), E i j = { e m } m = 1 | E j | is a sequence of clinical events (e.g., ICD diagnosis codes), and t i j is a collection of temporal data recorded in in-patient settings, including vital signs and laboratory measurements [10]. We refer to m i j = c i j , y i j , E i j collectively as static event data (high-dimensional and discrete data), and t i j as temporal data (low-dimensional and continuous data). Each temporal data element t i j is further represented as t i j = t i m e k , ( n k p , x k p ) p = 1 | L k | k = 1 | t i j | , where | t i j | denotes the total number of measurement time points, | L k | denotes the number of variables measured at time point t i m e k , n k p is the name of the pth variable measured at time t i m e k , and x k p is the corresponding observed value. In this paper, our task is to jointly generate high-quality and interpretable medical time-series data, including both static event data and temporal data.

2.2. Medical Time-Series Data Generation

Due to the complexity of medical time-series data, as shown in Table 1, few studies have addressed the joint generation of medical time-series data.
A considerable amount of research has focused on the generation of a sequence of static event data. Traditional methods, such as GAN-based methods [20,21] and VAE-based methods [23], have been widely used. However, the discrete and high-dimensional nature of static event data makes it challenging for them to guarantee the quality of the generated data. Recently, LLMs have been explored for static event data generation [19,24]. Empirical studies have demonstrated that LLMs outperform GAN- and VAE-based methods in static event data generation [31,32], largely due to their superior capabilities in semantic reasoning and contextual understanding.
Other research on temporal data generation has primarily focused on producing data with meaningful patterns and trends. Traditional generative techniques, such as GANs [25,26,27] and VAEs [28], have demonstrated strong potential for generating temporal data. Existing studies have shown that GAN-based methods generally outperform VAE-based methods in generating temporal data [33]. The application of LLMs to temporal data generation remains relatively underexplored [10]. This limitation is due to LLMs being inherently designed for text-based tasks, and they must treat numerical temporal data as sequences of tokens.
Few studies can jointly generate both static event data and temporal data. Traditional GAN-based methods [16] generate medical time-series data by a hybrid multi-generator framework. LLM-based methods, such as HALO [9] and SynEHRgy [10], employ novel tokenization strategies. Although these methods achieve joint generation of medical time-series data, they face notable limitations: GAN-based methods often produce unreliable static event data, while LLM-based methods struggle to generate high-quality temporal data and face high computational cost. Therefore, modularizing the medical time-series data generation task using two modules to generate static event and temporal data may yield better performance than relying on a single model.

2.3. Medical Time-Series Data Quality Assurance

Data quality assurance plays a crucial role in guaranteeing the quality of generated medical time-series data. This assurance must consider not only clinical constraints (value range limits on temporal variables and logical limits among sequential diagnoses in static event variables) but also variable dependencies (logical inferences exist between different variables). In the following, we provide an overview of existing data assurance methods.
A few existing studies have addressed clinical constraints in medical time-series data generation. As shown in Table 1, some methods perform pre-generation quality assurance by incorporating quality assurance mechanisms directly within the generative model architecture [16,24] or leveraging the reasoning capabilities of LLMs to ensure logical consistency [24]. Other methods [34] perform post-generation quality assurance by employing sampling and data transformation techniques to maintain value ranges and data integrity. Although these methods can enforce numerical value ranges for temporal data, they fall short in validating static event data, which require logical constraints. For example, a patient cannot have a diagnosis of poliomyelitis immediately followed by Alzheimer’s.
Due to the complex logical relationships among variables in medical time-series data, very few studies have addressed the assurance of variable dependencies in generated data. Existing work, such as PromptEHR [19], leverages the reasoning capabilities of LLMs to infer dependencies from one variable to another during the generation process. However, the generation process of LLMs is prone to hallucination [35], which limits the reliability of such inferred dependencies. As a result, though LLMs may capture dependencies among variables when generating data, it is also essential to validate the variable dependencies of the generated medical time-series data again by post-generation assurance.

2.4. Interpretability of Data Generation

Traditional medical time-series data generation methods, such as GANs [20,21,22,25,26] and VAEs [23,28], are inherently black-box models, making it difficult to provide meaningful explanations for the data generation process. Moreover, existing LLM-based methods [9,10,19,24] also lack the interpretability of data generation. However, the interpretability of medical time-series data generation is essential for researchers, making the interpretability of the data generation process crucial.

3. QAMT Overview

QAMT enables the joint generation of high-quality medical time-series data while preserving the interpretability throughout the generation process. As illustrated in Figure 1, this framework consists of four main modules. The health knowledge graph module provides domain-specific knowledge to LLMs. The GAN-based temporal data generation module and the LLM-based static event data generation module are responsible for jointly generating medical time-series data. The medical time-series data quality assurance module evaluates the generated data to ensure the overall quality and clinical plausibility of the final generated data. The entire medical time-series data generation and quality assurance process can be divided into the following steps:
(a)
Based on real-world medical time-series datasets and open knowledge bases, a health knowledge graph (HKG) is constructed using an existing Health Knowledge Graph Builder (HKGB). This knowledge graph serves as a domain-specific knowledge resource for downstream LLMs (Section 4.1).
(b)
Construct HKG-CoT, a chain-of-thought (CoT) reasoning process enriched with clinical knowledge from the HKG, which provides healthcare-specific inference capabilities (Section 4.2).
(c)
The static event data generation module uses Retrieval-Augmented Generation (RAG) guided by the HKG to generate static event data m (Section 5.1), referred to as Health Knowledge Graph-based Retrieval Augmented Generation (HKG-RAG):
m = HKG - RAG ( m , H K G )
(d)
The medical time-series data quality assurance module then evaluates the generated static event data using HKG-CoT (Section 6.1), obtaining constrained static event data m with clinical constraints and logical consistency. Then, the evaluation results are fed back to the static event data generation module:
Constraint 1 : C C 1 ( m ) = HKG - CoT ( m , H K G ) 0 , 1
m = C C 1 ( m ) · m
(e)
The temporal data generation module uses a GAN to generate temporal data t (Section 5.2), leveraging constrained static event data as conditional guidance:
t = GAN t , m
(f)
The generated temporal data are further validated by the medical time-series data quality assurance module, using the Concept Knowledge Graph (CKG), which is a subgraph within the HKG (Section 6.2), to check for clinical value range constraints and plausibility and obtain constrained temporal data:
Constraint 2 : C C 2 t = CKG t , H K G 0 , 1
(g)
The constrained temporal data are then input to an LLM-based diagnostic model, LLM-EvPredict, which predicts its corresponding static event data. Then, the medical time-series data quality assurance module compares the predicted static event data with the previously constrained static event data using the LLM-TSAssure (Section 6.3):
Constraint 3 : V D ( m , t ) = LLM - TSAssure m , LLM - EvPredict t 0 , 1
(h)
Finally, if the predicted static event data and constrained static event data are deemed consistent, the static event data and temporal data are considered to satisfy variable dependencies and are jointly assembled into final, reliable synthetic medical time-series data V (Section 6.3):
V = V D m , t C C 1 m · m C C 2 t · t
The following sections provide a detailed description of each module and explain how they contribute to the joint generation and quality assurance of medical time-series data.

4. Health Knowledge Graph Module

The emergence of LLMs has introduced new potential for data generation. However, existing LLMs lack sufficient domain-specific knowledge in the medical time-series data generation task, leading them to generate hallucinated information. One solution involves pretraining LLMs on domain-specific corpora to enhance their medical knowledge [36,37]. However, it is computationally expensive. Another strategy involves prompt tuning to improve LLM performance [37,38], which requires relevant domain knowledge. Therefore, an increasingly explored method is to integrate knowledge graphs (KGs) with LLMs [39,40,41,42,43]. By leveraging the structured domain knowledge encoded in KGs, LLMs can be guided toward more accurate generation and reasoning. In this paper, we construct a health knowledge graph module to provide domain-specific expertise to the LLMs. In Section 4.1, we describe how we construct the HKG. In Section 4.2, we explain how HKG is used to guide reliable CoT reasoning. In Section 4.3, we demonstrate how HKG assists in generating trustworthy medical information.

4.1. HKG

To construct a reliable HKG, we leverage an existing HKGB [30]. As shown in Figure 2, the HKGB is an end-to-end platform designed to construct disease-specific and scalable health knowledge graphs (HKGs) from diverse sources as well as evaluate its generated HKG. In this study, we built a reliable HKG by leveraging external knowledge bases together with medical time-series data (e.g., MIMIC-III and eICU). The HKG constructed by the HKGB consists of three types of nodes: concept nodes, entity nodes, and event nodes. Among them, concept nodes are extracted from open knowledge bases and include medical information such as diseases, symptoms, drugs, treatments, clinical indicators, and clinical constraints. We refer to the subgraph containing only concept nodes as the Concept Knowledge Graph (CKG). Since the construction of the CKG relies on external knowledge bases, its update is relatively slow.
In contrast, entity nodes and event nodes are derived from real medical time-series datasets. Entity nodes, which represent patient information (e.g., demographic) or contextual information, are obtained through event extraction, and event nodes, which capture static clinical events in the medical time-series data, are extracted through entity extraction. Furthermore, we use ER-RDF to extract column-level information from the datasets to enrich the node information. We refer to the subgraph containing entity and event nodes as the Instance Knowledge Graph (IKG). Since the construction of the IKG depends on real medical time-series datasets, it is frequently updated and continuously expanded as new medical time-series data become available.
In this work, the HKG is defined as the integration of the CKG and IKG, thereby capturing both general medical knowledge and instance-level clinical information. Given the complexity and heterogeneity of these nodes, the HKG defines edges to represent relationships between them, with several illustrative examples provided in Table 2.

4.2. HKG-CoT

Handling reasoning tasks such as arithmetic, commonsense, and symbolic reasoning has always been a challenge [44]. Previous studies have explored the use of KGs for logical reasoning [45]. However, selecting nodes based solely on similarity does not necessarily lead to correct or complete reasoning outcomes. With the rapid development of LLMs, recent work has explored LLMs’ potential in reasoning by constructing a CoT [38], which aims to solve complex reasoning problems by generating a sequence of intermediate reasoning steps through manually designed prompts. Although CoT prompting has demonstrated promising performance in reasoning, it often suffers from hallucinations when applied to knowledge-intensive tasks, primarily due to the lack of related knowledge. To address this limitation, recent work proposed the concept of KG-CoT [41], which combines the explicit relational structure of KGs with the step-by-step reasoning capabilities of CoT prompting. This method decomposes complex problem-solving into manageable steps, enhancing the reasoning capabilities of LLMs while providing an observable reasoning process that improves interpretability [35]. However, its multi-turn question mechanism inevitably leads to increased computational cost [46]. Since the generated static event data consist of various events for multiple patients, and there should be certain logical constraints among the events of each individual, it is reasonable to use a CoT to infer the logic between a patient’s multiple events. Therefore, in this paper, we construct an HKG-CoT to fulfill static event data quality assurance.

4.3. HKG-RAG

To address the limitations of LLMs in accessing external domain-specific knowledge, many studies have adopted Retrieval-Augmented Generation (RAG) [47], which enhances LLMs by incorporating few-shot prompts retrieved from external sources (e.g., real-world medical time-series data or open knowledge bases). However, information retrieved directly from data-rich databases often lacks reliability and interpretability [48]. Recent research has therefore focused on integrating KGs into retrieval strategies to strengthen LLMs’ generation ability [49,50]. Compared to databases, KGs provide structured and inferable knowledge, making them more suitable for enhancing RAG. Recent work proposed KG-RAG [40], demonstrating that KGs could effectively enhance the performance of LLMs. In the generation tasks, RAG has the advantage of encoding sensitive data in a secure manner, thereby reducing the risk of privacy leakage. Furthermore, compared to a CoT, RAG can eliminate certain intermediate generation steps, thus lowering computational costs. Given that static event data generation requires the rapid synthesis of large volumes of data while minimizing privacy risks during the generation process, we propose a method, HKG-RAG, to support the generation of static event data.

5. Medical Time-Series Data Generation Module

As analyzed in Section 2.2, since GANs and LLMs each have their advantages in static event data generation and temporal data generation, respectively, QAMT adopts dual modules to accomplish joint medical time-series data generation: the LLM-based static event data generation module (Section 5.1) and the GAN-based temporal data generation module (Section 5.2).

5.1. Static Event Data Generation Module

In the static event data generation module, we employ the HKG-RAG introduced in Section 4.3. As illustrated in Figure 3, the process of static event data generation consists of the following steps:
Step 1: Privacy-insensitive demographic sampling and prompt customization. Demographic information c, such as patient ID, age, diagnosis time, religion, and marital status, is randomly sampled from real-world medical time-series data. This information is first checked by an LLM, obtaining reliable demographic information c as shown in Equation (8), and then used to construct a customized prompt q, which is fed into another LLM:
c = LLM - judge S a m p l e c
Step 2: Entity recognition. Entities e = e n 1 , e n 2 e n m q are extracted from the input prompt q = N L c and matched to corresponding nodes in the HKG:
N q = n s i m n , e n i > τ HKG
where τ is the threshold value. During the entity extraction, QAMT employs zero-shot prompting on an LLM distinct from the previous one to ensure accuracy. Next, entity linking is performed to obtain the corresponding entities in the HKG. We utilize the sentence embedding model “all-MiniLM-L6-v2” to encode entity nodes into dense vector representations [51]. The similarity between the extracted entities and HKG nodes is then calculated in this embedding space to determine the best match.
Step 3: Contextual retrieval of clinical outcomes and events based on the HKG. In the HKG, entity nodes are connected to event nodes through contextual triples (Subject, Predicate, Object), following a defined schema as shown in Table 2. Since multiple entities may be extracted from one prompt and may correspond to several entity nodes in the HKG, we select the top-Krag most frequently occurring objects across inferred triples as the final context targets. These are then structured into n-ary tuples: (Subject1, Subject2, …, Subjectn, Predicate, Object). These tuples can be directly transformed into a natural language sentence using the following pattern: (Subject1, Subject2, …, Subjectn, Predicate, Object) → Subjects predicateName Object.
Step 4: Generation of clinical outcomes and events. The prompt-aware content is used as few-shot prompts for the LLM to generate the final output. The generated clinical outcomes and events are then matched with the corresponding demographic information to form the final generated static event data.
m = m S ˜ HKG - RAG m S ˜ ,   f e w s h o t .
Step 5: Feedback mechanism. In the initial generation process, we set the size of top-Krag to 5. Subsequently, the medical time-series data quality assurance module evaluates the generated static event data (Section 6.1). If the data generated by HKG-RAG fail to satisfy the clinical constraints, we expand top-Krag to 10, enabling the generation of more diverse results and correcting previously erroneous outputs. By feeding back the evaluation results from the time-series data quality assurance module to the static event data generation module, the efficiency of data generation can be improved.

5.2. Temporal Data Generation Module

In the temporal data generation module, we adopt a GAN, which has demonstrated strong performance in continuous data generation. To establish a connection between it and the static event data generation module, we incorporate the statistical information of constrained static event data m after validation (Section 6.1) as input to assist the generation of temporal data [52]:
min G max D E t P real [ log D ( t θ D ) ] + E z P z [ log ( 1 D ( G ( z , m θ G ) ) ) ]
In this GAN model, we introduce an attention mechanism to mitigate the impact of noise. Additionally, we incorporate learned positional encoding to integrate position information into the attention computation process, preserving relative distance information within the sequence.
The generator accepts input random noise vectors z and statistical information derived from the static event data c, generating time-series data D T = G t ( z , c ; θ g ) , which captures variable dependencies between the static event data and temporal data.

6. Medical Time-Series Data Quality Assurance Module

Real-world medical time-series data often exhibit multiple constraints, including logical constraints within a single patient’s static event data (e.g., a patient diagnosed with poliomyelitis is unlikely to be diagnosed with Alzheimer’s disease in a short time) and value constraints on temporal data (e.g., heart rate values cannot be zero). Moreover, there are typically variable dependencies among variables (e.g., patients with high systolic blood pressure are more likely to have hypertension).
To ensure higher quality in the generated medical time-series data, we apply the HKG-CoT (Section 6.1) and CKG (Section 6.2) to enforce constraints on the outputs from the static event data and the temporal data generation module, respectively. In addition, we use LLMs to predict the static event data corresponding to the constrained temporal data. By comparing the predicted static event data with the constrained static event data, we validate the variable dependencies in the generated medical time-series data (Section 6.3).

6.1. Clinical Constraint Assurance in Static Event Data

Since static event data often exhibit inherent logical relationships, we apply the HKG-CoT introduced in Section 4.2 to enforce logical constraints on the generated static event data. Figure 4 illustrates the detailed process of the HKG-CoT:
Step 1: Step-by-step graph reasoning model. Let n denote the number of entities in the HKG. We first initialize an entity state e 0 [ 0 , 1 ] n . If the ith entity is mentioned in the question, e i 0 e 0 is initialized to 1; otherwise, it is set to 0. The question is transformed into a one-dimensional vector q through embedding. Divide the graph reasoning process into T steps, and through the design of an attention-related function f t ( ) , obtain the question representation q t = f t ( q ) at step t, t = 1 , 2 , , T . The question representations focus on different parts of the question context at different steps. Then, we can obtain the scores of all relations in the HKG at step t by using a multi-layer perception (MLP). We define a transition matrix W t and the update formula for e is e t = e t 1 W t . After T steps of reasoning, we can obtain the confidence score of each entity based on e k .
Step 2: Reasoning path generation method. Based on the results of the step-by-step graph reasoning model, obtain the k entities with the highest confidence, denoted as E k . Perform T steps of reasoning, and for each reasoning path, calculate its score based on rules. Take the entity in question as the initial entity. At step t, select the top-k intermediate reasoning paths based on the scores and add these paths to the candidate paths. After T steps of reasoning, select the top-k candidate reasoning paths with the highest scores to form the final reasoning paths: ( Path 1 , Path 2 , , Path k ) .
Step 3: Reasoning. Serialize the selected k final reasoning paths and use detailed instructions to prompt the LLM to generate answers using these reasoning paths. Since generated static event data consist of multiple events of a single patient, we form a question by combining the first event (the event data are sorted by timestamp) with a pre-defined prompt. If the answer produced by the HKG-CoT does not contain the entity of the second event, we consider the patient’s event data to be incorrect. If this is the first time the data undergo validation, the validation results are fed back to the static event generation module for correction (Section 5.1). If the data remain incorrect even after feedback-based revision, they are discarded. Conversely, if the event data are deemed correct, we consider the first event as passed and continue to ask questions about the second event until the set of event data is discarded or all events of this patient are passed. After the data quality assurance conducted by the HKG-CoT, we consider the generated event data to be constrained.

6.2. Clinical Constraint Assurance in Temporal Data

Temporal data are often subject to value constraints. However, due to the structural complexity of medical time-series data, different datasets may record different variables, making it difficult to manually define value constraints for each variable. The CKG within the HKG introduced in Section 4.1 captures clinical constraints associated with various clinical indicators and serves as a reliable and comprehensive knowledge base. Therefore, QAMT leverages the CKG to apply value constraints to the temporal data, thereby enabling data quality assurance.
For the generated temporal data, we extract each variable and perform similarity matching with the clinical indicator nodes (concept nodes) in the CKG. By leveraging the relationships between clinical indicators and clinical constraints, we obtain the corresponding value constraints for each variable. If a variable in the temporal data exceeds its specified value range, the data are considered unreliable. Otherwise, if all values fall within the range, the data are constrained.

6.3. Assurance of Variable Dependencies

Due to the complex relationships among variables in medical time-series data, it is necessary to verify the dependencies between temporal data and static event data. For example, suppose a patient’s temporal data show elevated systolic blood pressure (within the acceptable value range), but the corresponding static event data indicate a diagnosis of hypotension. This inconsistency suggests that the generated medical time-series data are unreliable. Given the diagnostic relationship between temporal data and static event data, we utilize an LLM to predict diagnostic outcomes in the static event data based on the constrained temporal data. If the predicted static event data are similar to the originally constrained static event data, the generated medical time-series data are considered reliable.
Previous studies have shown that with carefully designed prompts, LLMs can successfully predict static event data from temporal inputs [53,54]. Therefore, we designed LLM-EvPredict, a static event prediction model based on temporal data. This LLM formats a patient’s temporal data into queries by organizing the variables into tuples (variable: values)n, and construct the corresponding prompt as: Prompt = Instruction s t a r t + Context + Instruction e n d .
The results obtained by LLM-EvPredict are written in the form of event tuples ( PreEvent 1 , PreEvent 2 , …, PreEvent n ). Another model, LLM-TSAssure, is then used to compare the predicted static event data (in tuple form) with the constrained static event data of the corresponding patient (also in tuple form). If LLM-TSAssure determines that the two sets of static events are likely to come from the same patient, we consider that variable dependencies exist between the generated temporal data and static event data and retain the data. Otherwise, the data are discarded.

7. The Interpretability of Medical Time-Series Data Generation

Due to the high modularity of QAMT and its multiple usage of LLMs throughout the time-series data generation and quality assurance, the framework offers an interpretable generation pipeline.
(1) Clearly defined modularization. As shown in Figure 1, QAMT consists of four modules, each functioning independently. Specifically, steps such as (c) static event data generation, (d) static event data quality assurance, (g) static event data prediction, and (h) assurance of variable dependencies involve the formulation of question prompts and the use of LLMs to produce outputs. These interactions with the LLMs, comprising formulated questions and corresponding answers, reflect the logical flow of medical time-series data generation, thereby providing QAMT’s interpretability.
(2) Clear collaboration between modules. Figure 5 illustrates the process of generating and validating static event data based on randomly sampled demographic information. The output of each step serves as input for the next, and the question–answer pairs at each step make the generation process interpretable, demonstrating the collaboration of the data generation task. Moreover, some steps, such as the clinical constraint assurance in static event data based on the HKG-CoT, further enhance interpretability by providing insight into the reasoning steps within the step itself.

8. Experimental Results

8.1. Experimental Setup

8.1.1. Datasets

We conducted experiments using the MIMIC-III [8] and eICU [7] datasets. MIMIC-III is a large, publicly available database that contains a wide range of medical time-series data. In this study, we extracted 40 thousand data samples following the methodology adopted in previous work [10]. Specifically, we employed the preprocessing pipeline in [55] to extract relevant data. Except for patient demographic information (age and gender), we selected 25 phenotype labels as static event data for each visit. Furthermore, we included 41 continuous temporal variables derived from vital signs and laboratory measurements. The eICU Collaborative Research Database is a comprehensive and publicly available resource that contains data on approximately 200,000 hospitalized patients. In this study, similar to previous work [52], we extracted 13 thousand data samples. In addition to patient demographic information (age, gender, and ethnicity), we selected seven phenotype labels as static event variables and 40 continuous vital signs and laboratory measurements as temporal variables.

8.1.2. Evaluation Metrics

We evaluated the generated time-series data based on fidelity, utility, and privacy [10,52].
  • Fidelity. For static event data, we assessed fidelity using the probabilities of unigram, bigram, and trigram within each visit, as well as the probabilities of sequential bigram between continuous visits. For example, the probability of a continuous visit as [ i c d 13 , i c d 920 ] was computed by dividing its frequency by the total number of patients. We then calculated the Pearson Correlation between the top 1000 n-gram probabilities in the real and generated datasets to evaluate the similarity in their distributions.
For temporal data, we first constructed embeddings for each patient by calculating the statistical features of their first 48 h of temporal data (minimum, maximum, mean, and standard deviation). Using these embeddings, we evaluated the fidelity by calculating the precision, recall, density, and coverage (PRDC) between the embeddings of the generated data and real data. Furthermore, we compared the correlation matrices of the real and generated data and report the mean squared error ( M S E c o r r ) as overall correlation fidelity.
  • Utility. We assessed the utility of the generated data by evaluating their performance across four downstream tasks involving two disease types: sepsis clustering [56], sepsis treatment strategy modeling [57], ARDS (Acute Respiratory Distress Syndrome) prediction [58], and ARDS treatment strategy modeling [59]. A smaller difference between the generated data and real data in downstream tasks indicates more similarity. For the sepsis clustering task, the study evaluated the results before and after applying its proposed method using the Sum of Squares Error (SSE) metric. Accordingly, we adopted the difference in SSE improvement between real and generated data as our evaluation metric. For the sepsis treatment strategy task, we measured the difference in patient condition improvement Δ Q between models trained on generated data and real data. For ARDS prediction, we compared the AUROC scores of classifiers trained on real data and generated data within a 12 h window before onset. For ARDS treatment strategy modeling, we compared the average reduction in mortality achieved by reinforcement learning algorithms when trained on real or generated data.
  • Privacy. We adopted the Membership Inference Attack (MIA) as the evaluation metric of privacy to determine whether specific data points were included in the data [27]. We fit a K-Nearest Neighbors (KNN) model on the generated data and the real dataset and calculated their nearest distances for each patient. A significant disparity between the distance distributions in the generated and real sets indicates lower privacy. We used the Hamming distance for static event sequences and the Euclidean distance for temporal embeddings. We then fit Gaussian distributions to these distances and assess the differences between the two distributions using the Wasserstein Distance (WD), Jensen–Shannon Divergence (JSD), and Area Under the Receiver Operating Characteristic (AUROC) metrics.

8.1.3. Baselines

Since QAMT is capable of jointly generating medical time-series data, we compared our method, based on Table 1, with HGAN [16], HALO [9], SynEHRgy [10], and SynTEG [20].

8.1.4. Experimental Details

Since QAMT utilizes LLMs across multiple steps, including static event data generation, clinical constraint assurance in static event data, static event data prediction, and assurance of variable dependencies, we employed different models for different steps, including LLaMA 2 [60], GPT-3.5, Gemini [61], and GPT-4 [62]. The experiments were conducted in an environment equipped with an NVIDIA RTX A5000 GPU (Santa Clara, CA, USA).

8.2. Medical Time-Series Data Fidelity Evaluation

Table 3 presents the correlation values of n-gram probabilities between real and generated static event data. On the MIMIC-III dataset, although SynEHRgy achieved the best performance on trigram probabilities, our method consistently outperformed all baselines, particularly in unigram, bigram, and sequential bigram probabilities. Similarly, on the eICU dataset, our method demonstrated superior performance in unigram, trigram, and sequential bigram probabilities. These results confirm that our method is capable of generating high-quality static event data.
Table 4 reports the PRDC metrics, along with the correlation difference ( M S E c o r r ) for evaluating the fidelity of temporal data. We found that our proposed framework achieved either the best or second-best performance in the PRDC metrics compared to the baselines. This is attributed to our modular design, which incorporates a GAN model better suited for temporal data generation. Moreover, our model achieved the lowest M S E c o r r , indicating that QAMT excels in capturing variable-level correlations, thanks to the variable dependencies described in Section 5.2 and Section 6.

8.3. Medical Time-Series Data Utility Evaluation

Table 5 presents the performance of different models across various downstream tasks. We found that our proposed method achieved the best performance on all tasks in the MIMIC-III dataset. On the eICU dataset, our model performed best on the SepsisClustering and SepsisTreatment tasks. Compared with the MIMIC-III dataset, we included three demographic variables (static event variables) in the eICU dataset. Although these variables were also subject to quality assurance, they were randomly generated and thus carried uncertainty. As a result, our model showed some instability in the ARDSPrediction and ARDSTreatment tasks of eICU.
In addition, we applied QAMT to generate sepsis time-series data based on the real data provided by the Emergency Department of Peking University People’s Hospital. First, the generated data improved the sepsis prediction model’s accuracy as the training data [57], which assisted doctors in clinical decisions. Second, the generated data supported clinical research on sepsis, particularly in sepsis subphenotypes, revealing significant heterogeneity in inflammatory biomarkers, treatments, and consistency across cohorts [63]. Due to privacy considerations, the related medical data cannot be publicly released.

8.4. Medical Time-Series Data Privacy Evaluation

MIA metrics are reported in Table 6. Here, we computed privacy metrics by using the Hamming distance of static event data and the Euclidean distance of temporal data. We found that none of the methods showed a privacy risk.

8.5. Robustness Analysis

8.5.1. Statistical Significance Test

To ensure the statistical reliability of the experimental conclusions, we repeated several experiments using the MIMIC-III dataset. For the static event data, we selected the probability of unigram and bigram in each visit as the evaluation metrics. For temporal data, we chose precision and recall as evaluation metrics. For time-series data, we selected the performance in the downstream tasks of sepsis clustering and sepsis treatment as the evaluation index. We measured the mean of the results and their 95% confidence intervals. The results are shown in Table 7. We found that the experimental results fell in a certain interval with high probability, and the worst value of that interval was still better than the vast majority of baseline methods. Therefore, our experimental conclusions are statistically significant.

8.5.2. Noise Robustness Analysis

We injected 0% to 20% Gaussian noise into MIMIC-III time-series data and used the fidelity of temporal data as the evaluation metric to evaluate the noise robustness of QAMT. The results are shown in Table 8. We found that QAMT was less affected by noise, as its quality assurance module performed quality validations on the generated data. As a result, QAMT demonstrates robustness against noise.

8.6. Parameter Sensitivity Analysis

We also tested the sensitivity of QAMT by varying k, which is the number of inference paths selected in the KG-CoT (Section 5.1). The results are shown in Table 9. We found that the improvement of the fidelity of temporal data became smaller as k increased. Moreover, when k > 7, the fidelity of temporal data was no longer significantly improved by increasing k.

8.7. Ablation Experiments

To validate the importance of each module in the proposed QAMT framework, we conducted an ablation study. Specifically, QAMT0 only used a GAN to generate all static event data and temporal data, QAMT1 only used an LLM to generate all static event data and temporal data, and QAMT2 used a GAN to generate temporal data and an LLM to generate static event data. On the basis of QAMT2, QAMT3 added CoT prompting to ensure clinical constraints to the generated static event data. QAMT4 introduced the external knowledge graph HKG based on QAMT3. We compared the full QAMT against QAMT0, QAMT1, QAMT2, QAMT3, and QAMT4 across fidelity, utility, and privacy metrics on the MIMIC-III dataset to demonstrate the contribution and necessity of each module in the framework.

8.7.1. Fidelity Evaluation

As shown in Figure 6, from QAMT to QAMT0, the fidelity of the data generated with fewer modules in QAMT shows a step-like decreasing trend. The results of the fidelity evaluation experiments clearly demonstrate that omitting the variable dependencies assurance (QAMT4) leads to a decline in data quality. Similarly, comparing QAMT4 and QAMT3, the lack of HKG leads to a significant performance decrease, with a more noticeable drop compared to the exclusion of variable dependencies. It shows that the external knowledge provided by the HKG helps with higher-quality medical time-series data generation. Moreover, the data generated without clinical constraints (QAMT2) also result in decreased fidelity performance. This is because during the data generation process, both GANs and LLMs may produce incorrect outputs. Therefore, comparing QAMT2 and QAMT4, we find that the clinical constraints applied after generation using the HKG-CoT and CKG can effectively eliminate these errors. We find that the medical time-series data generated with the simultaneous usage of LLM and GAN (QAMT2) have higher fidelity compared with the single usage of LLM or GAN for data generation (QAMT1 and QAMT0), showing that the modularization of the generation process in QAMT is important.

8.7.2. Utility Evaluation

Figure 7 demonstrates that, in downstream tasks, medical time-series data generated without applying the simultaneous usage of the LLM and GAN, HKG, clinical constraints, and variable dependencies perform worse than the data generated by QAMT. This further demonstrates the importance of our modular design in QAMT.

8.7.3. Privacy Evaluation

Table 10 reports that the privacy risk of the data generated by QAMT, QAMT4, QAMT3, QAMT2, QAMT1, and QAMT0 show a step-like increasing trend, indicating the significance of our modular design in QAMT.

9. Conclusions

In this study, we proposed QAMT, an LLM-based framework for quality-assured medical time-series data generation. The framework constructs a reliable HKG to inject medical expertise into LLMs and uses a dual-module method to jointly generate medical time-series data, including static event data and temporal data. In addition, QAMT incorporates a quality assurance module to evaluate the generated data. It provides clinical constraint assurance in static event data based on an HKG-CoT and in temporal data based on a CKG and employs LLM-based prediction to ensure variable dependencies. Unlike existing methods, QAMT maintains the modularity and high-level pipeline structure of the generation process, preserving interpretability.
Currently, the proposed QAMT is only applicable to the medical domain. Other domain-specific areas, such as energy [29], also involve time-series data generation tasks. Therefore, in the future, we plan to extend QAMT to other domains to support other time-series data generation tasks.

Author Contributions

Conceptualization, Y.L., Y.Z., C.X. and P.R.; Methodology, Y.L. and Y.Z.; Investigation, Y.L., P.R. and X.L.; Writing—original draft, Y.L. and X.L.; Writing—review & editing, Y.L., Y.Z., C.X. and P.R.; Visualization, Y.L. and X.L.; Supervision, Y.Z. and C.X. All authors have read and agreed to the published version of the manuscript.

Funding

This research was funded by the National Natural Science Foundation of China under Grant 42371480.

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

We use the open source data MIMIC-III and eICU. The data provided by the Emergency Department of Peking University People’s Hospital is unavailable due to privacy concerns.

Conflicts of Interest

The authors declare no conflicts of interest.

References

  1. Cruz-Vega, I.B.; Ávila Vanzzini, N.; González-Gómez, G.H.; Springall, R.; Echeverría, J.C.; Lerma, C. Dynamic Response of Heart Rate Variability to Active Standing in Aortic Valve Disease: Insights from Recurrence Quantification Analysis. Sensors 2025, 25, 1535. [Google Scholar] [CrossRef]
  2. Fan, Y.; Dang, Y.; Guo, Y. Fault Identification Model Using Convolutional Neural Networks with Transformer Architecture. Sensors 2025, 25, 3897. [Google Scholar] [CrossRef]
  3. Rupprechter, S.; Morinan, G.; Peng, Y.; Foltynie, T.; Sibley, K.; Weil, R.S.; Leyland, L.A.; Baig, F.; Morgante, F.; Gilron, R.; et al. A Clinically Interpretable Computer-Vision Based Method for Quantifying Gait in Parkinson’s Disease. Sensors 2021, 21, 5437. [Google Scholar] [CrossRef]
  4. Kalahasty, R.; Yerrapragada, G.; Lee, J.; Gopalakrishnan, K.; Kaur, A.; Muddaloor, P.; Sood, D.; Parikh, C.; Gohri, J.; Panjwani, G.A.R.; et al. A Novel You Only Listen Once (YOLO) Deep Learning Model for Automatic Prominent Bowel Sounds Detection: Feasibility Study in Healthy Subjects. Sensors 2025, 25, 4735. [Google Scholar] [CrossRef]
  5. Dang, T.H.; Kim, S.m.; Choi, M.s.; Hwan, S.n.; Min, H.k.; Bien, F. An Automated Algorithm for Obstructive Sleep Apnea Detection Using a Wireless Abdomen-Worn Sensor. Sensors 2025, 25, 2412. [Google Scholar] [CrossRef] [PubMed]
  6. Randazzo, V.; Caligari, S.; Pasero, E.; Giustetto, C.; Saglietto, A.; Bertarello, W.; Averbuch, A.; Marcus-Kalish, M.; Zheludev, V.; Gaita, F. A Vision Transformer Model for the Prediction of Fatal Arrhythmic Events in Patients with Brugada Syndrome. Sensors 2025, 25, 824. [Google Scholar] [CrossRef] [PubMed]
  7. Pollard, T.J.; Johnson, A.E.; Raffa, J.D.; Celi, L.A.; Mark, R.G.; Badawi, O. The eICU Collaborative Research Database, a freely available multi-center database for critical care research. Sci. Data 2018, 5, 180178. [Google Scholar] [CrossRef]
  8. Johnson, A.E.; Pollard, T.J.; Shen, L.; Lehman, L.w.H.; Feng, M.; Ghassemi, M.; Moody, B.; Szolovits, P.; Anthony Celi, L.; Mark, R.G. MIMIC-III, a freely accessible critical care database. Sci. Data 2016, 3, 160035. [Google Scholar] [CrossRef]
  9. Theodorou, B.; Xiao, C.; Sun, J. Synthesize high-dimensional longitudinal electronic health records via hierarchical autoregressive language model. Nat. Commun. 2023, 14, 5305. [Google Scholar] [CrossRef] [PubMed]
  10. Karami, H.; Atienza Alonso, D.; Ionescu, A. SynEHRgy: Synthesizing Mixed-Type Structured Electronic Health Records using Decoder-Only Transformers. In Proceedings of the 38th Annual Conference on Neural Information Processing Systems, Vancouver, BC, Canada, 10–15 December 2024. [Google Scholar]
  11. El Emam, K.; Buckeridge, D.; Tamblyn, R.; Neisa, A.; Jonker, E.; Verma, A. The re-identification risk of Canadians from longitudinal demographics. BMC Med. Inform. Decis. Mak. 2011, 11, 46. [Google Scholar] [CrossRef]
  12. Benitez, K.; Malin, B. Evaluating re-identification risks with respect to the HIPAA privacy rule. J. Am. Med. Inform. Assoc. 2010, 17, 169–177. [Google Scholar] [CrossRef]
  13. Abbas, S.R.; Abbas, Z.; Zahir, A.; Lee, S.W. Federated learning in smart healthcare: A comprehensive review on privacy, security, and predictive analytics with IoT integration. Healthcare 2024, 12, 2587. [Google Scholar] [CrossRef]
  14. Kingma, D.P.; Welling, M. An introduction to variational autoencoders. Found. Trends® Mach. Learn. 2019, 12, 307–392. [Google Scholar] [CrossRef]
  15. Goodfellow, I.; Pouget-Abadie, J.; Mirza, M.; Xu, B.; Warde-Farley, D.; Ozair, S.; Courville, A.; Bengio, Y. Generative adversarial networks. Commun. ACM 2020, 63, 139–144. [Google Scholar] [CrossRef]
  16. Yan, C.; Zhang, Z.; Nyemba, S.; Malin, B.A. Generating Electronic Health Records with Multiple Data Types and Constraints. In Proceedings of the AMIA 2020, American Medical Informatics Association Annual Symposium, Virtual, 14–18 November 2020. [Google Scholar]
  17. Xu, L.; Skoularidou, M.; Cuesta-Infante, A.; Veeramachaneni, K. Modeling tabular data using conditional gan. arXiv 2019, arXiv:1907.00503. [Google Scholar] [CrossRef]
  18. Pradhan, P.K.; Das, A.; Kumar, A.; Baruah, U.; Sen, B.; Ghosal, P. SwinSight: A hierarchical vision transformer using shifted windows to leverage aerial image classification. Multim. Tools Appl. 2024, 83, 86457–86478. [Google Scholar] [CrossRef]
  19. Wang, Z.; Sun, J. PromptEHR: Conditional Electronic Healthcare Records Generation with Prompt Learning. In Proceedings of the 2022 Conference on Empirical Methods in Natural Language Processing, EMNLP 2022, Abu Dhabi, United Arab Emirates, 7–11 December 2022; Goldberg, Y., Kozareva, Z., Zhang, Y., Eds.; Association for Computational Linguistics: Stroudsburg, PA, USA, 2022; pp. 2873–2885. [Google Scholar] [CrossRef]
  20. Zhang, Z.; Yan, C.; Lasko, T.A.; Sun, J.; Malin, B.A. SynTEG: A framework for temporal structured electronic health data simulation. J. Am. Med. Inform. Assoc. 2021, 28, 596–604. [Google Scholar] [CrossRef]
  21. Baowaly, M.K.; Lin, C.; Liu, C.; Chen, K. Synthesizing electronic health records using improved generative adversarial networks. J. Am. Med. Inform. Assoc. 2019, 26, 228–241. [Google Scholar] [CrossRef]
  22. Lu, C.; Reddy, C.K.; Wang, P.; Nie, D.; Ning, Y. Multi-Label Clinical Time-Series Generation via Conditional GAN. IEEE Trans. Knowl. Data Eng. 2024, 36, 1728–1740. [Google Scholar] [CrossRef]
  23. Nikolentzos, G.; Vazirgiannis, M.; Xypolopoulos, C.; Lingman, M.; Brandt, E.G. Synthetic electronic health records generated with variational graph autoencoders. Npj Digit. Med. 2023, 6, 83. [Google Scholar] [CrossRef]
  24. Pang, C.; Jiang, X.; Pavinkurve, N.P.; Kalluri, K.S.; Minto, E.L.; Patterson, J.; Zhang, L.; Hripcsak, G.; Elhadad, N.; Natarajan, K. CEHR-GPT: Generating Electronic Health Records with Chronological Patient Timelines. arXiv 2024, arXiv:2402.04400. [Google Scholar] [CrossRef]
  25. Esteban, C.; Hyland, S.L.; Rätsch, G. Real-valued (Medical) Time Series Generation with Recurrent Conditional GANs. arXiv 2017, arXiv:1706.02633. [Google Scholar] [CrossRef]
  26. Karami, H.; Hartley, M.; Atienza, D.; Ionescu, A. TimEHR: Image-based Time Series Generation for Electronic Health Records. arXiv 2024, arXiv:2402.06318. [Google Scholar] [CrossRef]
  27. Yoon, J.; Mizrahi, M.J.; Ghalaty, N.F.; Jarvinen, T.; Ravi, A.S.; Brune, P.; Kong, F.; Anderson, D.; Lee, G.; Meir, A.; et al. EHR-Safe: Generating high-fidelity and privacy-preserving synthetic electronic health records. Npj Digit. Med. 2023, 6, 141. [Google Scholar] [CrossRef]
  28. Lee, Y.; Chae, Y.; Jung, K. Leveraging VQ-VAE tokenization for autoregressive modeling of medical time series. Artif. Intell. Med. 2024, 154, 102925. [Google Scholar] [CrossRef]
  29. Zhou, X.; Jia, Q.; Hu, Y.; Xie, R.; Huang, T.; Yu, F.R. GenG: An LLM-Based Generic Time Series Data Generation Approach for Edge Intelligence via Cross-Domain Collaboration. In Proceedings of the IEEE INFOCOM 2024—IEEE Conference on Computer Communications Workshops, Vancouver, BC, Canada, 20 May 2024. [Google Scholar] [CrossRef]
  30. Zhang, Y.; Sheng, M.; Zhou, R.; Wang, Y.; Han, G.; Zhang, H.; Xing, C.; Dong, J. HKGB: An inclusive, extensible, intelligent, semi-auto-constructed knowledge graph framework for healthcare with clinicians’ expertise incorporated. Inf. Process. Manag. 2020, 57, 102324. [Google Scholar] [CrossRef]
  31. Borisov, V.; Seßler, K.; Leemann, T.; Pawelczyk, M.; Kasneci, G. Language Models are Realistic Tabular Data Generators. In Proceedings of the The Eleventh International Conference on Learning Representations, ICLR 2023, Kigali, Rwanda, 1–5 May 2023. [Google Scholar]
  32. Hernandez, M.; Epelde, G.; Alberdi, A.; Cilla, R.; Rankin, D. Synthetic data generation for tabular health records: A systematic review. Neurocomputing 2022, 493, 28–45. [Google Scholar] [CrossRef]
  33. Viana, D.; Teixeira, R.; Baptista, J.; Pinto, T. Synthetic Data Generation Models for Time Series: A Literature Review. In Proceedings of the International Conference on Electrical, Computer and Energy Technologies, ICECET 2024, Sydney, Australia, 25–27 July 2024. [Google Scholar] [CrossRef]
  34. Inan, M.S.K.; Hossain, S.; Uddin, M.N. Data augmentation guided breast cancer diagnosis and prognosis using an integrated deep-generative framework based on breast tumor’s morphological information. Inform. Med. Unlocked 2023, 37, 101171. [Google Scholar] [CrossRef]
  35. Chu, Z.; Chen, J.; Chen, Q.; Yu, W.; He, T.; Wang, H.; Peng, W.; Liu, M.; Qin, B.; Liu, T. Navigate through enigmatic labyrinth a survey of chain of thought reasoning: Advances, frontiers and future. arXiv 2023, arXiv:2309.15402. [Google Scholar]
  36. Lee, J.; Yoon, W.; Kim, S.; Kim, D.; Kim, S.; So, C.H.; Kang, J. BioBERT: A pre-trained biomedical language representation model for biomedical text mining. Bioinformatics 2020, 36, 1234–1240. [Google Scholar] [CrossRef] [PubMed]
  37. Peng, Y.; Yan, S.; Lu, Z. Transfer learning in biomedical natural language processing: An evaluation of BERT and ELMo on ten benchmarking datasets. arXiv 2019, arXiv:1906.05474. [Google Scholar] [CrossRef]
  38. Wei, J.; Wang, X.; Schuurmans, D.; Bosma, M.; Xia, F.; Chi, E.; Le, Q.V.; Zhou, D. Chain-of-thought prompting elicits reasoning in large language models. Adv. Neural Inf. Process. Syst. 2022, 35, 24824–24837. [Google Scholar]
  39. Lin, B.Y.; Chen, X.; Chen, J.; Ren, X. Kagnet: Knowledge-aware graph networks for commonsense reasoning. arXiv 2019, arXiv:1909.02151. [Google Scholar]
  40. Soman, K.; Rose, P.W.; Morris, J.H.; Akbas, R.E.; Smith, B.; Peetoom, B.; Villouta-Reyes, C.; Cerono, G.; Shi, Y.; Rizk-Jackson, A.; et al. Biomedical knowledge graph-optimized prompt generation for large language models. Bioinformatics 2024, 40, btae560. [Google Scholar] [CrossRef]
  41. Zhao, R.; Zhao, F.; Wang, L.; Wang, X.; Xu, G. Kg-cot: Chain-of-thought prompting of large language models over knowledge graphs for knowledge-aware question answering. In Proceedings of the Thirty-Third International Joint Conference on Artificial Intelligence (IJCAI-24), Jeju, Republic of Korea, 3–9 August 2024; pp. 6642–6650. [Google Scholar]
  42. Matsumoto, N.; Moran, J.; Choi, H.; Hernandez, M.E.; Venkatesan, M.; Wang, P.; Moore, J.H. KRAGEN: A knowledge graph-enhanced RAG framework for biomedical problem solving using large language models. Bioinformatics 2024, 40, btae353. [Google Scholar] [CrossRef]
  43. Shao, S.; Lin, S.; Huang, Z. A Medical Consultation System for Geriatric Disease Based on Multi-agent Architecture and Knowledge Graph. In Proceedings of the Health Information Science-13th International Conference, HIS 2024, Hong Kong, China, 8–10 December 2024; Proceedings; Siuly, S., Xing, C., Li, X., Zhou, R., Eds.; Lecture Notes in Computer Science; Springer: Singapore, 2024; Volume 15336, pp. 313–325. [Google Scholar] [CrossRef]
  44. Rae, J.W.; Borgeaud, S.; Cai, T.; Millican, K.; Hoffmann, J.; Song, F.; Aslanides, J.; Henderson, S.; Ring, R.; Young, S.; et al. Scaling language models: Methods, analysis & insights from training gopher. arXiv 2021, arXiv:2112.11446. [Google Scholar]
  45. Ma, K.; Cheng, H.; Liu, X.; Nyberg, E.; Gao, J. Open-domain question answering via chain of reasoning over heterogeneous knowledge. arXiv 2022, arXiv:2210.12338. [Google Scholar] [CrossRef]
  46. Xia, Y.; Wang, R.; Liu, X.; Li, M.; Yu, T.; Chen, X.; McAuley, J.; Li, S. Beyond chain-of-thought: A survey of chain-of-x paradigms for llms. arXiv 2024, arXiv:2404.15676. [Google Scholar]
  47. Lewis, P.; Perez, E.; Piktus, A.; Petroni, F.; Karpukhin, V.; Goyal, N.; Küttler, H.; Lewis, M.; Yih, W.t.; Rocktäschel, T.; et al. Retrieval-augmented generation for knowledge-intensive nlp tasks. Adv. Neural Inf. Process. Syst. 2020, 33, 9459–9474. [Google Scholar]
  48. Zhao, X.; Liu, S.; Yang, S.Y.; Miao, C. Medrag: Enhancing retrieval-augmented generation with knowledge graph-elicited reasoning for healthcare copilot. In Proceedings of the Proceedings of the ACM on Web Conference 2025, Sydney, Australia, 28 April–2 May 2025; pp. 4442–4457. [Google Scholar]
  49. Jiang, J.; Zhou, K.; Zhao, W.X.; Li, Y.; Wen, J.R. ReasoningLM: Enabling structural subgraph reasoning in pre-trained language models for question answering over knowledge graph. arXiv 2023, arXiv:2401.00158. [Google Scholar]
  50. Kang, M.; Kwak, J.M.; Baek, J.; Hwang, S.J. Knowledge graph-augmented language models for knowledge-grounded dialogue generation. arXiv 2023, arXiv:2305.18846. [Google Scholar]
  51. Reimers, N.; Gurevych, I. Sentence-bert: Sentence embeddings using siamese bert-networks. arXiv 2019, arXiv:1908.10084. [Google Scholar] [CrossRef]
  52. Luo, Y.; Sheng, M.; Liu, X.; Wang, K.; Zhang, Y.; Zhao, H. Ltgan: Multi-label time-series gan with constraints for electronic health records generation. In Proceedings of the International Conference on Health Information Science, Hong Kong, China, 8–10 December 2024; Springer: Singapore, 2025; pp. 36–47. [Google Scholar]
  53. Kim, Y.; Xu, X.; McDuff, D.; Breazeal, C.; Park, H.W. Health-llm: Large language models for health prediction via wearable sensor data. arXiv 2024, arXiv:2401.06866. [Google Scholar] [CrossRef]
  54. Jin, M.; Yu, Q.; Shu, D.; Zhang, C.; Fan, L.; Hua, W.; Zhu, S.; Meng, Y.; Wang, Z.; Du, M.; et al. Health-LLM: Personalized retrieval-augmented disease prediction system. arXiv 2024, arXiv:2402.00746. [Google Scholar]
  55. Harutyunyan, H.; Khachatrian, H.; Kale, D.C.; Ver Steeg, G.; Galstyan, A. Multitask learning and benchmarking with clinical time series data. Sci. Data 2019, 6, 96. [Google Scholar] [CrossRef] [PubMed]
  56. Hao, R.; Sheng, M.; Zhang, Y.; Zhao, H.; Hao, C.; Li, W.; Wang, L.; Li, C. Enhancing clustering performance in sepsis time series data using gravity field. In Proceedings of the International Conference on Health Information Science, Melbourne, Australia, 23–24 October 2023; Springer: Singapore, 2023; pp. 199–212. [Google Scholar]
  57. Wang, Z.; Zhao, H.; Ren, P.; Zhou, Y.; Sheng, M. Learning optimal treatment strategies for sepsis using offline reinforcement learning in continuous space. In Proceedings of the International Conference on Health Information Science, Virtual, 28–30 October 2022; Springer: Cham, Switzerland, 2022; pp. 113–124. [Google Scholar]
  58. Le, S.; Pellegrini, E.; Green-Saxena, A.; Summers, C.; Hoffman, J.; Calvert, J.; Das, R. Supervised machine learning for the early prediction of acute respiratory distress syndrome (ARDS). J. Crit. Care 2020, 60, 96–102. [Google Scholar] [CrossRef]
  59. Zheng, H.; Zhu, J.; Xie, W.; Zhong, J. Reinforcement learning assisted oxygen therapy for COVID-19 patients under intensive care. BMC Med. Inform. Decis. Mak. 2021, 21, 350. [Google Scholar] [CrossRef]
  60. Touvron, H.; Martin, L.; Stone, K.; Albert, P.; Almahairi, A.; Babaei, Y.; Bashlykov, N.; Batra, S.; Bhargava, P.; Bhosale, S.; et al. Llama 2: Open foundation and fine-tuned chat models. arXiv 2023, arXiv:2307.09288. [Google Scholar] [CrossRef]
  61. Team, G.; Georgiev, P.; Lei, V.I.; Burnell, R.; Bai, L.; Gulati, A.; Tanzer, G.; Vincent, D.; Pan, Z.; Wang, S.; et al. Gemini 1.5: Unlocking multimodal understanding across millions of tokens of context. arXiv 2024, arXiv:2403.05530. [Google Scholar] [CrossRef]
  62. Brown, T.; Mann, B.; Ryder, N.; Subbiah, M.; Kaplan, J.D.; Dhariwal, P.; Neelakantan, A.; Shyam, P.; Sastry, G.; Askell, A.; et al. Language models are few-shot learners. Adv. Neural Inf. Process. Syst. 2020, 33, 1877–1901. [Google Scholar]
  63. Hao, C.; Hao, R.; Zhao, H.; Zhang, Y.; Sheng, M.; An, Y. Identification and validation of sepsis subphenotypes using time-series data. Heliyon 2024, 10, e28520. [Google Scholar] [CrossRef] [PubMed]
Figure 1. The framework of QAMT, where the blue rounded rectangles represent data, red rectangles represent generative models, green rectangles represent medical time-series data quality assurance models, and yellow rectangles represent the health knowledge graphs. The parts marked with icons indicate the use of LLMs.
Figure 1. The framework of QAMT, where the blue rounded rectangles represent data, red rectangles represent generative models, green rectangles represent medical time-series data quality assurance models, and yellow rectangles represent the health knowledge graphs. The parts marked with icons indicate the use of LLMs.
Sensors 25 05482 g001
Figure 2. The detailed workflow of the HKGB to construct an HKG.
Figure 2. The detailed workflow of the HKGB to construct an HKG.
Sensors 25 05482 g002
Figure 3. HKG-RAG workflow. The demographic information is randomly generated. An LLM is then used to assess its validity to avoid logically inconsistent cases, such as an age of “6” with a marital status of “MARRIED”. Only demographic data that pass this logical check are used as input for generating complete static event data.
Figure 3. HKG-RAG workflow. The demographic information is randomly generated. An LLM is then used to assess its validity to avoid logically inconsistent cases, such as an age of “6” with a marital status of “MARRIED”. Only demographic data that pass this logical check are used as input for generating complete static event data.
Sensors 25 05482 g003
Figure 4. HKG-CoT workflow. It presents an example of how the HKG-CoT is used for static event data quality assurance. The solid lines indicate that the first event in the generated data is extracted and inserted into a customized prompt, which is then submitted to the LLM for reasoning. The HKG-CoT generates a response Type 2 Diabetes Mellitus, Diabetes Insipidus, Primary Polydipsia for the question. Since Type 2 Diabetes Mellitus matches the second event in the data sequence, showing as a check mark, the first event passes the validation, and the process continues by questioning the next event, and so on. If all events pass this validation, the entire dataset is considered trustworthy. In contrast, as shown by the dashed lines, if the LLM’s response does not match the event, showing as a cross mark, and does not contain the next event, the event sequence is deemed invalid and discarded.
Figure 4. HKG-CoT workflow. It presents an example of how the HKG-CoT is used for static event data quality assurance. The solid lines indicate that the first event in the generated data is extracted and inserted into a customized prompt, which is then submitted to the LLM for reasoning. The HKG-CoT generates a response Type 2 Diabetes Mellitus, Diabetes Insipidus, Primary Polydipsia for the question. Since Type 2 Diabetes Mellitus matches the second event in the data sequence, showing as a check mark, the first event passes the validation, and the process continues by questioning the next event, and so on. If all events pass this validation, the entire dataset is considered trustworthy. In contrast, as shown by the dashed lines, if the LLM’s response does not match the event, showing as a cross mark, and does not contain the next event, the event sequence is deemed invalid and discarded.
Sensors 25 05482 g004
Figure 5. Interpretable prompt example, where (c), (d), (g) and (h) corresponds to the step introduced in Section 3.
Figure 5. Interpretable prompt example, where (c), (d), (g) and (h) corresponds to the step introduced in Section 3.
Sensors 25 05482 g005
Figure 6. Fidelity metrics for medical time-series data in ablation experiments. Static event data results are on the left, and temporal data results are on the right.
Figure 6. Fidelity metrics for medical time-series data in ablation experiments. Static event data results are on the left, and temporal data results are on the right.
Sensors 25 05482 g006
Figure 7. Utility metrics for medical time-series data in ablation experiments. Each point in the line chart represents the difference between QAMT n and the real data.
Figure 7. Utility metrics for medical time-series data in ablation experiments. Each point in the line chart represents the difference between QAMT n and the real data.
Sensors 25 05482 g007
Table 1. Comparison of medical time-series data generation methods.
Table 1. Comparison of medical time-series data generation methods.
MethodTypeDomainJointGenerationQualityAssurance 1Interpretability
SynTEG [20]GAN-basedMedical-
BGAN [21]GAN-basedMedical-
MTGAN [22]GAN-basedMedical-
VGAE [23]VAE-basedMedical-
PromptEHR [19]LLM-basedMedicalCC, VD
CEHR-GPT [24]LLM-basedMedicalCC
RTSGAN [25]GAN-basedMedical-
TimEHR [26]GAN-basedMedical-
EHR-Safe [27]GAN-basedMedical-
CodeAR [28]VAE-basedMedical-
GenG [29]LLM-basedEnergy-
HGAN [16]GAN-basedMedicalCC, VD
HALO [9]LLM-basedMedicalCC
SynEHRgy [10]LLM-basedMedical-
QAMT (ours)LLM-basedmedicalCC, VD
1 “CC” refers to clinical constraints, and “VD” refers to variable dependencies. “✓” indicates that the method has solved the corresponding challenge, while “✗” indicates that it has not been solved.
Table 2. Examples of edge relationships in the HKG.
Table 2. Examples of edge relationships in the HKG.
RelationSource Node 1,2Target Node 1,2Example
Concept → Concept
_has_symptom_DisSymCOPD -_has_symptom_→ fever
_is_treated_by_Dis/SymMed/TreCOPD -_is_treated_by_→ β -agonist
_indicated_by_DisCIDiabetes -_indicated_by_→ HbA1c
_progresses_to_DisDisPrediabetes -_progresses_to_→ Type 2 Diabetes
_treated_Med/TreDis/Sym β -agonist -_treated_→ COPD
_cause_SymDisfever -_cause_→ COPD
_measured_CIDisPaO2 -_measured_→ COPD
_constraints_CICCPaO2 -_constraints_→ 20–50 mmHg
Entity/Event → Concept
_conforms_to_Entity/EventCCEUI:67178.Lab Result -_conforms_to_→ HbA1c > 7%
_diagnoses_EventDisEUI:43613.Description -_diagnoses_→ Hypertension
_prescribes_EventMed/TreEUI:54241.Prescription -_prescribes_→ Amoxicillin
_associated_with_EventCIEUI:67129.Indicators -_associated_with_→ PaO 2
Entity/Event → Entity/Event
_participate_in_EntityEventIUI:7657 -_participate_in_→ EUI:67129
_associate_with_Entity (A/Eth)EventAge 65 -_associate_with_→ EUI:412546
_occurs_at_EventEntity (L)EUI:63415 -_occurs_at_→ Beijing
_time_is_EventEntity (T)EUI:132807 -_time_is_→ Year 2010
_event_type_is_EventEntity (ET)EUI:67129 -_event_type_is_→ Surgery
_after_/_before_EventEventEUI:67129 -_after_→ EUI:162308
1 For concept nodes, “Dis” refers to diseases, “Sym” refers to symptoms, “Med” refers to medications, “Tre” refers to treatments, “CI” refers to clinical indicators, and “CC” refers to clinical constraints. 2 For entity nodes, “A” refers to age, “Eth” refers to ethnicity, “L” refers to location, “T” refers to time, “ET” refers to static event type.
Table 3. Fidelity metrics for static event data 1.
Table 3. Fidelity metrics for static event data 1.
MIMIC-III dataset
UnigramBigramTrigramSequential Bigram
HGAN0.8320.4450.5130.487
SynEHRgy0.9070.7170.7380.571
HALO0.8720.2870.3130.521
SynTEG0.8580.5010.6470.562
QAMT0.9280.7210.7180.631
eICU dataset
UnigramBigramTrigramSequential Bigram
HGAN0.7990.3190.4830.366
SynEHRgy0.8480.7630.7110.500
HALO0.7690.2930.2810.474
SynTEG0.7870.5790.6630.492
QAMT0.8970.7550.7360.592
1 Bold and underline values indicate the best and second-best results, respectively.
Table 4. Fidelity metrics for temporal data 1.
Table 4. Fidelity metrics for temporal data 1.
MIMIC-III dataset
PrecisionRecallDensityCoverage M S E c o r r
SynEHRgy0.781 (0.011)0.853 (0.003)0.711 (0.016)0.852 (0.008)0.036
HGAN0.731 (0.023)0.617 (0.028)0.745 (0.045)0.315 (0.004)0.083
HALO0.503 (0.038)0.461 (0.002)0.372 (0.029)0.215 (0.009)0.075
SynTEG0.610 (0.013)0.721 (0.006)0.672 (0.031)0.507 (0.005)0.045
QAMT0.811 (0.009)0.859 (0.011)0.739 (0.036)0.638 (0.003)0.024
eICU dataset
PrecisionRecallDensityCoverage M S E c o r r
SynEHRgy0.814 (0.018)0.822 (0.005)0.701 (0.011)0.714 (0.007)0.042
HGAN0.669 (0.037)0.691 (0.042)0.728 (0.032)0.594 (0.004)0.045
HALO0.400 (0.061)0.417 (0.024)0.296 (0.018)0.395 (0.013)0.062
SynTEG0.523 (0.020)0.806 (0.012)0.633 (0.022)0.678 (0.006)0.051
QAMT0.863 (0.017)0.817 (0.009)0.698 (0.025)0.743 (0.005)0.033
1 Bold and underline values indicate the best and second-best results, respectively.
Table 5. Utility metrics for medical time-series data 1.
Table 5. Utility metrics for medical time-series data 1.
MIMIC-III dataset
SpesisClustering
( Δ S S E %)
SpesisTreatment
( Δ Q )
ARDSPrediction
(AUROC)
ARDSTreatment
( Δ M o r t a l i t y R a t e %)
Real data−32.370.2170.809−2.33
HGAN−28.670.1890.764−1.73
SynEHRgy−28.110.2030.818−2.17
HALO−27.720.1950.793−2.11
SynTEG−28.020.1910.799−2.13
QAMT−29.910.2140.801−2.28
eICU dataset
SpesisClustering
( Δ S S E %)
SpesisTreatment
( Δ Q )
ARDSPrediction
(AUROC)
ARDSTreatment
( Δ M o r t a l i t y R a t e %)
Real data−40.820.1720.813−2.49
HGAN−38.820.1610.825−2.01
SynEHRgy−37.410.1670.816−2.31
HALO−43.580.1650.814−2.16
SynTEG−40.070.1620.820−2.20
QAMT−42.170.1730.807−2.24
1 Bold and underline values indicate the best and second-best results, respectively.
Table 6. Privacy metrics for medical time-series data.
Table 6. Privacy metrics for medical time-series data.
MIMIC-III Dataset
Static Event Data Temporal Data
MethodJSDWDAUROC MethodJSDWDAUROC
HGAN0.0150.0010.482 HGAN0.0010.0030.482
SynEHRgy0.0140.0010.461 SynEHRgy0.0020.0020.492
HALO0.0130.0000.477 HALO0.0030.0010.493
SynTEG0.0140.0000.469 SynTEG0.0020.0010.488
QAMT0.0130.0000.456 QAMT0.0010.0020.477
eICU Dataset
Static Event Data Temporal Data
MethodJSDWDAUROC MethodJSDWDAUROC
HGAN0.0150.0020.496 HGAN0.0010.0010.497
SynEHRgy0.0150.0020.479 SynEHRgy0.0010.0020.508
HALO0.0150.0010.486 HALO0.0030.0020.509
SynTEG0.0150.0010.481 SynTEG0.0020.0020.506
QAMT0.0150.0010.456 QAMT0.0010.0020.504
Table 7. Statistical significance test.
Table 7. Statistical significance test.
Static event data
UnigramBigram
Value 0.9280.721
95% CI [0.905, 0.951][0.712, 0.730]
p p < 0.05p < 0.05
Temporal data
PrecisionRecall
Value 0.8110.859
95% CI [0.802, 0.820][0.852, 0.866]
p p < 0.05p < 0.05
Time-series data
SpesisClustering
( Δ S S E %)
SpesisTreatment
( Δ Q )
Value −29.910.214
95% CI [−30.27, −29.55][0.211, 0.217]
p p < 0.05p < 0.01
Table 8. Noise robustness analysis.
Table 8. Noise robustness analysis.
Noise IntensityPrecisionRecallDensityCoverage MSE corr
0%0.8110.8590.7390.6380.024
10%0.8060.8480.7350.6360.025
20%0.8030.8410.7280.6330.025
Influence0.6%2.1%1.5%0.8%4.2%
Table 9. Parameter sensitivity analysis.
Table 9. Parameter sensitivity analysis.
kPrecisionRecallDensityCoverage MSE corr
30.7900.8370.7280.6150.022
50.8110.8590.7390.6380.024
70.8140.8710.7430.6440.025
100.8150.8730.7440.6460.025
Table 10. Privacy metrics for medical time-series data in ablation experiments.
Table 10. Privacy metrics for medical time-series data in ablation experiments.
Static Event Data Temporal Data
MethodJSDWDAUROC MethodJSDWDAUROC
QAMT0.0130.0000.456 QAMT0.0010.0020.477
QAMT40.0160.0040.457 QAMT40.0020.0020.478
QAMT30.0440.0130.462 QAMT30.0050.0060.484
QAMT20.0500.0140.465 QAMT20.0050.0070.486
QAMT10.0710.0190.470 QAMT10.0060.0080.491
QAMT00.0680.0180.469 QAMT00.0060.0080.492
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Luo, Y.; Zhang, Y.; Xing, C.; Ren, P.; Liu, X. QAMT: An LLM-Based Framework for Quality-Assured Medical Time-Series Data Generation. Sensors 2025, 25, 5482. https://doi.org/10.3390/s25175482

AMA Style

Luo Y, Zhang Y, Xing C, Ren P, Liu X. QAMT: An LLM-Based Framework for Quality-Assured Medical Time-Series Data Generation. Sensors. 2025; 25(17):5482. https://doi.org/10.3390/s25175482

Chicago/Turabian Style

Luo, Yi, Yong Zhang, Chunxiao Xing, Peng Ren, and Xinhao Liu. 2025. "QAMT: An LLM-Based Framework for Quality-Assured Medical Time-Series Data Generation" Sensors 25, no. 17: 5482. https://doi.org/10.3390/s25175482

APA Style

Luo, Y., Zhang, Y., Xing, C., Ren, P., & Liu, X. (2025). QAMT: An LLM-Based Framework for Quality-Assured Medical Time-Series Data Generation. Sensors, 25(17), 5482. https://doi.org/10.3390/s25175482

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop