Next Article in Journal
Circulating Tumour Cells as Potential Biomarkers for Oral Squamous Cell Carcinoma
Previous Article in Journal
Characterization and Performance of Non-Activated Apricot Stone Powder for the Remediation of Zn2+-Rich Galvanizing Effluents
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

A Multitask Active Learning Framework with Probabilistic Modeling for Multi-Species Acute Toxicity Prediction

1
Shanghai Key Laboratory of Power Station Automation Technology, School of Mechatronics Engineering and Automation, Shanghai University, Shanghai 200444, China
2
School of Environmental and Chemical Engineering, Shanghai University, Shanghai 200444, China
3
School of Medicine, Shanghai University, Shanghai 200444, China
4
Academy of Military Medical Sciences, Beijing 100850, China
*
Authors to whom correspondence should be addressed.
These authors contributed equally to this work.
Molecules 2026, 31(7), 1144; https://doi.org/10.3390/molecules31071144
Submission received: 20 February 2026 / Revised: 17 March 2026 / Accepted: 26 March 2026 / Published: 30 March 2026

Abstract

Predicting acute toxicity across species is essential for early-stage drug safety evaluation. While recent efforts have primarily focused on improving predictive accuracy, they often fail to address two critical issues: the substantial divergence in toxicity mechanisms among different species, and the inherent noise present in experimental data. To bridge this gap, we introduce a Probabilistic Multitask Active Learning (PMAL) framework for multi-species acute toxicity prediction. Our framework integrates two key modules: a Probabilistic Multitask Learning (PML) component which jointly models the predictive distributions of multiple toxicity endpoints from a probabilistic viewpoint, and an Uncertainty-based Active Learning (UAL) component which strategically selects the most informative compounds for experimental annotation based on predictive uncertainty. Empirical evaluations demonstrate that PMAL surpasses state-of-the-art methods and is capable of providing well-calibrated uncertainty estimates for small molecules across diverse toxicity endpoints. Beyond advancing multi-species toxicity prediction, the core design principles of PMAL offer a generalizable paradigm for learning in noisy multi-task environments.

1. Introduction

Acute toxicity is defined as adverse physiological effects arising from a single or repeated exposure to exogenous chemical compounds within a 24-h period in humans or experimental animals. Notably, toxicity-related attrition accounts for over 30% of candidate drug failures during early-stage pharmaceutical development [1,2], underscoring the critical need for reliable in silico acute toxicity prediction [3,4]. Robust predictive models enable efficient prioritization of compound libraries by identifying high-risk molecules prior to costly and time-consuming in vitro and in vivo assays, thereby reducing development costs, accelerating lead optimization, and improving overall attrition rates. Nevertheless, accurate and generalizable acute toxicity prediction remains challenging due to data scarcity across species, assay variability, structural diversity of chemical space, and intrinsic biological complexity.
Small-molecule toxicity frequently exhibits pronounced species specificity, i.e., marked inter-species variation in toxicological response profiles, including differences in target engagement, metabolic activation, detoxification pathways, and downstream physiological outcomes. Such variability critically undermines the reliability of safety extrapolation from preclinical animal studies to human clinical trials and imposes stringent demands on the robustness and transferability of computational toxicity prediction models [5]. Consequently, conventional Single-Task Learning (STL) approaches (e.g., training independent models for each species) are inherently limited in their capacity for cross-species generalization. To address this limitation, Multi-Task Learning (MTL) has been increasingly adopted in toxicological modeling. By jointly optimizing predictions across multiple species-specific toxicity endpoints under a shared representational framework, MTL improves predictive accuracy, calibration, and out-of-distribution generalization.
Current MTL approaches [6] fall broadly into two methodological paradigms: (i) architectural design, where task relationships are encoded via shared and task-specific neural modules; and (ii) optimization strategy, where training dynamics are controlled to balance competing objectives. Architectural methods commonly employ encoder–decoder frameworks in which a shared encoder extracts cross-species invariant representations while task-specific decoders model species-dependent toxicological responses. In contrast, optimization-based methods explicitly manage gradient conflicts or loss imbalances to promote equitable learning across toxicity endpoints. Despite these advances, a fundamental bottleneck persists in that MTL performance remains critically constrained by the scarcity and uneven distribution of high-quality and experimentally validated toxicity labels across species, limiting both model calibration and out-of-distribution generalization.
Generating high-quality toxicity annotations requires resource-intensive biological assays, entailing substantial financial cost, prolonged experimental timelines, and rigorous ethical oversight. Consequently, large-scale, consistently curated, and species-balanced toxicity datasets remain scarce. To mitigate this data bottleneck, Active Learning (AL) has emerged as a principled strategy for synergistic model–data co-development in toxicology. Rather than random or passive sampling, AL iteratively selects compounds with maximal information for targeted experimental validation, thereby maximizing knowledge gain per assay and enabling high-accuracy modeling with significantly reduced labeling effort.
Current AL selection strategies [7] fall into two principal categories: uncertainty-based and diversity-based. Uncertainty-based methods prioritize compounds for which the model exhibits high predictive uncertainty, under the premise that such instances are most likely to refine decision boundaries and reveal model deficiencies. In contrast, diversity-based methods seek to maximize coverage of the underlying chemical space by selecting samples that are both representative and minimally redundant relative to the already-labeled set, ensuring robust generalization. A critical limitation of both paradigms is their reliance on deterministic deep learning models, which fail to distinguish between epistemic uncertainty (model ignorance, reducible via data) and aleatoric uncertainty (inherent noise, e.g., assay variability, inter-laboratory discrepancies, or biological stochasticity). Consequently, these approaches cannot reliably quantify or correct for toxicity annotation bias arising from heterogeneous experimental conditions.
In real-world small-molecule toxicity assessment, experimental measurements for the same compound at a given toxicity endpoint frequently exhibit inter-laboratory and intra-assay variability [8]. Deterministic models produce point predictions without uncertainty quantification, making them inherently vulnerable to such label noise and consequently undermining their reliability and credibility. To address this fundamental limitation, probabilistic modeling has emerged as a principled framework for robust toxicity prediction. Unlike deterministic models, probabilistic models explicitly parameterize the full predictive distribution, which not only mitigates bias induced by noisy or inconsistent toxicity labels but also supports calibrated and risk-aware inference for decision-critical toxicological assessment.
Existing probabilistic modeling approaches (e.g., Monte Carlo dropout [9]) rely on repeated stochastic forward passes to approximate the predictive posterior distribution. This “posterior sampling” paradigm, while conceptually straightforward, imposes substantial computational overhead and increases model deployment complexity, particularly in resource-constrained or real-time settings. To overcome these limitations, we propose a direct parameterization framework that explicitly learns the parameters of the predictive distribution (e.g., mixture weights, means, and variances) in a single forward pass, thereby enabling efficient, scalable, and fully differentiable uncertainty quantification. Furthermore, we adopt Gaussian Mixture Models (GMMs) [10] as the parametric form of the predictive distribution. GMMs offer superior flexibility for modeling heterogeneous acute toxicity data, naturally accommodating both irreducible experimental noise and systematic mechanistic differences across species. By explicitly representing multi-task uncertainty structures, our GMM-based formulation yields not only better-calibrated prediction intervals but also interpretable and mechanism-aware uncertainty estimates.
Building on the foregoing analysis, we propose the Probabilistic Multitask Active Learning (PMAL) framework for multi-species acute toxicity prediction (Figure 1). The core model employs a Graph Neural Network (GNN)-based encoder–decoder architecture in which a shared graph encoder learns invariant molecular representations across species and toxicity endpoints while task-specific probabilistic decoders parameterize predictive distributions. Critically, PMAL jointly quantifies epistemic uncertainty and aleatoric uncertainty for each unlabeled compound. These uncertainties are weighted and aggregated into a unified query score. Compounds are then ranked by descending query score and the top-k compounds are selected for targeted experimental validation. Newly acquired labels are incrementally integrated into the training set in order to refine both representation learning and uncertainty calibration, enabling iterative improvement in predictive accuracy, robustness, and decision-relevant uncertainty quantification.
The contributions of this work are as follows:
  • We introduce the Probabilistic Multitask Active Learning (PMAL) framework tailored for multi-species acute toxicity prediction, which unifies multi-task learning and active learning to jointly address data scarcity and label noise in toxicological modeling.
  • We formulate a probabilistic multi-task model that explicitly accounts for both epistemic and aleatoric uncertainty; leveraging this dual uncertainty quantification, we further design an active learning query strategy that prioritizes high-uncertainty compounds for experimental labeling.
  • We conduct comprehensive experiments across 59 diverse acute toxicity endpoints spanning multiple species, demonstrating consistent improvements over established state-of-the-art methods under realistic low-labeling-budget settings.

2. Related Work

Early toxicity research predominantly employed statistical learning methods to develop Quantitative Structure–Activity Relationship (QSAR) models such as linear regression [11], support vector machine [12], and random forest [13]. The advent of deep learning catalyzed a paradigm shift, with deep neural networks now routinely integrated into QSAR modeling. A representative advance is the Deep Learning Consensus Architecture (DLCA) [14], which synergistically unifies consensus modeling and MTL to scale QSAR prediction across diverse endpoints. Subsequent studies [15,16] systematically benchmarked multi-task toxicity prediction across broad chemical spaces, varying both molecular descriptors and algorithmic frameworks. Crucially, progress has accelerated not only in model sophistication but also in data infrastructure and biological relevance; recent efforts have established large-scale and high-quality toxicity data platforms [17,18], adopted more advanced methods [19], and conducted organ-specific toxicity assessment grounded in human-relevant biology [20].
MTL and AL have advanced rapidly in computer vision and natural language processing, and their adoption in computational toxicology is now accelerating. In MTL, researchers have developed novel neural architectures that jointly predict multiple tasks [21,22,23,24], significantly enhanced by integration with GNNs [25,26,27] to exploit molecular topology for unified structure-aware representation learning. Critically, these efforts confront the persistent challenge of task imbalance, in which divergent loss magnitudes and convergence rates across tasks can induce gradient conflict and degrade shared representation quality. To mitigate this, recent work has introduced principled optimization strategies such as dynamic loss weighting [28,29,30] and gradient surgery techniques that orthogonalize task-specific gradients [31,32,33], leading to improved training stability and cross-task generalization. In AL, the high cost of experimental annotation has driven intensive innovation in query strategies. While uncertainty-based selection remains dominant, e.g., entropy-maximization [34,35,36] or loss-guided identification of ambiguous samples [37,38], diversity-promoting approaches [39,40] are increasingly recognized for their ability to ensure broad coverage of chemical space, enabling more efficient model refinement under stringent labeling budgets.

3. Results

3.1. Experimental Settings

3.1.1. Dataset

We evaluated the effectiveness of PMAL using the multi-species acute toxicity dataset ToxAcute [41]. This dataset comprises 120,384 experimentally measured data points across 59 distinct toxicity endpoints, each defined by a unique combination of species (mouse, rat, mammal, guinea pig, rabbit, dog, cat, wild bird, quail, duck, chicken, frog, and human), administration route (intraperitoneal, intravenous, oral, transdermal, subcutaneous, intramuscular, and non-intestinal), and toxicity metric (median lethal dose (LD50), lowest observed lethal dose (LDLo), and lowest observed toxic dose (TDLo)). For each endpoint, we performed scaffold split into training, validation, and test sets at an 8:1:1 ratio to ensure structural independence. To stabilize model optimization and improve numerical conditioning, all dose values were transformed to their negative logarithm in units of mol/kg (i.e., −log(mol/kg)). Under this transformation, higher values correspond to greater toxicity for a given compound–endpoint pair.

3.1.2. Metrics

We evaluate model performance using the Root Mean Squared Error (RMSE) and the Coefficient of Determination (R2):
  • Root Mean Squared Error (RMSE): Quantifies the average magnitude of prediction errors in the target variable. It is defined as follows:
    RMSE = 1 n i = 1 n ( y i y ^ i ) 2 ,
    where n is the total number of samples, y i denotes the true value, and y ^ i denotes the predicted value for the i-th sample.
  • Coefficient of Determination (R2): Measures the proportion of variance in the observed outcomes explained by the model, computed as follows:
    R 2 = 1 i = 1 n ( y i y ^ i ) 2 i = 1 n ( y i y ¯ ) 2 ,
    where y ¯ = 1 n i = 1 n y i is the mean of the observed values.

3.1.3. Training Details

All experiments were conducted on a single NVIDIA RTX 4090 GPU. We implemented the training and inference pipeline using PyTorch 1.8.1 and PyTorch Geometric 2.2.0, and performed molecular processing using RDKit 2022.3.3. Each molecule was represented as an undirected graph in which node features were encoded as a 96-dimensional vector comprising the following: atomic number (44 dimensions), atom degree (17 dimensions), implicit hydrogen count (17 dimensions), valence (17 dimensions), and aromaticity flag (1 dimension). For each endpoint, the AL protocol began with an initial labeled set containing 10% of labeled samples; three iterative rounds followed, each querying the top 10% of unlabeled pool based on annotation utility. Within each round, models were trained for 50 epochs with a batch size of 64, using the Adam optimizer (learning rate 1 × 10−3, weight decay 1 × 10−5). To ensure robustness, all experiments were repeated three times, and results are reported as mean ± standard deviation.

3.2. Comparative Experiments

To evaluate PMAL, we benchmarked it against a comprehensive set of established baselines on the ToxAcute dataset under two learning paradigms: MTL and AL. For MTL, we compared PMAL with three GNN-based approaches (MTGCN [25], MTGAT [26], and MTGIN [27]), each representing distinct architectural inductive biases for molecular representation learning. Under AL, we evaluated PMAL against five representative strategies (Random, Entropy [34], CoreSet [39], ProbCover [40], and TiDAL [38]). Model performance is assessed using both RMSE and R2, with each reported value representing the mean across all 59 toxicity endpoints.

3.2.1. Multi-Task Learning Setting

The comparative performance of PMAL against representative MTL methods is summarized in Table 1 and Table 2. As shown in the results, PMAL achieved state-of-the-art performance on 6 out of 14 evaluation tasks (including 13 individual species and their averaged score), surpassing all other MTGNNs by a substantial margin. Moreover, model performance across species exhibited a positive correlation with the volume of available labeled data, confirming the expected data scalability behavior. Notably, PMAL achieved the highest performance among all compared methods on several particularly challenging species (e.g., guinea pig, cat, and frog) which are characterized by limited annotated samples. This consistent superiority on low-resource species strongly supports PMAL’s enhanced generalization capability under small-sample conditions.

3.2.2. Active Learning Setting

The comparative performance of PMAL against representative AL methods is summarized in Table 3 and Table 4. As shown in the results, Entropy underperformed even Random sampling, highlighting its limited utility for acute toxicity prediction in this multi-endpoint setting. In contrast, diversity-based approaches (CoreSet and ProbCover) consistently outperformed uncertainty-based ones (Entropy and TiDAL), indicating that structural coverage of the chemical space is more informative than prediction variance alone for this task. While PMAL exhibited marginally lower initial performance relative to these baselines, it demonstrated consistent improvement across successive AL rounds and ultimately achieved superior final performance. Crucially, PMAL integrates dual-source uncertainty estimation, a capability that is absent in existing diversity-based methods.

3.3. Effectiveness Analysis

This section empirically evaluates PMAL along three complementary dimensions: (i) its probabilistic modeling capability, specifically whether the uncertainty estimates it produces are well calibrated and meaningfully reflect true prediction uncertainty; (ii) its MTL efficacy, assessing whether joint optimization yields performance competitive with or superior to that of the STL setting; and (iii) the effectiveness of its active sample annotation, investigating whether the instances selected by PMAL consistently drive measurable improvements in downstream model performance.

3.3.1. Probability Modeling

To evaluate PMAL’s probabilistic modeling capability across diverse toxicity prediction tasks, we monitored the Negative Log-Likelihood (NLL) over training epochs; the results are presented in Figure 2. For the majority of species, PMAL’s NLL consistently decreases with training progress, demonstrating robust and well-calibrated probabilistic predictions on multi-species acute toxicity data. In contrast, NLL increases for a small subset of species (e.g., duck and frog) indicating task conflict during multi-task optimization. The performance gains on dominant tasks appear to come at the expense of predictive calibration for these underrepresented or structurally dissimilar species. Such tradeoffs reflect a known challenge in MTL, particularly when task objectives exhibit misaligned gradients or imbalanced data coverage. Nevertheless, PMAL achieves superior overall performance compared to competing methods (Table 1 and Table 2), confirming its effectiveness in multi-task uncertainty modeling.

3.3.2. Multi-Task Training

To assess whether multi-task joint training yields superior generalization compared to single-task training, we conducted a per-species side-by-side comparison of RMSE performance, as presented in Figure 3. For most species, the multi-task model achieved significantly lower RMSE than its single-task counterpart, suggesting that acute toxicity profiles across these species share underlying structural or mechanistic regularities. These shared patterns enable effective knowledge transfer and cross-species regularization, thereby improving predictive robustness. Notably, however, the multi-task approach underperformed for a minority of species (e.g., frog and human). This degradation implies limited transferability between certain phylogenetically or physiologically distant species, highlighting a fundamental limitation in domain adaptation: when source and target distributions diverge substantially in feature space or biological mechanism, shared representations may introduce negative transfer. This phenomenon remains an active area of investigation in toxicological modeling and cross-species extrapolation.

3.3.3. Active Annotation

An effective query strategy should prioritize label-beneficial samples, that is, those with annotations that yield the largest marginal gain in model performance, rather than selecting samples based solely on task-irrelevant attributes. As shown in Figure 4 (Lower), we compute both Pearson and Spearman correlation coefficients between the uncertainty and the corresponding marginal performance gain (measured as the decrease/increase in RMSE/R2 metric after incorporating each labeled sample into the training set). The resulting coefficients range from 0.3 to 0.6, indicating a statistically meaningful and moderately positive monotonic association. Crucially, this correlation holds across two evaluation metrics, suggesting that higher uncertainty reliably identifies samples with high annotation utility.

3.4. Hyperparameter Analysis

This section analyzes two critical hyperparameters in PMAL: the number of components (C) in the PML, and the balance factor ( α ) in the UAL. Specifically, C determines the capacity of the Gaussian mixture model to capture uncertainty patterns across acute toxicity endpoints; meanwhile, α governs the tradeoff between epistemic and aleatory uncertainty in the composite uncertainty score.

3.4.1. Number of Components

To investigate the impact of the number of components (C) on the capacity of PML, we conducted an analysis over C { 2 , 4 , 8 , 16 } , evaluating mean RMSE and R2 across all 59 toxicity endpoints. The results are presented in Figure 4. As shown, the lowest RMSE is achieved at C = 8 , whereas the highest R2 occurs at C = 4 . Moreover, increasing C leads to a monotonic rise in the standard deviation of performance, reflecting growing sensitivity to training instability in mixture models. Given that R2 better reflects predictive reliability for multi-endpoint toxicity regression and that C = 4 achieves both competitive RMSE and markedly lower variance, we adopt C = 4 as the default setting.

3.4.2. Balanced Factor

To assess the influence of the balance factor ( α ) on the capacity of UAL, we performed an ablation study over α { 0.0 , 0.25 , 0.5 , 0.75 , 1.0 } , evaluating mean RMSE and R2 across all 59 toxicity endpoints (Figure 4). As shown, α = 1.0 (i.e., exclusive reliance on epistemic uncertainty) yields the worst performance. This degradation reflects the model’s inability to account for inherent noise within the outcome. Consequently, integrating both uncertainty sources is essential for robust sample selection. Among all configurations, α = 0.75 achieves the strongest performance, striking the best empirical tradeoff between error minimization and explained variance. We therefore adopt α = 0.75 as the default balance factor.

3.5. Visualizations

This section presents a three-fold visualization analysis of PMAL: (i) prediction credibility, (ii) species-level uncertainty distribution, and (iii) species-level embedding distributions. Each aspect is examined in turn through visualizations.

3.5.1. Prediction Visualizations

This subsection evaluates the credibility of PMAL’s predictions. Figure 5 presents scatter plots (left) and histograms (right) comparing model predictions against ground-truth values across species. As shown in the scatter plot, PMAL predictions exhibit alignment with the y = x line, with overall-level results outperforming species-level results. This gap mainly stems from the highly unbalanced distribution of toxicity data among different species. The histogram confirms strong distributional fidelity: the distributions of the predicted and ground-truth values exhibit high overlap, and their statistical distributions are highly similar.

3.5.2. Uncertainty Visualizations

This subsection characterizes the species-level uncertainty distribution estimated by PMAL. For each species, we aggregate uncertainty scores across all test samples and visualize the resulting distributions using Gaussian Kernel Density Estimation (KDE) [42] (Figure 5). We observe that uncertainty distributions correlate with both phylogenetic proximity and data coverage. Specifically, mouse and rat exhibit elevated mean uncertainty and broader dispersion, consistent with their extensive chemical space coverage. In contrast, bird and quail show nearly identical KDE curves, reflecting shared avian physiology and comparable assay protocols.

3.5.3. Representation Visualizations

This subsection examines the species-level embedding distributions learned by PMAL. Figure 6 shows t-SNE projections [43] of embeddings for thirteen species. The projections reveal pronounced inter-species divergence in latent geometry: high-sample species (e.g., mouse and rat) exhibit smooth, dose-ordered manifolds; by contrast, low-sample species (e.g., duck and frog) yield fragmented isotropic clouds with no discernible dose gradient, reflecting insufficient capacity for structured embedding formation. This stark contrast underscores a fundamental limitation in which representation quality in PMAL remains strongly contingent on per-species data volume, highlighting the persistent challenge of few-shot toxicological generalization.

4. Method

4.1. Notations

We formalize each acute toxicity endpoint (e.g., specific biological species, route of administration, and toxicity metric) as an individual learning task. Let T = { T 1 , T 2 , , T K } denote the set of K such tasks. For each task T k , we define three sets: a labeled set X k label , an unlabeled set X k unlabel , and a label set Y k . Every labeled instance x X k label is associated with a ground-truth label y Y k ; in contrast, instances x X k unlabel lack corresponding annotations. Crucially, each molecule x is represented as an undirected graph G = { V , E } in which vertices V correspond to atoms and edges E encode chemical bonds between them.
The MTL model f adopts an encoder–decoder architecture comprising a shared encoder f en and K task-specific decoders { f k de } k = 1 K . Its learnable parameters are partitioned into two disjoint sets: a shared parameter θ s governing the encoder f en , and task-specific parameters { θ k } k = 1 K , each associated with decoder f k de . Given a molecule x, the encoder first computes a task-invariant latent representation z = f en ( x ; θ s ) . This shared representation is then fed into the k-th decoder to yield the task-specific prediction y ^ k = f k de ( z ; θ k ) . The overall objective is to jointly optimize all parameters by minimizing the weighted sum of expected task losses:
min θ s , θ 1 , , θ K k = 1 K E { x , y } { X k label , Y k } [ L k ( y ^ k , y k ) ] ,
where L k is the loss function of task T k .
The AL is formulated as an iterative refinement process. At iteration t, the current multi-task model f scores every unlabeled molecule x X unlabel according to a labeling utility metric. The top-B samples with the highest utility are selected to make up the query set X t query . The labeled and unlabeled sets are then updated deterministically:
X t + 1 label = X t label X t query ,
X t + 1 unlabel = X t unlabel X t query ,
with this cycle repeating until either a predefined iterations is reached or the annotation budget is fully consumed.

4.2. PMAL Framework

This paper introduces PMAL, a probabilistic multi-task active learning framework for multi-species acute toxicity prediction, illustrated in Figure 1. PMAL integrates two synergistic components: (i) Probabilistic Multitask Learning (PML), which employs a shared graph encoder to extract invariant molecular representations and task-specific probabilistic decoders to model the full predictive distribution of each toxicity endpoint; and (ii) Uncertainty-based Active Learning (UAL), which jointly quantifies epistemic uncertainty and aleatoric uncertainty and fuses them into a calibrated overall uncertainty score. The active selection module then prioritizes unlabeled molecules by descending overall uncertainty and queries the top-B instances for expert annotation. By iteratively acquiring maximally informative samples under strict annotation budgets, PMAL achieves significant reductions in labeling effort while consistently improving out-of-distribution generalization across diverse species and endpoints.

4.3. Probabilistic Multi-Task Learning

PML establishes a principled graph-to-distribution mapping via a shared graph encoder and task-specific probabilistic decoders. The encoder extracts invariant molecular representations from the input graph, while each decoder transforms these representations into a well-calibrated predictive distribution. Critically, this architecture yields not only probabilistic predictions but also intrinsic quantifiable uncertainty estimates (e.g., epistemic and aleatoric uncertainty) directly encoded in the output distribution. As such, it provides a theoretically grounded and empirically actionable uncertainty signal that serves as the essential input to UAL.
The shared graph encoder consists of L consecutive Graph Convolutional Network (GCN) layers. Given a molecular graph G = { V , E } , it computes a molecule-level embedding z via iterative message passing across nodes followed by a Readout operation. Specifically, at layer l + 1 , the embedding h v ( l + 1 ) of node v V is updated as follows:
h v ( l + 1 ) = σ u N ( v ) { v } 1 deg ( v ) deg ( u ) h u ( l ) θ s ( l ) ,
where N ( · ) denotes the first-order neighborhood, deg ( · ) is the degree, h u ( l ) is the embedding of node u, θ s ( l ) is the learnable weight matrix, and σ ( · ) is the activation function. After L layers of message passing, the final node embeddings h v ( L ) are aggregated into a fixed-size molecular representation
z = READOUT h v ( L ) | v V ,
with this embedding z serving as the input to each task-specific probabilistic decoder.
Each task-specific probabilistic decoder comprises a Multi-Layer Perceptron (MLP) that maps the shared molecular embedding z to the parameters of a C-component Gaussian Mixture Model (GMM). For regression endpoints, the conditional distribution is defined as
p ( y ^ c | z ) = N ( y ^ c | μ c ( z ) , σ c 2 ( z ) ) ,
where μ c ( · ) and σ c 2 ( · ) respectively denote the mean and variance and y ^ c denotes the component-wise prediction. For classification endpoints, we adopt a stochastic softmax formulation grounded in the reparameterization trick:
p ( y ^ c | z ) = Softmax μ c ( z ) + σ c 2 ( z ) · γ ,
where γ follows the standard Gaussian distribution. The loss for the k-th toxicity endpoint is defined as
L k = c = 1 C π c ( z ) log p ( y ^ c | z ) ,
with π c ( · ) denoting the mixture weight subject to Σ c = 1 C π c = 1 .

4.4. Uncertainty-Based Active Learning

Upon completion of PML training, the model’s learned probabilistic outputs provide a principled basis for quantifying sample-level uncertainty [44]. Specifically, uncertainty is decomposed into the two components of epistemic uncertainty and aleatoric uncertainty. Such decomposition enables targeted selection of highly informative samples while explicitly distinguishing between uncertainty that is reducible through labeling (epistemic) and uncertainty that is inherent to the measurement process (aleatoric).
For regression endpoints, aleatoric uncertainty is computed as the mixture-weighted average of component-wise variances, while epistemic uncertainty is quantified as the mixture-weighted variance of the differences between the component-wise means and overall-wise mean:
U al = c = 1 C π c · σ c 2 ,
U ep = c = 1 C π c · μ c c = 1 C π c · μ c 2 .
For classification endpoints, we estimate aleatoric and epistemic uncertainty via Monte Carlo (MC) sampling over the learned Gaussian mixture model. First, a component is sampled based on the mixed weights, then an independent sample is drawn from the corresponding predictive distribution of that component. This process is repeated T times to obtain the predictive probability sequence { p t } t = 1 T . Aleatory uncertainty is expressed as the expectation of the predictive entropy of the sampling sequence, and epistemic uncertainty can be obtained by subtracting aleatory uncertainty from the total uncertainty:
U al = 1 T t = 1 T [ p t log p t ( 1 p t ) log ( 1 p t ) ] ,
U ep = U total U al ,
where U total = p ¯ log p ¯ ( 1 p ¯ ) log ( 1 p ¯ ) is the total uncertainty, reflecting the model’s overall confidence in the final class assignment, and p ¯ = 1 T Σ t = 1 T p t is the average of multiple sampling prediction probabilities.
Finally, we define the annotation utility of each unlabeled sample as a combination of the epistemic and aleatoric uncertainties:
U = α · U ep + ( 1 α ) · U al ,
where higher values of U indicate greater informativeness. The unlabeled pool is then sorted in descending order and the top-B samples are selected for expert annotation.

5. Conclusions

In this study, we introduce Probabilistic Multitask Active Learning (PMAL), a unified framework designed to tackle two fundamental challenges in multi-species acute toxicity prediction: (i) pronounced interspecies divergence in toxicity response patterns, and (ii) irreducible experimental noise inherent to biological assays. PMAL synergistically integrates Probabilistic Multitask Learning (PML) and Uncertainty-based Active Learning (UAL). PML jointly models the predictive distributions across 59 toxicity endpoints, while UAL leverages this calibrated uncertainty to iteratively prioritize compounds with maximal expected information gain for experimental validation. Empirical evaluation demonstrates that PMAL achieves state-of-the-art prediction accuracy and reduces required experimental effort. Critically, the PMAL paradigm is broadly applicable to any domain in which label scarcity, measurement noise, and cross-task structure coexist, including environmental risk assessment and preclinical safety profiling.

Author Contributions

Conceptualization, T.H.; methodology, T.H. and J.W.; software, T.H. and Y.Z.; validation, J.W., Y.L., and L.Y.; formal analysis, J.W.; investigation, J.W. and Y.Z.; resources, X.B.; data curation, S.H.; writing—original draft preparation, T.H.; writing—review and editing, J.W. and Y.Z.; visualization, Y.Z. and Y.L.; supervision, S.H., P.Z., and X.B.; project administration, P.Z.; funding acquisition, Y.Z. and S.H. All authors have read and agreed to the published version of the manuscript.

Funding

This research was funded by the Natural Science Foundation of Shanghai (Yanpeng Zhao, grant number 25ZR1402171), the Key Technologies Research and Development Program (Song He, grant number 2023YFC2604400), and the National Natural Science Foundation of China (Song He, grant number 62573425).

Institutional Review Board Statement

Ethical review and approval were waived for this study due to all data being derived from publicly available datasets included in TOXRIC [41].

Informed Consent Statement

Patient consent was waived due to all data being derived from publicly available datasets included in TOXRIC [41].

Data Availability Statement

The data are available in https://toxric.bioinforai.tech/ (accessed on 18 November 2022).

Conflicts of Interest

The authors declare no conflicts of interest.

References

  1. Vo, A.H.; Van Vleet, T.R.; Gupta, R.R.; Liguori, M.J.; Rao, M.S. An overview of machine learning and big data for drug toxicity evaluation. Chem. Res. Toxicol. 2019, 33, 20–37. [Google Scholar] [CrossRef]
  2. Hua, Y.; Shi, Y.; Cui, X.; Li, X. In silico prediction of chemical-induced hematotoxicity with machine learning and deep learning methods. Mol. Divers. 2021, 25, 1585–1596. [Google Scholar] [CrossRef] [PubMed]
  3. Tran, T.T.V.; Surya Wibowo, A.; Tayara, H.; Chong, K.T. Artificial intelligence in drug toxicity prediction: Recent advances, challenges, and future perspectives. J. Chem. Inf. Model. 2023, 63, 2628–2643. [Google Scholar] [CrossRef] [PubMed]
  4. Zhang, J.; Li, H.; Zhang, Y.; Huang, J.; Ren, L.; Zhang, C.; Zou, Q.; Zhang, Y. Computational toxicology in drug discovery: Applications of artificial intelligence in ADMET and toxicity prediction. Briefings Bioinform. 2025, 26, bbaf533. [Google Scholar] [CrossRef] [PubMed]
  5. Pérez Santín, E.; Rodríguez Solana, R.; González García, M.; García Suárez, M.D.M.; Blanco Díaz, G.D.; Cima Cabal, M.D.; Moreno Rojas, J.M.; López Sánchez, J.I. Toxicity prediction based on artificial intelligence: A multidisciplinary overview. Wiley Interdiscip. Rev. Comput. Mol. Sci. 2021, 11, e1516. [Google Scholar] [CrossRef]
  6. Vandenhende, S.; Georgoulis, S.; Van Gansbeke, W.; Proesmans, M.; Dai, D.; Van Gool, L. Multi-task learning for dense prediction tasks: A survey. IEEE Trans. Pattern Anal. Mach. Intell. 2021, 44, 3614–3633. [Google Scholar] [CrossRef]
  7. Li, D.; Wang, Z.; Chen, Y.; Jiang, R.; Ding, W.; Okumura, M. A survey on deep active learning: Recent advances and new frontiers. IEEE Trans. Neural Netw. Learn. Syst. 2025, 36, 5879–5899. [Google Scholar] [CrossRef]
  8. Wu, Z.; Chen, J.; Li, Y.; Deng, Y.; Zhao, H.; Hsieh, C.Y.; Hou, T. From black boxes to actionable insights: A perspective on explainable artificial intelligence for scientific discovery. J. Chem. Inf. Model. 2023, 63, 7617–7627. [Google Scholar] [CrossRef]
  9. Gal, Y.; Ghahramani, Z. Dropout as a bayesian approximation: Representing model uncertainty in deep learning. In Proceedings of the International Conference on Machine Learning; PMLR; JMLR.org: Norfolk, MA, USA, 2016; pp. 1050–1059. [Google Scholar]
  10. Stauffer, C.; Grimson, W.E.L. Adaptive background mixture models for real-time tracking. In Proceedings of the 1999 IEEE Computer Society Conference on Computer Vision and Pattern Recognition (Cat. No PR00149); IEEE: Piscataway, NJ, USA, 1999; Volume 2, pp. 246–252. [Google Scholar]
  11. Toropov, A.A.; Rasulev, B.F.; Leszczynski, J. QSAR modeling of acute toxicity for nitrobenzene derivatives towards rats: Comparative analysis by MLRA and optimal descriptors. QSAR Comb. Sci. 2007, 26, 686–693. [Google Scholar] [CrossRef]
  12. Mazzatorta, P.; Cronin, M.T.; Benfenati, E. A QSAR study of avian oral toxicity using support vector machines and genetic algorithms. QSAR Comb. Sci. 2006, 25, 616–628. [Google Scholar] [CrossRef]
  13. Polishchuk, P.G.; Muratov, E.N.; Artemenko, A.G.; Kolumbin, O.G.; Muratov, N.N.; Kuz’min, V.E. Application of random forest approach to QSAR prediction of aquatic toxicity. J. Chem. Inf. Model. 2009, 49, 2481–2488. [Google Scholar] [CrossRef]
  14. Zakharov, A.V.; Zhao, T.; Nguyen, D.T.; Peryea, T.; Sheils, T.; Yasgar, A.; Huang, R.; Southall, N.; Simeonov, A. Novel consensus architecture to improve performance of large-scale multitask deep learning QSAR models. J. Chem. Inf. Model. 2019, 59, 4613–4624. [Google Scholar] [CrossRef]
  15. Sosnin, S.; Karlov, D.; Tetko, I.V.; Fedorov, M.V. Comparative study of multitask toxicity modeling on a broad chemical space. J. Chem. Inf. Model. 2018, 59, 1062–1072. [Google Scholar] [CrossRef]
  16. Jain, S.; Siramshetty, V.B.; Alves, V.M.; Muratov, E.N.; Kleinstreuer, N.; Tropsha, A.; Nicklaus, M.C.; Simeonov, A.; Zakharov, A.V. Large-scale modeling of multispecies acute toxicity end points using consensus of multitask deep learning methods. J. Chem. Inf. Model. 2021, 61, 653–663. [Google Scholar] [CrossRef] [PubMed]
  17. Daghighi, A.; Casanola-Martin, G.M.; Iduoku, K.; Kusic, H.; Gonzalez-Diaz, H.; Rasulev, B. Multi-endpoint acute toxicity assessment of organic compounds using large-scale machine learning modeling. Environ. Sci. Technol. 2024, 58, 10116–10127. [Google Scholar] [CrossRef] [PubMed]
  18. Fu, L.; Shi, S.; Yi, J.; Wang, N.; He, Y.; Wu, Z.; Peng, J.; Deng, Y.; Wang, W.; Wu, C.; et al. ADMETlab 3.0: An updated comprehensive online ADMET prediction platform enhanced with broader coverage, improved performance, API functionality and decision support. Nucleic Acids Res. 2024, 52, W422–W431. [Google Scholar] [CrossRef] [PubMed]
  19. Gao, J.; Wu, L.; Lin, G.; Zou, J.; Yan, B.; Liu, K.; He, S.; Bo, X. Multi-task multi-view and iterative error-correcting random forest for acute toxicity prediction. Expert Syst. Appl. 2025, 274, 126972. [Google Scholar] [CrossRef]
  20. Wu, P.Y.; Chou, W.C.; Kamineni, V.N.; Chen, C.Y.; Hsieh, J.H.; Vulpe, C.D.; Lin, Z. Development of machine learning-based multi-task quantitative structure–activity relationship models for predicting toxicities in six human organ systems. Comput. Toxicol. 2025, 37, 100399. [Google Scholar] [CrossRef]
  21. Misra, I.; Shrivastava, A.; Gupta, A.; Hebert, M. Cross-stitch networks for multi-task learning. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition; IEEE: Piscataway, NJ, USA, 2016; pp. 3994–4003. [Google Scholar]
  22. Ma, J.; Zhao, Z.; Yi, X.; Chen, J.; Hong, L.; Chi, E.H. Modeling task relationships in multi-task learning with multi-gate mixture-of-experts. In Proceedings of the 24th ACM SIGKDD International Conference on Knowledge Discovery & Data Mining; Association for Computing Machinery: New York, NY, USA, 2018; pp. 1930–1939. [Google Scholar]
  23. Liu, S.; Johns, E.; Davison, A.J. End-to-end multi-task learning with attention. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition; IEEE: Piscataway, NJ, USA, 2019; pp. 1871–1880. [Google Scholar]
  24. Ding, C.; Lu, Z.; Wang, S.; Cheng, R.; Boddeti, V.N. Mitigating task interference in multi-task learning via explicit task routing with non-learnable primitives. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition; IEEE: Piscataway, NJ, USA, 2023; pp. 7756–7765. [Google Scholar]
  25. Kipf, T. Semi-supervised classification with graph convolutional networks. In Proceedings of the International Conference on Learning Representations; Curran Associates, Inc.: Red Hook, NY, USA, 2017; pp. 1–14. [Google Scholar]
  26. Veličković, P.; Cucurull, G.; Casanova, A.; Romero, A.; Lio, P.; Bengio, Y. Graph attention networks. In Proceedings of the International Conference on Learning Representations; Curran Associates, Inc.: Red Hook, NY, USA, 2018; pp. 1–12. [Google Scholar]
  27. Xu, K.; Hu, W.; Leskovec, J.; Jegelka, S. How powerful are graph neural networks? In Proceedings of the International Conference on Learning Representations; Curran Associates, Inc.: Red Hook, NY, USA, 2019; pp. 1–17. [Google Scholar]
  28. Kendall, A.; Gal, Y.; Cipolla, R. Multi-task learning using uncertainty to weigh losses for scene geometry and semantics. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition; IEEE: Piscataway, NJ, USA, 2018; pp. 7482–7491. [Google Scholar]
  29. Chen, Z.; Badrinarayanan, V.; Lee, C.Y.; Rabinovich, A. Gradnorm: Gradient normalization for adaptive loss balancing in deep multitask networks. In Proceedings of the International Conference on Machine Learning; PMLR; JMLR.org: Norfolk, MA, USA, 2018; pp. 794–803. [Google Scholar]
  30. He, Y.; Zhou, S.; Zhang, G.; Yun, H.; Xu, Y.; Zeng, B.; Chilimbi, T.; Zhao, H. Robust multi-task learning with excess risks. In Proceedings of the International Conference on Machine Learning; PMLR; JMLR.org: Norfolk, MA, USA, 2024; pp. 1–21. [Google Scholar]
  31. Yu, T.; Kumar, S.; Gupta, A.; Levine, S.; Hausman, K.; Finn, C. Gradient surgery for multi-task learning. Adv. Neural Inf. Process. Syst. 2020, 33, 5824–5836. [Google Scholar]
  32. Liu, B.; Liu, X.; Jin, X.; Stone, P.; Liu, Q. Conflict-averse gradient descent for multi-task learning. Adv. Neural Inf. Process. Syst. 2021, 34, 18878–18890. [Google Scholar]
  33. Wang, Z.; Tsvetkov, Y.; Firat, O.; Cao, Y. Gradient vaccine: Investigating and improving multi-task optimization in massively multilingual models. In Proceedings of the International Conference on Learning Representations; Openreview.net: Newton Highlands, MA, USA, 2021; pp. 1–22. [Google Scholar]
  34. Wang, D.; Shang, Y. A new active labeling method for deep learning. In Proceedings of the International Joint Conference on Neural Networks; IEEE: Piscataway, NJ, USA, 2014; pp. 112–119. [Google Scholar]
  35. Gal, Y.; Islam, R.; Ghahramani, Z. Deep bayesian active learning with image data. In Proceedings of the International Conference on Machine Learning; PMLR; JMLR.org: Norfolk, MA, USA, 2017; pp. 1–10. [Google Scholar]
  36. Kirsch, A.; Van Amersfoort, J.; Gal, Y. Batchbald: Efficient and diverse batch acquisition for deep bayesian active learning. Adv. Neural Inf. Process. Syst. 2019, 32, 7026–7037. [Google Scholar]
  37. Yoo, D.; Kweon, I.S. Learning loss for active learning. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition; IEEE: Piscataway, NJ, USA, 2019; pp. 93–102. [Google Scholar]
  38. Kye, S.M.; Choi, K.; Byun, H.; Chang, B. TiDAL: Learning training dynamics for active learning. In Proceedings of the IEEE/CVF International Conference on Computer Vision; IEEE: Piscataway, NJ, USA, 2023; pp. 22335–22345. [Google Scholar]
  39. Sener, O.; Savarese, S. Active learning for convolutional neural networks: A core-set approach. In Proceedings of the International Conference on Learning Representations; Curran Associates, Inc.: Red Hook, NY, USA, 2018; pp. 1–13. [Google Scholar]
  40. Yehuda, O.; Dekel, A.; Hacohen, G.; Weinshall, D. Active learning through a covering lens. Adv. Neural Inf. Process. Syst. 2022, 35, 22354–22367. [Google Scholar]
  41. Wu, L.; Yan, B.; Han, J.; Li, R.; Xiao, J.; He, S.; Bo, X. TOXRIC: A comprehensive database of toxicological data and benchmarks. Nucleic Acids Res. 2023, 51, D1432–D1445. [Google Scholar] [CrossRef]
  42. Parzen, E. On estimation of a probability density function and mode. Ann. Math. Stat. 1962, 33, 1065–1076. [Google Scholar] [CrossRef]
  43. Van der Maaten, L.; Hinton, G. Visualizing data using t-SNE. J. Mach. Learn. Res. 2008, 9, 2579–2605. [Google Scholar]
  44. Choi, J.; Elezi, I.; Lee, H.J.; Farabet, C.; Alvarez, J.M. Active learning for deep object detection via probabilistic modeling. In Proceedings of the IEEE/CVF International Conference on Computer Vision; IEEE: Piscataway, NJ, USA, 2021; pp. 10264–10273. [Google Scholar]
Figure 1. The Probabilistic Multitask Active Learning (PMAL) framework comprises two components: (i) Probabilistic Multitask Learning (PML), which jointly models all toxicity endpoints using a shared representation encoder and endpoint-specific probabilistic decoders to output calibrated predictive distributions; and (ii) Uncertainty-based Active Learning (UAL), which leverages both epistemic and aleatoric uncertainty estimates from PML’s outputs to guide iterative information-maximizing compound selection for experimental validation-aware model calibration.
Figure 1. The Probabilistic Multitask Active Learning (PMAL) framework comprises two components: (i) Probabilistic Multitask Learning (PML), which jointly models all toxicity endpoints using a shared representation encoder and endpoint-specific probabilistic decoders to output calibrated predictive distributions; and (ii) Uncertainty-based Active Learning (UAL), which leverages both epistemic and aleatoric uncertainty estimates from PML’s outputs to guide iterative information-maximizing compound selection for experimental validation-aware model calibration.
Molecules 31 01144 g001
Figure 2. Probability modeling effectiveness analysis. Each subplot displays the evolution of the negative log-likelihood over training epochs, with a distinct color assigned to each species.
Figure 2. Probability modeling effectiveness analysis. Each subplot displays the evolution of the negative log-likelihood over training epochs, with a distinct color assigned to each species.
Molecules 31 01144 g002
Figure 3. Multi-task integration effectiveness analysis. Each subplot presents a side-by-side comparison of model performance under the MTL and STL paradigms, with a distinct color assigned to each species.
Figure 3. Multi-task integration effectiveness analysis. Each subplot presents a side-by-side comparison of model performance under the MTL and STL paradigms, with a distinct color assigned to each species.
Molecules 31 01144 g003
Figure 4. Hyperparameter and active annotation effectiveness analysis. (Upper): Component number (C) analysis. (Middle): Balance factor ( α ) analysis. (Lower): Active annotation effectiveness analysis.
Figure 4. Hyperparameter and active annotation effectiveness analysis. (Upper): Component number (C) analysis. (Middle): Balance factor ( α ) analysis. (Lower): Active annotation effectiveness analysis.
Molecules 31 01144 g004
Figure 5. Prediction and uncertainty visualization. (Upper): PMAL’s predicted toxicity values (grouped by species). (Lower): Species-level uncertainty estimates. Species are distinguished by distinct colors.
Figure 5. Prediction and uncertainty visualization. (Upper): PMAL’s predicted toxicity values (grouped by species). (Lower): Species-level uncertainty estimates. Species are distinguished by distinct colors.
Molecules 31 01144 g005
Figure 6. Representation visualizations, showing t-SNE visualization of PMAL representations across species. Toxicity values are encoded via a continuous color; darker hues correspond to lower toxicity, while lighter hues correspond to higher toxicity.
Figure 6. Representation visualizations, showing t-SNE visualization of PMAL representations across species. Toxicity values are encoded via a continuous color; darker hues correspond to lower toxicity, while lighter hues correspond to higher toxicity.
Molecules 31 01144 g006
Table 1. RMSE performance comparison under MTL settings. The best result is highlighted in bold and the second-best result is underlined.
Table 1. RMSE performance comparison under MTL settings. The best result is highlighted in bold and the second-best result is underlined.
SpeciesLabelsMTGCN [25]MTGAT [26]MTGIN [27]PMAL
Mouse86,2310.7795 ± 0.02040.7818 ± 0.02180.7825 ± 0.01030.7924 ± 0.0131
Rat21,7490.7255 ± 0.00890.7359 ± 0.01960.7356 ± 0.01570.7559 ± 0.0088
Mammal24530.7132 ± 0.00940.7114 ± 0.02730.6987 ± 0.00790.7062 ± 0.0412
Guinea Pig17430.9406 ± 0.02100.9293 ± 0.01170.9072 ± 0.01180.8091 ± 0.0385
Rabbit42830.7138 ± 0.01090.7352 ± 0.02090.7187 ± 0.01690.7807 ± 0.0092
Dog15340.7698 ± 0.03610.7639 ± 0.02410.7910 ± 0.02530.8361 ± 0.0165
Cat6931.1136 ± 0.03491.1623 ± 0.00371.0830 ± 0.05840.8833 ± 0.0828
Bird3290.6364 ± 0.03420.6654 ± 0.03070.6689 ± 0.04900.7325 ± 0.0539
Quail3490.7651 ± 0.07300.7605 ± 0.01570.7428 ± 0.04900.7026 ± 0.0635
Duck1861.1770 ± 0.02951.1528 ± 0.04621.1830 ± 0.12931.1836 ± 0.1012
Chicken3470.8196 ± 0.05940.9200 ± 0.01510.8702 ± 0.05740.7940 ± 0.0280
Frog1061.2954 ± 0.12031.2520 ± 0.05441.3084 ± 0.05851.2309 ± 0.0977
Human3811.0756 ± 0.09091.0534 ± 0.05901.1105 ± 0.03251.1618 ± 0.0224
Mean-0.8865 ± 0.04220.8941 ± 0.02690.8924 ± 0.04020.8745 ± 0.0444
Table 2. R2 performance comparison under MTL settings. The best result is highlighted in bold and the second-best result is underlined.
Table 2. R2 performance comparison under MTL settings. The best result is highlighted in bold and the second-best result is underlined.
SpeciesLabelsMTGCN [25]MTGAT [26]MTGIN [27]PMAL
Mouse86,2310.3059 ± 0.03680.2825 ± 0.03690.2913 ± 0.01650.2828 ± 0.0179
Rat21,7490.3879 ± 0.01820.3824 ± 0.03160.3787 ± 0.02510.3506 ± 0.0150
Mammal24530.1885 ± 0.03100.1961 ± 0.06770.2265 ± 0.02260.2328 ± 0.0916
Guinea Pig1743−0.1206 ± 0.0767−0.0297 ± 0.05650.0436 ± 0.04720.3235 ± 0.0739
Rabbit42830.3645 ± 0.04510.3325 ± 0.02760.3774 ± 0.00310.2685 ± 0.0480
Dog15340.4616 ± 0.05670.4658 ± 0.03330.4330 ± 0.03230.3520 ± 0.0191
Cat693−0.0676 ± 0.0717−0.1395 ± 0.0180−0.0031 ± 0.13870.3860 ± 0.1081
Bird3290.1811 ± 0.08910.1052 ± 0.08150.0939 ± 0.1336−0.0828 ± 0.1698
Quail3490.3744 ± 0.11570.3856 ± 0.02520.4122 ± 0.07600.4757 ± 0.1033
Duck1860.3352 ± 0.03350.3618 ± 0.05160.3233 ± 0.14250.3279 ± 0.1243
Chicken3470.2817 ± 0.10470.0979 ± 0.02970.1906 ± 0.10460.3275 ± 0.0477
Frog1060.2234 ± 0.14710.2779 ± 0.06340.2112 ± 0.07120.3028 ± 0.1191
Human381−0.1243 ± 0.1911−0.0813 ± 0.0960−0.2018 ± 0.0870−0.1384 ± 0.0196
Mean-0.2148 ± 0.07830.2029 ± 0.04760.2136 ± 0.06930.2622 ± 0.0736
Table 3. RMSE performance comparison under AL settings. The best result is highlighted in bold and the second-best result is underlined.
Table 3. RMSE performance comparison under AL settings. The best result is highlighted in bold and the second-best result is underlined.
MethodRound 1Round 2Round 3Round 4p Value
Random0.9465 ± 0.01690.9065 ± 0.01390.8751 ± 0.00780.8637 ± 0.0087-
Entropy [34]0.9465 ± 0.01690.9180 ± 0.01370.8944 ± 0.01460.8917 ± 0.01560.0899
CoreSet [39]0.9465 ± 0.01690.8687 ± 0.00770.8418 ± 0.00890.8418 ± 0.01040.0705
ProbCover [40]0.9465 ± 0.01690.8585 ± 0.02390.8546 ± 0.02000.8504 ± 0.01760.1365
TiDAL [38]0.9465 ± 0.01690.8673 ± 0.01540.8656 ± 0.01670.8647 ± 0.01660.2938
PMAL0.9465 ± 0.01690.8741 ± 0.00820.8439 ± 0.00830.8277 ± 0.00960.0587
Table 4. R2 performance comparison under AL settings. The best result is highlighted in bold and the second-best result is underlined.
Table 4. R2 performance comparison under AL settings. The best result is highlighted in bold and the second-best result is underlined.
MethodRound 1Round 2Round 3Round 4p Value
Random0.0550 ± 0.03450.1226 ± 0.03230.1788 ± 0.01000.1968 ± 0.0121-
Entropy [34]0.0550 ± 0.03450.1188 ± 0.02150.1666 ± 0.02090.1699 ± 0.02220.1700
CoreSet [39]0.0550 ± 0.03450.1702 ± 0.03210.2258 ± 0.02240.2323 ± 0.02120.0622
ProbCover [40]0.0550 ± 0.03450.2158 ± 0.03710.2230 ± 0.03080.2326 ± 0.02390.1094
TiDAL [38]0.0550 ± 0.03450.2087 ± 0.02770.2131 ± 0.02970.2131 ± 0.02970.1646
PMAL0.0550 ± 0.03450.1759 ± 0.02000.2263 ± 0.01660.2514 ± 0.01660.0586
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Han, T.; Wang, J.; Zhao, Y.; Lin, Y.; Yu, L.; He, S.; Zan, P.; Bo, X. A Multitask Active Learning Framework with Probabilistic Modeling for Multi-Species Acute Toxicity Prediction. Molecules 2026, 31, 1144. https://doi.org/10.3390/molecules31071144

AMA Style

Han T, Wang J, Zhao Y, Lin Y, Yu L, He S, Zan P, Bo X. A Multitask Active Learning Framework with Probabilistic Modeling for Multi-Species Acute Toxicity Prediction. Molecules. 2026; 31(7):1144. https://doi.org/10.3390/molecules31071144

Chicago/Turabian Style

Han, Tianyu, Jingjing Wang, Yanpeng Zhao, Ying Lin, Lu Yu, Song He, Peng Zan, and Xiaochen Bo. 2026. "A Multitask Active Learning Framework with Probabilistic Modeling for Multi-Species Acute Toxicity Prediction" Molecules 31, no. 7: 1144. https://doi.org/10.3390/molecules31071144

APA Style

Han, T., Wang, J., Zhao, Y., Lin, Y., Yu, L., He, S., Zan, P., & Bo, X. (2026). A Multitask Active Learning Framework with Probabilistic Modeling for Multi-Species Acute Toxicity Prediction. Molecules, 31(7), 1144. https://doi.org/10.3390/molecules31071144

Article Metrics

Back to TopTop