Next Article in Journal
On the Relativity of Quantumness as Implied by Relativity of Arithmetic and Probability
Previous Article in Journal
The Hungry Daemon: Does an Energy-Harvesting Active Particle Have to Obey the Second Law of Thermodynamics?
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

A Graph Contrastive Learning Method for Enhancing Genome Recovery in Complex Microbial Communities

1
Department of Computer Science, Yangzhou University, Yangzhou, 225100, China
2
State Key Laboratory of Pharmaceutical Biotechnology, School of Life Sciences, Nanjing University, Nanjing 210023, China
*
Author to whom correspondence should be addressed.
Entropy 2025, 27(9), 921; https://doi.org/10.3390/e27090921
Submission received: 22 June 2025 / Revised: 20 August 2025 / Accepted: 28 August 2025 / Published: 31 August 2025
(This article belongs to the Special Issue Network-Based Machine Learning Approaches in Bioinformatics)

Abstract

Accurate genome binning is essential for resolving microbial community structure and functional potential from metagenomic data. However, existing approaches—primarily reliant on tetranucleotide frequency (TNF) and abundance profiles—often perform sub-optimally in the face of complex community compositions, low-abundance taxa, and long-read sequencing datasets. To address these limitations, we present MBGCCA, a novel metagenomic binning framework that synergistically integrates graph neural networks (GNNs), contrastive learning, and information-theoretic regularization to enhance binning accuracy, robustness, and biological coherence. MBGCCA operates in two stages: (1) multimodal information integration, where TNF and abundance profiles are fused via a deep neural network trained using a multi-view contrastive loss, and (2) self-supervised graph representation learning, which leverages assembly graph topology to refine contig embeddings. The contrastive learning objective follows the InfoMax principle by maximizing mutual information across augmented views and modalities, encouraging the model to extract globally consistent and high-information representations. By aligning perturbed graph views while preserving topological structure, MBGCCA effectively captures both global genomic characteristics and local contig relationships. Comprehensive evaluations using both synthetic and real-world datasets—including wastewater and soil microbiomes—demonstrate that MBGCCA consistently outperforms state-of-the-art binning methods, particularly in challenging scenarios marked by sparse data and high community complexity. These results highlight the value of entropy-aware, topology-preserving learning for advancing metagenomic genome reconstruction.

1. Introduction

Metagenomics enables the comprehensive characterization of microbial communities by directly sequencing the collective genomic content of environmental samples, providing a cultivation-independent approach to studying microbial diversity and function [1,2,3]. This methodology has become foundational across a wide range of disciplines, including human health, agriculture, food safety, and climate science [4,5,6]. In contrast to traditional culture-dependent techniques, metagenomics circumvents the limitations imposed by microbial cultivability, allowing for the discovery and functional annotation of a vast array of previously inaccessible microorganisms and their metabolic potential [7,8].
Over the past decades, the field of computational metagenomics has witnessed significant advancements, resulting in the development of key analytical tools such as metagenomic assembly [9], contig binning [10,11], and microbial feature profiling [12]. Among these, contig binning is particularly critical; it seeks to cluster DNA fragments derived from the same or phylogenetically related genomes into distinct bins. This step is essential for reconstructing microbial community structure, elucidating metabolic potential, and identifying novel microbial lineages with potential biotechnological value [13]. Nonetheless, the high complexity, sparsity, and heterogeneity characteristics [14] of environmental metagenomic datasets continue to pose substantial challenges to achieving both accurate and scalable genome binning.
To overcome the challenges of metagenomic binning, modern algorithms increasingly incorporate diverse features such as guanine–cytosine (GC) content, k-mer frequencies (i.e., the count or normalized occurrence of all possible nucleotide subsequences of length k), and coverage profiles, often employing multimodal data fusion strategies. Binning methods are typically categorized into three groups [15]: composition-based, abundance-based, and hybrid approaches. Among these, hybrid models generally achieve better performance by leveraging complementary signals across feature types. However, the effective integration of heterogeneous features remains a core challenge, particularly in capturing non-linear dependencies and ensuring scalability. Early methods such as CONCOCT [16] relied on dimensionality reduction (e.g., principal component analysis [17]) over composition and coverage features, followed by clustering via a Gaussian Mixture Model (GMM) [18]. MaxBin2 [19] adopted a probabilistic framework based on tetranucleotide frequencies and coverage, using an Expectation-Maximization [20] algorithm to infer genome assignments. MetaBAT2 [21] computed pairwise similarities between contigs using k-mer and coverage information and employed a modified Label Propagation Algorithm for clustering. These approaches laid the groundwork but often suffered from normalization sensitivity, limited flexibility, or shallow integration strategies. More recent methods aim to improve robustness and resolution. Binny [22] uses multiple k-mer profiles and coverage data in an iterative, non-linear dimensionality reduction pipeline, followed by clustering via HDBSCAN [23]. MetaDecoder [24] employs a two-phase approach: initial clustering with a modified Dirichlet Process Gaussian Mixture Model (DPGMM) [25], refined by a semi-supervised probabilistic model based on k-mer frequencies and a modified GMM. MetaBinner [26], in contrast, takes an ensemble learning approach, combining partial seed k-means across multiple feature types and integrating results using a two-stage ensemble framework. Despite these advances, many methods still fall short in capturing complex, high-order dependencies across features, particularly in high-dimensional biological data. Moreover, few frameworks incorporate information-theoretic principles—such as entropy or mutual information—which can provide theoretical guidance for feature selection, redundancy reduction, and improved representation learning.
In recent years, deep learning has markedly advanced metagenomic binning by enabling the extraction of informative representations from complex, high-dimensional data sources [27,28]. For instance, variational autoencoders (VAEs) have been employed to jointly embed k-mer composition and abundance features, while semi-supervised frameworks have incorporated auxiliary constraints to guide representation learning. Moreover, contrastive multi-view learning strategies [29] have improved the integration of heterogeneous modalities by aligning different feature views. However, despite these innovations, most existing methods rarely incorporate information-theoretic principles, such as mutual information maximization or entropy-based regularization, to explicitly control feature informativeness or redundancy. Furthermore, these approaches often neglect the intrinsic topological structure encoded in assembly graphs, which captures valuable contig–contig relationships, crucial for accurate genome binning [30].
To address the aforementioned challenges, and inspired by recent advances in graph representation learning and canonical correlation analysis [31,32], we propose MBGCCA (Metagenomic Binning using Graph-based Canonical Correlation Analysis), a novel framework that integrates graph neural networks (GNNs) [33] with contrastive learning to improve contig binning performance. MBGCCA explicitly models the assembly graph to capture structural dependencies between contigs, and jointly optimizes feature embeddings using a contrastive objective that encourages proximity between structurally or functionally related contigs while increasing separation among unrelated ones in the latent space. By harmonizing graph topology with multimodal feature correlation, MBGCCA not only improves binning performance in terms of accuracy, robustness, and generalization, but also implicitly incorporates information-theoretic regularization through contrastive learning and CCA-based embedding decorrelation. This promotes high-entropy, low-redundancy representations that more effectively capture the underlying genomic structure.

2. Materials and Methods

2.1. Computational Framework of MBGCCA

The computational framework of MBGCCA comprises two primary stages, as illustrated in Figure 1.
The MBGCCA framework operates in two sequential stages, each designed to address key challenges in metagenomic binning by integrating compositional, abundance, and structural information.
Stage one focuses on the fusion of two fundamental features: tetranucleotide frequency (TNF) and abundance profiles. Initially, assembled genome contigs are length-normalized through extension to ensure consistency in input size. TNF features capture sequence compositional patterns, while abundance profiles reflect sample-specific sequencing depth. These features are processed through a dedicated feature fusion network that generates high-dimensional embeddings, effectively encoding both structural and quantitative signals. This multimodal representation enhances the model’s ability to distinguish contigs from different microbial genomes based on both their intrinsic sequence characteristics and their relative abundance.
Stage two incorporates neighborhood context through a GNN module, which models the topological structure among contigs. To prevent overfitting and improve generalization, stochastic edge dropout is applied during training. Two perturbed graph views are generated and independently updated via message-passing mechanisms, allowing for local aggregation of neighborhood information. The resulting node embeddings are then aligned using CCA, which encourages maximum correlation between the two views. This contrastive alignment strategy facilitates the clustering of contigs from the same genome while promoting separation from unrelated contigs in the latent space.
By jointly optimizing multimodal feature integration and structure-aware representation learning, MBGCCA significantly improves binning performance. The final genome bins demonstrate enhanced biological coherence, benefiting from both topology-aware modeling and entropy-aware representation learning, which together enable robust and generalizable binning across diverse and complex microbial datasets. The overall framework jointly leverages contrastive learning and mutual information regularization in two stages: feature fusion and structural embedding. These learning objectives improve both the intra-contig consistency and inter-contig separation required for accurate genome binning.

2.2. Fusion of Tetranucleotide Frequency and Abundance Features

The initial step of the MBGCCA framework, illustrated in Figure 1A, focuses on multimodal feature integration through a feature fusion module. The overall process consists of four main components.

2.2.1. Data Augmentation

To enhance the robustness and generalizability of the model, we apply a Random Subsequence Extraction strategy. For each contig in the dataset, five additional augmented versions are generated by randomly cropping subsequences of varying lengths. This simulates the incomplete or fragmented nature of contigs that commonly arise in real-world metagenomic assemblies due to sequencing errors or low coverage. As a result, each original contig is expanded into six views: the original sequence and five augmented variants. These multiple views are later used to reinforce the model’s ability to generalize across contig length variability and handle short or noisy sequences.

2.2.2. TNF Calculation and Normalization

Following the methodology outlined by Teeling et al. [34] and Wang et al. [35], tetranucleotide frequencies (TNF) are calculated to construct compositional feature vectors for DNA contigs, providing a sequence-based signal that is complementary to coverage-based abundance features.
Step 1: K-mer Extraction. For each contig, a sliding window of size k = 4 with stride 1 bp is applied to generate all possible overlapping tetranucleotides. For a contig of length L i , this produces ( L i k + 1 ) 4-mers. For example, the sequence “ATGCATG” yields the 4-mers “ATGC”, “TGCA”, “GCAT”, and “CATG”.
Step 2: Reverse Complement (RC) Merging. Because DNA is double-stranded, a 4-mer and its reverse complement carry identical biological information. To eliminate strand bias, each 4-mer is merged with its reverse complement into a canonical representation. For instance, “ATGC” and “GCAT” are counted as a single canonical k-mer. Palindromic k-mers are self-complementary and thus counted once. This reduces the dimensionality from 256 raw 4-mers to 136 canonical 4-mers.
Step 3: Frequency Computation. Let N i , j denote the count of the j-th canonical 4-mer in the i-th contig. The raw relative frequency is computed as
f i , j = N i , j L i k + 1 ,
which normalizes by the total number of 4-mers in the contig, thereby correcting for length-dependent bias.
Step 4: Smoothing. To reduce sparsity and avoid zero-valued features (especially in shorter contigs or rare k-mers), Laplace smoothing [36] is applied:
f i , j = f i , j + 1 .
Step 5: Normalization. The smoothed frequencies are further normalized to form a probability distribution:
q i , j = f i , j k = 1 136   f i , k ,
ensuring that feature vectors are comparable across contigs, regardless of length or coverage.
Step 6: Vector Construction. The final TNF feature vector for contig i is defined as
x i ( com ) =   [ q i , 1 , q i , 2 , , q i , 136 ] ,
yielding a 136-dimensional representation of its compositional profile that is highly informative for downstream tasks such as genome binning, clustering, and embedding-based representation learning.

2.2.3. Calculation of Abundance Features

The abundance feature of a contig is a key indicator reflecting the relative occurrence of genomic fragments in sequencing samples. Unlike compositional features that are derived from intrinsic nucleotide patterns, abundance features are estimated from sequencing depth profiles obtained by mapping reads back to assembled contigs. This provides a complementary source of information, as contigs belonging to the same genome typically exhibit similar coverage patterns across samples.
Step 1: Read Mapping and Depth Profiling. For each sample, raw sequencing reads are aligned to the assembled contigs using a short-read aligner such as Bowtie2 or BWA. This produces alignment files (SAM/BAM) that record how many reads map to each base position of each contig. The per-base coverage profile is then aggregated to compute mean coverage and variability. Specifically, the mean coverage quantifies the overall abundance of a contig in one sample, while the standard deviation reflects unevenness of coverage caused by sequencing bias or repeat regions.
Step 2: Coverage Vector Construction. For contig i across M samples, two parallel descriptors are extracted: the mean coverage vector
C i ( mean ) = C i , 1 ( mean ) , C i , 2 ( mean ) , , C i , M ( mean ) ,
and the standard deviation coverage vector
C i ( std ) = C i , 1 ( std ) , C i , 2 ( std ) , , C i , M ( std ) ,
where C i , m ( mean ) and C i , m ( std ) denote the average and variability of coverage of contig i in the m-th sample, respectively. Concatenating these two descriptors yields the raw abundance feature vector:
C i = C i ( mean ) , C i ( std ) .
This representation explicitly encodes both abundance levels and their stability across samples.
Step 3: Handling Missing Coverage. Some contigs may not be detected in certain samples, leading to zero coverage. To avoid issues caused by sparsity and division by zero, a small constant ϵ is added:
C i , m = C i , m + ϵ , ϵ = 10 5 .
This smoothing step ensures numerical stability while preserving relative abundance differences.
Step 4: Cross-Sample Normalization. Because sequencing depth differs across samples, direct comparison of raw coverage values can be misleading. To ensure comparability, normalization is applied across all contigs within each sample:
C i , m ( cov ) = C i , m max k = 1 N C k , m ,
where N is the total number of contigs. This max-normalization rescales coverage values into [ 0 , 1 ] , making contigs comparable across samples with different sequencing depths.
Step 5: Final Feature Representation. After smoothing and normalization, the final abundance feature vector for contig i is defined as
x i ( abund ) = x i , 1 ( abund ) , x i , 2 ( abund ) ,   , x i , 2 M ( abund ) ,
which is 2 M -dimensional, combining both mean and variability descriptors for each of the M samples. This enriched representation captures not only the overall abundance profile of each contig but also its heterogeneity across sequencing datasets. When integrated with compositional TNF features, the abundance vector provides a robust foundation for downstream machine learning tasks such as clustering, genome binning, and multimodal representation learning.

2.2.4. Fusion Network Construction

The objective function guides the optimization of the MBGCCA neural network, with the core being the Normalized Temperature-scaled Cross-Entropy [37] (NT-Xent) loss function. This contrastive learning-based loss function optimizes instance representations by pulling similar instances closer and pushing dissimilar ones apart.
During training, different views of the same contig are treated as positive samples, while views from other contigs are treated as negative samples. Specifically, suppose that each contig has V different views in a batch of size N bs , then the NT-Xent loss is defined as
L NT Xent = 1 N bs V ( V 1 ) i = 1 N bs v = 1 V v 1 = 1 v 1 v V log exp cos ( z i , v , z i , v 1 ) τ j = 1 N bs v 2 = 1 V exp cos ( z i , v , z j , v 2 ) τ
Here, z i , v denotes the representation of the i-th contig under the v-th view, τ is the temperature parameter that controls the sensitivity of the contrastive learning, and cos ( a , b ) denotes the cosine similarity between two representations. The numerator captures the similarity between positive pairs (different views of the same contig), while the denominator aggregates similarities across all samples to form the contrastive objective.
Compared with traditional single-view contrastive learning, this multi-view approach enhances the robustness of binning by ensuring that different fragments of the same contig are correctly clustered together. It also avoids the information loss caused by relying on a single hand-crafted view. This objective enables MBGCCA to achieve accurate and stable contig classification in metagenomic binning. Moreover, the NT-Xent loss encourages high mutual information between positive pairs while implicitly reducing feature redundancy across views, serving as a form of information-theoretic regularization.

2.2.5. Training Details of Step 1

During training, TNF vectors and abundance feature vectors are used as inputs. A fusion network is employed for feature learning, consisting of three fully connected layers with 2048, 2048, and 128 units, designed to integrate multi-modal feature information and improve contig separability. The network is implemented in PyTorch v1.9.1 and trained over 200 epochs with a batch size of 1024. To prevent overfitting, early stopping is adopted, dynamically halting training based on validation loss and enhancing model generalization. The optimization target is the NT-Xent loss (Equation (11)), which improves binning performance by maximizing similarity between different views of the same contig (positive pairs) while minimizing similarity with other contigs (negative pairs).

2.3. Aggregation of Contig Neighborhood Information

The assembly graph, typically represented in GFA (Graphical Fragment Assembly) format, serves as the core structural input to the graph neural network (GNN) model in metagenomic binning tasks. This graph is generated during the de novo genome assembly process, where sequencing reads—especially from long-read or hybrid platforms—are first aligned and then progressively assembled into contiguous sequences (contigs). Rather than producing a linear genome sequence, modern assemblers (e.g., metaFlye [38], SPAdes [39]) construct an assembly graph to capture the inherent ambiguities and structural complexity present in metagenomic datasets, such as genomic repeats, strain-level variation, and incomplete coverage.
During graph initialization, each contig is treated as a node, and the edges denote observed connections (e.g., overlaps or scaffold links) between contigs c i and c j . The adjacency matrix A R N × N , where N is the number of contigs, is constructed accordingly. Each entry A i j reflects the edge weight between c i and c j , determined by the read coverage supporting their connection. If no direct linkage exists, A i j = 0 .
To mitigate the influence of sequencing depth variability and improve comparability across datasets, the edge weights in A are normalized. This normalization strategy ensures that connections supported by higher-confidence sequencing data exert proportionally greater influence during GNN training, while still controlling for sample-specific depth biases.
To further enhance the discriminative power of the GNN, MBGCCA adopts a GraphSAGE-based neighborhood sampling and aggregation strategy [40]. In this approach, each contig (node) updates its representation by sampling a fixed-size subset of its neighbors, rather than aggregating over all connected nodes. This makes the model computationally efficient and scalable to large assembly graphs. For each sampled neighborhood, MBGCCA applies mean aggregation to combine the features of neighboring contigs and integrates them with the central contig’s own features, enabling multi-hop message passing and capturing broader structural context.
Through message passing on the constructed graph, the GraphSAGE mechanism enables each contig node to iteratively update its representation by aggregating information from its local neighborhood. This process integrates sequence composition (e.g., TNF), abundance features, and topological context, resulting in robust and biologically coherent embeddings. Consequently, MBGCCA enhances binning accuracy, particularly in complex and heterogeneous metagenomic environments.

2.3.1. Self-Supervised Graph Representation Learning with CCA

In recent years, GNNs have demonstrated significant success in modeling graph-structured data, particularly in the domain of self-supervised learning [41]. While traditional contrastive learning frameworks operate at the instance level to distinguish between different samples, GNN-based methods can adopt a non-contrastive strategy, such as CCA, to efficiently learn meaningful graph representations without the need for negative sampling.
The essence of self-supervised learning lies in the generation of diverse views of the data, enabling the model to learn view-invariant and semantically consistent representations. In the context of graph data, we adopt two widely used augmentation strategies to construct such views:
  • Edge Dropping [42]: A subset of edges is randomly removed from the graph to simulate structural variations. This perturbation enhances the model’s ability to generalize to different graph topologies and encourages reliance on robust neighborhood structures.
  • Node Feature Masking [43]: Portions of node features are randomly masked, introducing uncertainty in node attributes and forcing the model to avoid over-dependence on specific dimensions. This strategy improves robustness and promotes the learning of more generalizable representations.
These augmented views are subsequently encoded via shared or parallel GNN encoders, and their representations are aligned using CCA-based objectives [44] to maximize inter-view correlation. This approach enables the model to capture structural and semantic consistency across perturbations, thereby yielding high-quality graph embeddings suitable for downstream tasks such as clustering or classification.
At each training iteration, two augmentations, t A and t B , are sampled from the transformation set T to generate two graph views:
G A = ( X A ,   A A ) and G B = ( X B ,   A B )
A shared-parameter GNN encoder is used to extract node embeddings. The core principle of the GNN is to update node representations using the adjacency structure (Figure 1B), defined as
Z = f θ ( X ,   A )
where X R N × F is the node feature matrix, N is the number of nodes, F is the feature dimension, A R N × N is the adjacency matrix, f θ is the GNN encoder with parameters θ , and  Z R N × D is the output embedding matrix with dimension D.
A two-layer Graph Convolutional Network (GCN) [45] is adopted as the encoder. The node representations are updated according to the following propagation rule:
H ( l + 1 ) = σ ( A ^ H ( l ) W ( l ) )
where H ( l ) is the feature matrix at layer l with H ( 0 ) = X , A ^ = D 1 / 2 A D 1 / 2 is the symmetric normalized adjacency matrix, W ( l ) is the trainable weight matrix at layer l, and  σ denotes the activation function (ReLU).
The final representations from the two augmented views are computed as
Z A = f θ ( X A ,   A A ) , Z B = f θ ( X B ,   A B )
To prevent numerical instability, node embeddings are standardized:
Z ˜ = Z μ ( Z ) σ ( Z ) × N
where μ ( Z ) and σ ( Z ) are the mean and standard deviation of the embedding matrix Z, ensuring zero mean and 1 / N standard deviation across each dimension.
Objective Function
The graph neural network employed in this work adopts CCA as its optimization objective. The loss function [44] is defined as
L CCA = Z ˜ A Z ˜ B F 2 + λ Z ˜ A Z ˜ A I F 2 + Z ˜ B Z ˜ B I F 2
The first term, Z ˜ A Z ˜ B F 2 , is the invariance loss [46], which minimizes the difference between embeddings from two augmented views, encouraging the model to learn view-invariant features.
The second term, Z ˜ A Z ˜ A I F 2 + Z ˜ B Z ˜ B I F 2 , is the decorrelation loss [44], designed to reduce redundancy between different feature dimensions and ensure that each dimension captures distinct information. This decorrelation mechanism is inspired by information-theoretic principles, aiming to maximize the entropy of learned representations by enforcing low mutual information among embedding dimensions. The hyperparameter λ controls the strength of the decorrelation term. Empirical results show that setting λ between 0.0005 and 0.001 yields optimal performance.

2.3.2. Training Details of Step 2

We set the training epochs to 100 for the self-supervised stage. The learning rates for the encoder and the linear classifier were set to 1 × 10−3 and 1 × 10−2, respectively. Weight decay was set to 0 for the encoder and 1 × 10−4 for the linear evaluator. The hidden and output dimensions of the encoder were both set to 512, and we used a two-layer GNN architecture. For graph augmentations, the drop edge ratio and drop feature ratio were both set to 0.2. All experiments were conducted using a single GPU unless otherwise specified. Upon convergence, the learned node embeddings capture robust structural and semantic patterns, making them suitable for a variety of downstream graph-based tasks, such as node classification, link prediction, and graph clustering.

2.4. Evaluation Metrics

In this study, bins (i.e., genome classification units) generated by GraphMB and other binning tools are evaluated using CheckM [47] (parameters: –reduced-tree, version 1.1.2) to assess both completeness and contamination. To eliminate redundancy, dereplication is performed using dRep [48]. During filtering, bins with completeness greater than 90% and contamination less than 5% are classified as high-quality (HQ) bins.
Additionally, for simulated datasets, AMBER [49] is employed to provide gold-standard evaluation by comparing predicted bins against reference labels. In this setting, several key metrics are used.

2.4.1. Average Purity (AP)

Purity measures the proportion of contig length in a bin that originates from a single reference genome. The average purity across all bins is defined as
AP = 1 B i = 1 B max j ( L i , j ) j L i , j ,
where B is the number of predicted bins, L i , j is the total length of contigs in bin i that map to reference genome j, and  max j ( L i , j ) corresponds to the dominant genome assignment for that bin. A higher AP indicates lower contamination.

2.4.2. Average Completeness (AC)

Completeness reflects the proportion of a reference genome recovered in binning. The average completeness is defined as
AC = 1 R j = 1 R i L i , j G j ,
where R is the total number of reference genomes, i L i , j is the sum of contig lengths assigned to genome j across bins, and  G j is the total genome length. A higher AC indicates that bins better capture the full genome content.

2.4.3. Contamination

In the context of AMBER, contamination is related to purity and defined as
Contamination = 1 Purity .

2.4.4. F1 Score

To capture a balance between purity and completeness, the F1 score is calculated for each bin as
F 1 i = 2 · Purity i · Completeness i Purity i + Completeness i ,
and the overall score is averaged across all bins. This metric emphasizes bins that achieve both high purity and high completeness simultaneously.

2.4.5. High-Quality (HQ) Bins

Following the MIMAG standard, bins are defined as HQ if they satisfy
Completeness 90 % , Contamination 5 % .
The number of HQ bins is reported as a practical summary of binning performance.

2.4.6. Dereplication and Uniqueness

During dereplication, dRep clusters highly similar bins (based on average nucleotide identity, ANI) into bin clusters. We define a bin as a unique HQ bin if it does not share a cluster with other HQ bins. This avoids redundancy and ensures the fair evaluation of distinct genome recovery.

2.4.7. Consensus Integration

Finally, to integrate the strengths of different binning tools, DASTool [50] is used to combine bins from all methods into a consolidated set, which improves both the quality and reliability of the final binning results.

2.5. Datasets

2.5.1. Pretraining Datasets

In this study, we aim to pretrain our model to improve the representation capacity of contigs, reduce reliance on manual feature engineering, and enhance binning performance for low-coverage and short-fragment contigs. Pretraining also enables the model to better capture underlying genomic patterns, thereby increasing its adaptability to diverse metagenomic datasets, especially in the context of uncultured microbial binning, ultimately boosting generalization, robustness, and efficiency.
We selected six datasets for training, including four simulated datasets and two real-world datasets. The four simulated datasets were provided by the CAMI II [51] challenge organizers (https://cami-challenge.org) (accessed on 20 August 2025), covering diverse microbial ecosystems: CAMI Mouse Gut (mouse gut microbiome), CAMI Skin (human skin microbiome), CAMI Airways (human airway microbiome), CAMI Gastrointestinal Tract (abbreviated as CAMI Gt).
To ensure accuracy in binning, gold standard cross-sample assemblies were used. The two real datasets include STEC, a collection of 53 fecal samples from the European Nucleotide Archive (ENA) project (https://www.ebi.ac.uk/ena/browser/view/PRJEB1775) (accessed on 10 April 2013) and Water Group1, a river water microbiome dataset with 8 samples, (retrieved from https://www.ebi.ac.uk/ena/browser/view/PRJNA542960) (accessed on 5 October 2019), mainly used to test performance under low-sample-count conditions.

2.5.2. Benchmark Datasets

To comprehensively evaluate the performance of MBGCCA in metagenomic binning tasks, we conducted experiments on a diverse set of datasets:
  • Simulated dataset: The simulated dataset Strong100 was generated using the Badread tool [52], following the methodology proposed by Quince et al. [53]. Specifically, sequencing reads were simulated from 100 bacterial strains representing 50 different species, with abundances assigned randomly to mimic natural community variability.
  • Real wastewater treatment plant (WWTP) datasets: Hjor, Viby, Damh, Mari, AalE, and Hade.
  • Real environmental dataset: Soil, representing complex microbial communities in natural soil environments.
To benchmark MBGCCA on long-read metagenomic binning, we generated synthetic long-read datasets using badread [52], following the simulation protocol adopted in LRBinner [54]. Specifically, reads were simulated from the 100 strains of Strong100 with randomly assigned abundances, and the assemblies were produced using metaFlye v2.9 [38] to ensure consistency and comparability with baseline methods.
Details of all test datasets are summarized in Table 1.
WWTP datasets were derived from Singleton et al. [55] (PRJNA629478). For each WWTP dataset, contig coverage was calculated from long-read assemblies, and three additional short-read samples from the same site (different time points) were used to compute contig abundance. Assemblies were generated using metaFlye, polished in three rounds with Racon v1.3.3 [56], and further corrected using short reads. Only one sample per WWTP dataset contains long-read data (used for assembly), while the other three are used for abundance estimation. The Soil dataset comes from Brunbjerg et al. [57], and was used alongside other datasets for model development and hyperparameter optimization—except for Damh and Hade, which were held out to evaluate potential overfitting.
Long-Read Simulation Protocol
For the simulation of long-read metagenomic data, we sampled strain-level abundances from a Dirichlet distribution, where p g denotes the proportion of reads assigned to strain g, and the total number of simulated bases was set to N = 7.5  Gbp. The synthetic sequencing reads were generated using badread [52] with the following configuration:
  • Mean read length: 10,000 bp.
  • Standard deviation: 7000 bp.
  • Read identity: mean = 98%; max = 99.9%; standard deviation = 5.
  • Error model: nanopore2020 (default).
To emulate real-world long-read metagenomic workflows, all reads were assembled using metaFlye v2.9 [38], thereby ensuring high simulation fidelity and comparability across methods.

3. Results

In this study, we compared the performance of MBGCCA with five mainstream binning tools on the same datasets, including GraphBin [58], MaxBin2 [19], SemiBin-ocean [59], SemiBin-train [59], VAMB [27], MetaBAT2 [21], and GraphMB [28], using default parameters for all methods. All methods took contig sequences and their abundance information as input.

3.1. Evaluation on Simulated Dataset Strong100

We evaluated MBGCCA and baseline methods on the simulated dataset Strong100 using the AMBER toolkit [49]. Evaluation metrics included average purity (AP, bp-weighted), average completeness (AC, bp-weighted), F1 score (harmonic mean of AP and AC), and the number of high-quality (HQ) bins (completeness > 90%, contamination < 5%).
As shown in Table 2, MBGCCA consistently outperforms all competing methods across key metrics. It achieves the highest AP of 0.986, outperforming VAMB (0.969) and GraphMB (0.967), indicating superior ability in reducing contamination and improving bin purity. For AC, MBGCCA attains 0.784, slightly surpassing SemiBin-ocean (0.783) and significantly exceeding GraphBin (0.613) and MetaBAT2 (0.592), demonstrating robust genome recovery. MBGCCA also achieves the highest F1 score (0.864), outperforming GraphMB (0.852) and VAMB (0.849), and recovers the most HQ bins (32), a 10.3% improvement over the next-best method (GraphMB, 29).
While MBGCCA adopts a self-supervised learning strategy without requiring labeled genomes, methods like SemiBin-train rely on pretraining with external annotations. Despite this, MBGCCA still surpasses SemiBin-train across all metrics (AP: 0.986 vs. 0.826; F1: 0.864 vs. 0.823; HQ bins: 32 vs. 20). This highlights MBGCCA’s strength in generalizability and scalability, particularly in environments where reference genomes are unavailable or incomplete. We have emphasized this contrast in the Section 4.
In summary, MBGCCA achieves top performance across all four core metrics—AP, AC, F1, and HQ bins—solidifying its effectiveness and reliability for microbial genome reconstruction. Its consistent accuracy makes it well-suited for metagenomic analysis in complex or low-abundance microbial environments.

3.2. Evaluation on Real-World Datasets

Table 3 presents the number of high-quality (HQ) bins produced by various binning tools across multiple real-world datasets, including GraphBin, MaxBin2, SemiBin-ocean, SemiBin-train, VAMB, MetaBAT2, GraphMB, and the proposed method, MBGCCA. The number of HQ bins serves as a key indicator for assessing the effectiveness of each binning method on different datasets.
As shown in Table 3, MBGCCA achieves the highest number of HQ bins across all datasets, outperforming all other tools. For example, in the Hjor, Viby, Damh, and Hade datasets, MBGCCA obtains 28, 32, 46, and 56 HQ bins respectively, exceeding all competitors.
In the Hade dataset, MBGCCA achieves 56 HQ bins, representing a 27.3% increase over MetaBAT2 (44 bins) and a 7.7% improvement over GraphMB (52 bins), highlighting its superior performance in HQ bin identification. In the AalE dataset, MBGCCA and MetaBAT2 both achieve 43 HQ bins, substantially outperforming VAMB and GraphMB, indicating MBGCCA’s robustness and consistency across datasets. Although MetaBAT2 is widely regarded as a strong binning tool, MBGCCA consistently matches or exceeds its performance in all scenarios. For instance, on the Viby dataset, MBGCCA yields 32 HQ bins compared to MetaBAT2’s 29, an improvement of 10.3%. On the Damh dataset, MBGCCA produces 46 HQ bins versus MetaBAT2’s 41, a 12.2% increase.
These results demonstrate that MBGCCA offers superior or at least equivalent performance compared to mainstream binning tools, with generally higher HQ bin counts, supporting its effectiveness, accuracy, and robustness in microbial genome reconstruction across diverse real-world environments.

3.3. Runtime and Memory Usage Analysis of MBGCCA

We evaluated the runtime and peak memory usage of three deep learning-based binning tools on the Soil dataset. All tools were executed on a workstation equipped with dual AMD EPYC 7H12 64-Core Processors and an NVIDIA RTX 4090 GPU. Each tool was run using 48 threads, and the time spent on aligning reads to contigs was excluded, as this step is required by all binning methods.
Each method was executed three times per dataset, and the average runtime and memory usage were recorded.
As illustrated in Table 4 and Table 5, MBGCCA exhibits moderate increases in runtime and memory usage compared to other deep learning-based binning methods. Specifically, MBGCCA’s runtime is slightly longer than VAMB but shorter than GraphMB on both Hade and Soil datasets. In terms of peak memory usage, MBGCCA consumes more memory than both VAMB and GraphMB, especially on the larger Soil dataset (10.5 GB).
Despite these resource requirements, MBGCCA achieves significantly higher binning performance (Table 1 and Table 2), particularly in terms of genome completeness, purity, and recovery of low-abundance genomes. These gains stem from the integration of multi-modal features, graph-based neighborhood aggregation, and contrastive learning strategies. Thus, the slightly increased computational cost is a worthwhile trade-off for its enhanced accuracy and robustness in complex metagenomic scenarios.

3.4. Ablation Study

To evaluate the contribution of different input features and model components in the MBGCCA framework, we conducted an ablation study with the following variants. GraphMB (TNF): This variant uses only tetranucleotide frequency (TNF) features as input, excluding abundance information, to assess the compositional contribution. GraphMB (Abundance): This model uses only abundance profiles without TNF features to evaluate the impact of sequencing-depth-based information. GraphMB (Step 1): This version retains the multimodal feature fusion module and uses the initial contig embeddings from stage one, but excludes neighborhood aggregation via the graph neural network. As such, it does not incorporate contig–contig topological relationships and serves to isolate the effectiveness of the graph-based learning in stage two. These variants allow us to dissect the individual impact of input features and structural information, and to better understand the importance of multimodal integration and graph representation learning in our model.
The results (Table 6) demonstrate that while both compositional (TNF) and abundance-based features contribute to performance, their combination significantly improves binning outcomes. Notably, GraphMB (Step 1) outperforms single-feature variants, confirming the utility of multimodal feature fusion. The full MBGCCA model achieves the highest number of HQ bins across all datasets, highlighting the critical role of graph-based neighborhood aggregation in enhancing genome recovery, especially in complex microbial environments such as soil.

4. Discussion

In this study, we proposed a metagenomic binning method based on assembly graphs, termed MBGCCA, and conducted comprehensive evaluations on both simulated and real-world environmental datasets. Compared with conventional binning approaches, MBGCCA integrates GNNs with contrastive learning and information-theoretic regularization, enabling a more principled incorporation of assembly graph topology and genomic features. Experimental results demonstrate that MBGCCA not only performs well on simulated assemblies, but also exhibits strong generalization and adaptability to complex environmental metagenomic data.
Current metagenomic binning methods primarily rely on the fusion of sequence composition features (such as TNF) and abundance coverage. However, these approaches face challenges in recovering low-abundance species, handling noise in complex environments, and adapting to long-read sequencing data. MBGCCA addresses these limitations by leveraging GNNs to model the assembly graph and applying contrastive learning with entropy-aware objectives to optimize the embedding space. This allows the model to better capture topological structure while maximizing the information content of learned representations. The assembly graph encodes connectivity between contigs, and MBGCCA utilizes these connections to ensure that contigs belonging to the same genome are accurately grouped.
Traditional methods based on TNF and abundance are often sensitive to environmental noise and produce redundant or low-informative embeddings. In contrast, MBGCCA improves robustness by introducing decorrelation loss, which reduces feature redundancy and encourages the learning of high-entropy, information-rich representations. This entropy-aware design enhances both discriminative power and generalization ability across heterogeneous samples. Additionally, MBGCCA incorporates multi-view data augmentation and mutual information maximization to better recover low-abundance microbes, leading to improved performance on real-world metagenomic datasets.
Although MBGCCA offers improved computational efficiency over GraphMB, it still incurs higher computational costs on ultra-large datasets such as Soil. Future work may explore lightweight GNN architectures to reduce resource consumption. Currently, MBGCCA primarily utilizes genomic topology, TNF, and abundance features. Future extensions may incorporate multi-omics data, such as metatranscriptomics and metabolomics, to further enhance the information-theoretic richness and biological interpretability of the learned embeddings. The current study focuses on assembled contig sequences. Further optimization could extend MBGCCA to support novel data modalities, such as single-cell sequencing and spatial transcriptomics, thereby broadening its utility and resolution in microbial genome analysis.

5. Conclusions

In this study, we presented MBGCCA, a novel metagenomic binning framework that combines graph neural networks, contrastive learning, and information-theoretic regularization to improve the accuracy and robustness of genome reconstruction from complex metagenomic data. By leveraging assembly graph topology and integrating compositional and abundance features, MBGCCA generates high-entropy, low-redundancy embeddings that more faithfully represent the underlying genomic structure. Our experiments on both synthetic and real-world environmental datasets demonstrate that MBGCCA consistently outperforms existing binning methods, particularly in recovering low-abundance genomes and handling noisy, heterogeneous samples. The incorporation of entropy-aware objectives, such as mutual information maximization and embedding decorrelation, further enhances the discriminative power and generalization of the model. Future extensions of MBGCCA may explore lightweight GNNs, integration with multi-omics data, and support for new data modalities such as single-cell and spatial metagenomics, thereby expanding its applicability to a broader range of microbiome research scenarios.

Author Contributions

Conceptualization, G.W. and Y.L.; methodology, G.W.; software, G.W.; validation, G.W. and Y.L.; formal analysis, G.W.; investigation, G.W.; resources, Y.L.; data curation, G.W.; writing—original draft preparation, G.W.; writing—review and editing, Y.L.; visualization, G.W.; supervision, Y.L.; project administration, Y.L.; funding acquisition, Y.L. All authors have read and agreed to the published version of the manuscript.

Funding

This research was funded by the National Natural Science Foundation of China (Grant No. 62306142), the Jiangsu Funding Program for Excellent Postdoctoral Talent (Grant No. 2023ZB224), the Natural Science Foundation of the Higher Education Institutions of Jiangsu Province of China (Grant No. 24KJB520041).

Conflicts of Interest

The authors declare no conflicts of interest.

References

  1. Blanco-Míguez, A.; Beghini, F.; Cumbo, F.; McIver, L.J.; Thompson, K.N.; Zolfo, M.; Manghi, P.; Dubois, L.; Huang, K.D.; Thomas, A.M.; et al. Extending and improving metagenomic taxonomic profiling with uncharacterized species using MetaPhlAn 4. Nat. Biotechnol. 2023, 41, 1633–1644. [Google Scholar] [CrossRef]
  2. Liu, X.; Liu, Y.; Liu, J.; Zhang, H.; Shan, C.; Guo, Y.; Gong, X.; Cui, M.; Li, X.; Tang, M. Correlation between the gut microbiome and neurodegenerative diseases: A review of metagenomics evidence. Neural Regen. Res. 2024, 19, 833–845. [Google Scholar] [CrossRef]
  3. Kim, N.; Ma, J.; Kim, W.; Kim, J.; Belenky, P.; Lee, I. Genome-resolved metagenomics: A game changer for microbiome medicine. Exp. Mol. Med. 2024, 56, 1501–1512. [Google Scholar] [CrossRef]
  4. Liu, S.; Rodriguez, J.S.; Munteanu, V.; Ronkowski, C.; Sharma, N.K.; Alser, M.; Andreace, F.; Blekhman, R.; Błaszczyk, D.; Chikhi, R.; et al. Analysis of metagenomic data. Nat. Rev. Methods Prim. 2025, 5, 5. [Google Scholar] [CrossRef]
  5. Lin, G.L.; Drysdale, S.B.; Snape, M.D.; O’connor, D.; Brown, A.; MacIntyre-Cockett, G.; Mellado-Gomez, E.; De Cesare, M.; Ansari, M.A.; Bonsall, D.; et al. Targeted metagenomics reveals association between severity and pathogen co-detection in infants with respiratory syncytial virus. Nat. Commun. 2024, 15, 2379. [Google Scholar] [CrossRef]
  6. Masuda, Y.; Mise, K.; Xu, Z.; Zhang, Z.; Shiratori, Y.; Senoo, K.; Itoh, H. Global soil metagenomics reveals distribution and predominance of Deltaproteobacteria in nitrogen-fixing microbiome. Microbiome 2024, 12, 95. [Google Scholar] [CrossRef]
  7. Wolińska, A. Metagenomic achievements in microbial diversity determination in croplands: A review. In Microbial Diversity in the Genomic Era; Academic Press: Cambridge, MA, USA, 2019; pp. 15–35. [Google Scholar]
  8. Zheng, M.; Wen, L.; He, C.; Chen, X.; Si, L.; Li, H.; Liang, Y.; Zheng, W.; Guo, F. Sequencing-guided re-estimation and promotion of cultivability for environmental bacteria. Nat. Commun. 2024, 15, 9051. [Google Scholar] [CrossRef]
  9. Delgado, L.F.; Andersson, A.F. Evaluating metagenomic assembly approaches for biome-specific gene catalogues. Microbiome 2022, 10, 72. [Google Scholar] [CrossRef] [PubMed]
  10. Mallawaarachchi, V.G.; Wickramarachchi, A.S.; Lin, Y. Improving metagenomic binning results with overlapped bins using assembly graphs. Algorithms Mol. Biol. 2021, 16, 3. [Google Scholar] [CrossRef] [PubMed]
  11. Feng, X.; Li, H. Evaluating and improving the representation of bacterial contents in long-read metagenome assemblies. Genome Biol. 2024, 25, 92. [Google Scholar] [CrossRef] [PubMed]
  12. Zhang, Y.; Thompson, K.N.; Branck, T.; Yan, Y.; Nguyen, L.H.; Franzosa, E.A.; Huttenhower, C. Metatranscriptomics for the human microbiome and microbial community functional profiling. Annu. Rev. Biomed. Data Sci. 2021, 4, 279–311. [Google Scholar] [CrossRef]
  13. Lui, L.M.; Nielsen, T.N.; Arkin, A.P. A method for achieving complete microbial genomes and improving bins from metagenomics data. PLoS Comput. Biol. 2021, 17, e1008972. [Google Scholar] [CrossRef]
  14. Lema, N.K.; Gemeda, M.T.; Woldesemayat, A.A. Recent advances in metagenomic approaches, applications, and challenges. Curr. Microbiol. 2023, 80, 347. [Google Scholar] [CrossRef]
  15. Pérez-Cobas, A.E.; Gomez-Valero, L.; Buchrieser, C. Metagenomic approaches in microbial ecology: An update on whole-genome and marker gene sequencing analyses. Microb. Genom. 2020, 6, e000409. [Google Scholar] [CrossRef]
  16. Alneberg, J.; Bjarnason, B.S.; de Bruijn, I.; Schirmer, M.; Quick, J.; Ijaz, U.Z.; Loman, N.J.; Andersson, A.F.; Quince, C. CONCOCT: Clustering contigs on coverage and composition. arXiv 2013, arXiv:1312.4038. [Google Scholar] [CrossRef]
  17. Abdi, H.; Williams, L.J. Principal component analysis. Wiley Interdiscip. Rev. Comput. Stat. 2010, 2, 433–459. [Google Scholar] [CrossRef]
  18. Reynolds, D. Gaussian mixture models. In Encyclopedia of Biometrics; Springer: Berlin/Heidelberg, Germany, 2015; pp. 827–832. [Google Scholar]
  19. Wu, Y.W.; Simmons, B.A.; Singer, S.W. MaxBin 2.0: An automated binning algorithm to recover genomes from multiple metagenomic datasets. Bioinformatics 2016, 32, 605–607. [Google Scholar] [CrossRef]
  20. Moon, T.K. The expectation-maximization algorithm. IEEE Signal Process. Mag. 1996, 13, 47–60. [Google Scholar] [CrossRef]
  21. Kang, D.D.; Li, F.; Kirton, E.; Thomas, A.; Egan, R.; An, H.; Wang, Z. MetaBAT 2: An adaptive binning algorithm for robust and efficient genome reconstruction from metagenome assemblies. PeerJ 2019, 7, e7359. [Google Scholar] [CrossRef] [PubMed]
  22. Hickl, O.; Queirós, P.; Wilmes, P.; May, P.; Heintz-Buschart, A. Binny: An automated binning algorithm to recover high-quality genomes from complex metagenomic datasets. Briefings Bioinform. 2022, 23, bbac431. [Google Scholar] [CrossRef] [PubMed]
  23. McInnes, L.; Healy, J.; Astels, S. hdbscan: Hierarchical density based clustering. J. Open Source Softw. 2017, 2, 205. [Google Scholar] [CrossRef]
  24. Liu, C.C.; Dong, S.S.; Chen, J.B.; Wang, C.; Ning, P.; Guo, Y.; Yang, T.L. MetaDecoder: A novel method for clustering metagenomic contigs. Microbiome 2022, 10, 46. [Google Scholar] [CrossRef] [PubMed]
  25. Görür, D.; Edward Rasmussen, C. Dirichlet process gaussian mixture models: Choice of the base distribution. J. Comput. Sci. Technol. 2010, 25, 653–664. [Google Scholar] [CrossRef]
  26. Wang, Z.; Huang, P.; You, R.; Sun, F.; Zhu, S. MetaBinner: A high-performance and stand-alone ensemble binning method to recover individual genomes from complex microbial communities. Genome Biol. 2023, 24, 1. [Google Scholar] [CrossRef]
  27. Nissen, J.N.; Johansen, J.; Allesøe, R.L.; Sønderby, C.K.; Armenteros, J.J.A.; Grønbech, C.H.; Jensen, L.J.; Nielsen, H.B.; Petersen, T.N.; Winther, O.; et al. Improved metagenome binning and assembly using deep variational autoencoders. Nat. Biotechnol. 2021, 39, 555–560. [Google Scholar] [CrossRef] [PubMed]
  28. Lamurias, A.; Sereika, M.; Albertsen, M.; Hose, K.; Nielsen, T.D. Metagenomic binning with assembly graph embeddings. Bioinformatics 2022, 38, 4481–4487. [Google Scholar] [CrossRef]
  29. Wang, Z.; You, R.; Han, H.; Liu, W.; Sun, F.; Zhu, S. Effective binning of metagenomic contigs using contrastive multi-view representation learning. Nat. Commun. 2024, 15, 585. [Google Scholar] [CrossRef] [PubMed]
  30. Zhang, C.; Li, W.; Deng, M.; Jiang, Y.; Cui, X.; Chen, P. SIG: Graph-Based Cancer Subtype Stratification With Gene Mutation Structural Information. IEEE/ACM Trans. Comput. Biol. Bioinform. 2024, 21, 1752–1764. [Google Scholar] [CrossRef]
  31. Chen, L.; Li, W.; Cui, X.; Wang, Z.; Berretti, S.; Wan, S. MS-GDA: Improving heterogeneous recipe representation via multinomial sampling graph data augmentation. ACM Trans. Multimed. Comput. Commun. Appl. 2024, 20, 1–23. [Google Scholar] [CrossRef]
  32. Hardoon, D.R.; Szedmak, S.; Shawe-Taylor, J. Canonical correlation analysis: An overview with application to learning methods. Neural Comput. 2004, 16, 2639–2664. [Google Scholar] [CrossRef] [PubMed]
  33. Corso, G.; Stark, H.; Jegelka, S.; Jaakkola, T.; Barzilay, R. Graph neural networks. Nat. Rev. Methods Prim. 2024, 4, 17. [Google Scholar] [CrossRef]
  34. Teeling, H.; Meyerdierks, A.; Bauer, M.; Amann, R.; Glöckner, F.O. Application of tetranucleotide frequencies for the assignment of genomic fragments. Environ. Microbiol. 2004, 6, 938–947. [Google Scholar] [CrossRef]
  35. Wang, Z.; Wang, Z.; Lu, Y.Y.; Sun, F.; Zhu, S. SolidBin: Improving metagenome binning with semi-supervised normalized cut. Bioinformatics 2019, 35, 4229–4238. [Google Scholar] [CrossRef] [PubMed]
  36. Field, D.A. Laplacian smoothing and Delaunay triangulations. Commun. Appl. Numer. Methods 1988, 4, 709–712. [Google Scholar] [CrossRef]
  37. Khaertdinov, B.; Asteriadis, S.; Ghaleb, E. Dynamic temperature scaling in contrastive self-supervised learning for sensor-based human activity recognition. IEEE Trans. Biom. Behav. Identity Sci. 2022, 4, 498–507. [Google Scholar] [CrossRef]
  38. Kolmogorov, M.; Bickhart, D.M.; Behsaz, B.; Gurevich, A.; Rayko, M.; Shin, S.B.; Kuhn, K.; Yuan, J.; Polevikov, E.; Smith, T.P.; et al. metaFlye: Scalable long-read metagenome assembly using repeat graphs. Nat. Methods 2020, 17, 1103–1110. [Google Scholar] [CrossRef]
  39. Bankevich, A.; Nurk, S.; Antipov, D.; Gurevich, A.A.; Dvorkin, M.; Kulikov, A.S.; Lesin, V.M.; Nikolenko, S.I.; Pham, S.; Prjibelski, A.D.; et al. SPAdes: A new genome assembly algorithm and its applications to single-cell sequencing. J. Comput. Biol. 2012, 19, 455–477. [Google Scholar] [CrossRef]
  40. Hamilton, W.; Ying, Z.; Leskovec, J. Inductive representation learning on large graphs. Adv. Neural Inf. Process. Syst. 2017, 30. [Google Scholar]
  41. Xie, Y.; Xu, Z.; Zhang, J.; Wang, Z.; Ji, S. Self-supervised learning of graph neural networks: A unified review. IEEE Trans. Pattern Anal. Mach. Intell. 2022, 45, 2412–2429. [Google Scholar] [CrossRef] [PubMed]
  42. Xu, Y.; Huang, S.; Zhang, H.; Li, X. Why does dropping edges usually outperform adding edges in graph contrastive learning? In Proceedings of the AAAI Conference on Artificial Intelligence, Philadelphia, PA, USA, 25 February–4 March 2025; Volume 39, pp. 21824–21832. [Google Scholar] [CrossRef]
  43. Mishra, P.; Piktus, A.; Goossen, G.; Silvestri, F. Node masking: Making graph neural networks generalize and scale better. arXiv 2020, arXiv:2001.07524. [Google Scholar]
  44. Zhang, H.; Wu, Q.; Yan, J.; Wipf, D.; Yu, P.S. From canonical correlation analysis to self-supervised graph neural networks. Adv. Neural Inf. Process. Syst. 2021, 34, 76–89. [Google Scholar]
  45. Zhang, S.; Tong, H.; Xu, J.; Maciejewski, R. Graph convolutional networks: A comprehensive review. Comput. Soc. Netw. 2019, 6, 1–23. [Google Scholar] [CrossRef] [PubMed]
  46. Ming, Q.; Miao, L.; Zhou, Z.; Yang, X.; Dong, Y. Optimization for arbitrary-oriented object detection via representation invariance loss. IEEE Geosci. Remote Sens. Lett. 2021, 19, 1–5. [Google Scholar] [CrossRef]
  47. Parks, D.H.; Imelfort, M.; Skennerton, C.T.; Hugenholtz, P.; Tyson, G.W. CheckM: Assessing the quality of microbial genomes recovered from isolates, single cells, and metagenomes. Genome Res. 2015, 25, 1043–1055. [Google Scholar] [CrossRef]
  48. Olm, M.R.; Brown, C.T.; Brooks, B.; Banfield, J.F. dRep: A tool for fast and accurate genomic comparisons that enables improved genome recovery from metagenomes through de-replication. ISME J. 2017, 11, 2864–2868. [Google Scholar] [CrossRef]
  49. Meyer, F.; Hofmann, P.; Belmann, P.; Garrido-Oter, R.; Fritz, A.; Sczyrba, A.; McHardy, A.C. AMBER: Assessment of metagenome BinnERs. Gigascience 2018, 7, giy069. [Google Scholar] [CrossRef] [PubMed]
  50. Sieber, C.M.; Probst, A.J.; Sharrar, A.; Thomas, B.C.; Hess, M.; Tringe, S.G.; Banfield, J.F. Recovery of genomes from metagenomes via a dereplication, aggregation and scoring strategy. Nat. Microbiol. 2018, 3, 836–843. [Google Scholar] [CrossRef]
  51. Sczyrba, A.; Hofmann, P.; Belmann, P.; Koslicki, D.; Janssen, S.; Dröge, J.; Gregor, I.; Majda, S.; Fiedler, J.; Dahms, E.; et al. Critical assessment of metagenome interpretation—A benchmark of metagenomics software. Nat. Methods 2017, 14, 1063–1071. [Google Scholar] [CrossRef]
  52. Wick, R.R. Badread: Simulation of error-prone long reads. J. Open Source Softw. 2019, 4, 1316. [Google Scholar] [CrossRef]
  53. Quince, C.; Nurk, S.; Raguideau, S.; James, R.; Soyer, O.S.; Summers, J.K.; Limasset, A.; Eren, A.M.; Chikhi, R.; Darling, A.E. STRONG: Metagenomics strain resolution on assembly graphs. Genome Biol. 2021, 22, 214. [Google Scholar] [CrossRef] [PubMed]
  54. Wickramarachchi, A.; Lin, Y. LRBinner: Binning long reads in metagenomics datasets. In Proceedings of the 21st International Workshop on Algorithms in Bioinformatics (WABI 2021), Virtual, 2–4 August 2021; Schloss Dagstuhl–Leibniz-Zentrum für Informatik: Saarbrücken, Germany; Wadern, Germany, 2021; pp. 11:1–11:18. [Google Scholar]
  55. Singleton, C.M.; Petriglieri, F.; Kristensen, J.M.; Kirkegaard, R.H.; Michaelsen, T.Y.; Andersen, M.H.; Kondrotaite, Z.; Karst, S.M.; Dueholm, M.S.; Nielsen, P.H.; et al. Connecting structure to function with the recovery of over 1000 high-quality metagenome-assembled genomes from activated sludge using long-read sequencing. Nat. Commun. 2021, 12, 2009. [Google Scholar] [CrossRef] [PubMed]
  56. Vaser, R.; Sović, I.; Nagarajan, N.; Šikić, M. Fast and accurate de novo genome assembly from long uncorrected reads. Genome Res. 2017, 27, 737–746. [Google Scholar] [CrossRef]
  57. Brunbjerg, A.K.; Bruun, H.H.; Brøndum, L.; Classen, A.T.; Dalby, L.; Fog, K.; Frøslev, T.G.; Goldberg, I.; Hansen, A.J.; Hansen, M.D.; et al. A systematic survey of regional multi-taxon biodiversity: Evaluating strategies and coverage. BMC Ecol. 2019, 19, 43. [Google Scholar] [CrossRef]
  58. Mallawaarachchi, V.; Wickramarachchi, A.; Lin, Y. GraphBin: Refined binning of metagenomic contigs using assembly graphs. Bioinformatics 2020, 36, 3307–3313. [Google Scholar] [CrossRef]
  59. Pan, S.; Zhu, C.; Zhao, X.M.; Coelho, L.P. SemiBin: Incorporating information from reference genomes with semi-supervised deep learning leads to better metagenomic assembled genomes (MAGs). bioRxiv 2021. [Google Scholar] [CrossRef]
Figure 1. Overview of the MBGCCA framework. (A) Step 1: Fusion of TNF and abundance features. (B) Step 2: Aggregation of contig neighborhood information.
Figure 1. Overview of the MBGCCA framework. (A) Step 1: Fusion of TNF and abundance features. (B) Step 2: Aggregation of contig neighborhood information.
Entropy 27 00921 g001
Table 1. Test datasets overview.
Table 1. Test datasets overview.
DatasetTotalReads N50AssemblyContigs N50MeanContigsEdgesSamples
(Gbp) (kbp) (Gbp) (kbp) Cov.
Strong1007.513.30.17175.0428526701
Hjor16.08.70.8680.41319,49659374
Viby17.214.01.32101.0723,38978004
Damh26.714.31.93119.0832,77114,0664
Mari23.310.11.6983.1836,61112,6514
AalE27.710.21.9283.4840,82712,4254
Hade45.29.83.0173.9970,40227,9524
Soil115.07.71.9893.31951,13569,5221
Table 2. AMBER assessment metrics for binning tools on the simulated Strong100 dataset.
Table 2. AMBER assessment metrics for binning tools on the simulated Strong100 dataset.
MethodAP (bp)AC (bp)F1HQ Bins
GraphBin0.8480.6130.71223
MaxBin20.8180.7650.79114
SemiBin-ocean0.8580.7830.81926
SemiBin-train0.8260.8200.82320
VAMB0.9690.7550.84926
MetaBAT20.9050.5920.71626
GraphMB0.9670.7620.85229
MBGCCA0.9860.7840.86432
Note: Higher values of AP (average purity), AC (average completeness), F1 score, and HQ bins (high-quality bins) indicate better performance.
Table 3. Comparison of HQ bins obtained using MBGCCA and state-of-the-art binning tools.
Table 3. Comparison of HQ bins obtained using MBGCCA and state-of-the-art binning tools.
MethodHjorVibyDamhMariAalEHadeSoil
GraphBin111514161260
MaxBin21219161412190
SemiBin-ocean11122182170
SemiBin-train74232232250
VAMB221222303780
MetaBAT22329413943442
GraphMB2523434846523
MBGCCA2832464943563
Table 4. Runtime comparison (in seconds) of three deep learning-based methods.
Table 4. Runtime comparison (in seconds) of three deep learning-based methods.
MethodHade (s)Soil (s)
VAMB13231241
GraphMB20453768
MBGCCA18323387
Table 5. Peak memory usage (in MB) of three deep learning-based methods.
Table 5. Peak memory usage (in MB) of three deep learning-based methods.
MethodHade (MB)Soil (MB)
VAMB15641322
GraphMB19029343
MBGCCA206710,532
Table 6. Comparison of HQ bins obtained using different variants of MBGCCA.
Table 6. Comparison of HQ bins obtained using different variants of MBGCCA.
MethodHjorVibyDamhMariAalEHadeSoil
GraphMB (TNF)2023181622281
GraphMB (Abundance)1418191719160
GraphMB (Step 1)2426404240482
MBGCCA2832464943563
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Wei, G.; Liu, Y. A Graph Contrastive Learning Method for Enhancing Genome Recovery in Complex Microbial Communities. Entropy 2025, 27, 921. https://doi.org/10.3390/e27090921

AMA Style

Wei G, Liu Y. A Graph Contrastive Learning Method for Enhancing Genome Recovery in Complex Microbial Communities. Entropy. 2025; 27(9):921. https://doi.org/10.3390/e27090921

Chicago/Turabian Style

Wei, Guo, and Yan Liu. 2025. "A Graph Contrastive Learning Method for Enhancing Genome Recovery in Complex Microbial Communities" Entropy 27, no. 9: 921. https://doi.org/10.3390/e27090921

APA Style

Wei, G., & Liu, Y. (2025). A Graph Contrastive Learning Method for Enhancing Genome Recovery in Complex Microbial Communities. Entropy, 27(9), 921. https://doi.org/10.3390/e27090921

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop