Cosmos Week
MetaGEM: Bottom-Up Reconstruction of Genome-Scale Metabolic Networks via Deep Enzyme-Metabolite Anchoring
CosmologyEnglish editionPreprintPreliminary result

MetaGEM: Bottom-Up Reconstruction of Genome-Scale Metabolic Networks via Deep Enzyme-Metabolite Anchoring

Genome-scale metabolic models are essential tools for systems biology and rational chassis design, but conventional top-down reconstruction depends heavily on sequence homology.

Original source cited and editorially framed by Cosmos Week. arXiv Quantitative Biology
Editorial signatureCosmos Week Editorial Desk
Published14 May 2026 13: 25 UTC
Updated2026-05-14
Coverage typePreprint
Evidence levelPreliminary result
Read time4 min read

Key points

  • Focus: Genome-scale metabolic models are essential tools for systems biology and rational chassis design, but conventional top-down reconstruction depends
  • Editorial reading: provisional result, not yet formally peer reviewed.
Full story

Genome-scale metabolic models are essential tools for systems biology and rational chassis design, but conventional top-down reconstruction depends heavily on sequence homology and often leaves unknown enzymes and metabolic dark matter. The new analysis still awaits peer review, but it already lays out the central claim clearly.

It is relevant because cosmology operates at the edge of what current instruments can measure, where systematic errors and model assumptions are never trivial. Small discrepancies between independent measurements have historically pointed toward missing physics rather than simple calibration errors, and the ongoing tension in the Hubble constant is a live example of how a persistent disagreement between methods can reshape the theoretical landscape. Each new dataset that approaches this territory with independent systematics adds real information to a problem that has resisted easy resolution for more than a decade. Genome-scale metabolic models (GEMs) are essential tools for systems biology and rational chassis design, but conventional top-down reconstruction depends heavily on sequence. Direct reconstruction from metabolomics is also difficult because mapping observed metabolites to reactions is an ill-posed inverse problem with combinatorial ambiguity and.

Here we present MetaGEM, a bottom-up framework that uses enzymes as physical anchors to convert system-level network inference into enzyme-metabolite interaction prediction. MetaGEM uses a multimodal dual-tower architecture that combines protein evolutionary semantics from a protein language model with three-dimensional metabolite representations.

It further introduces contrastive learning with hard negative mining to separate structurally similar metabolites and reduce false positive interactions. In downstream reconstruction, MetaGEM generates functional genome-scale metabolic models for Escherichia coli, Bacillus subtilis, and Pseudomonas aeruginosa.

The reconstructed models improve network connectivity, capture promiscuous enzymes, and show strong agreement with experimental phenotype microarray and gene essentiality data. These results indicate that MetaGEM provides a practical route from metabolomic evidence to computable metabolic networks and offers a foundation for automated AI-driven virtual.

The relevance goes beyond one dataset because even small shifts in measured parameters can matter when the field is testing the limits of the standard cosmological model. The Lambda-CDM framework describes the observable universe with remarkable economy, but its success rests on two components, dark matter and dark energy, whose physical nature remains entirely unknown. Any credible measurement that tightens or loosens the constraints on those components moves the entire theoretical enterprise forward, regardless of whether the immediate result looks dramatic on its own terms.

Both individuals and organizations that work with arXivLabs have embraced and accepted our values of openness, community, excellence, and user data privacy. ArXiv is committed to these values and only works with partners that adhere to them.

Because this is still a preprint, the result should be read with genuine interest and proportionate caution. Peer review is not a guarantee of correctness, but it is a process that forces authors to respond to technical criticism from specialists who have no stake in a particular outcome. Preprints that survive that process, often with substantive revisions, emerge with a stronger evidential base than the version that first appeared. Until that stage is complete, the responsible reading keeps uncertainty explicitly visible rather than treating the claims as established findings.

The next step is to see whether the effect survives when independent surveys, different calibration strategies and tighter control of systematic uncertainties enter the picture. Programmes such as Euclid, DESI and the Rubin Observatory will deliver datasets over the next several years that cover the same parameter space with largely independent methods. If the current signal persists through those tests, its theoretical implications will become impossible to set aside. Until peer review and independent follow-up address those open questions, skepticism is not a failure of appreciation for the work; it is part of how science decides what to keep.

Source