[2505.18150] Generative Distribution Embeddings: Lifting autoencoders to the space of distributions for multiscale representation learning

[2505.18150] Generative Distribution Embeddings: Lifting autoencoders to the space of distributions for multiscale representation learning

arXiv - Machine Learning 4 min read Article

Summary

The paper introduces Generative Distribution Embeddings (GDE), a novel framework that enhances autoencoders for multiscale representation learning by operating on entire distributions rather than single data points.

Why It Matters

GDEs address the need for models that can reason across multiple scales in real-world problems, particularly in computational biology. By improving representation learning, GDEs can lead to better predictive models and insights in complex biological datasets, making them highly relevant for researchers in machine learning and bioinformatics.

Key Takeaways

  • Generative Distribution Embeddings (GDE) enhance traditional autoencoders by focusing on distributions.
  • GDEs demonstrate superior performance on synthetic datasets compared to existing methods.
  • The framework is applicable to various computational biology challenges, including RNA sequencing and synthetic promoter design.

Computer Science > Machine Learning arXiv:2505.18150 (cs) [Submitted on 23 May 2025 (v1), last revised 20 Feb 2026 (this version, v2)] Title:Generative Distribution Embeddings: Lifting autoencoders to the space of distributions for multiscale representation learning Authors:Nic Fishman, Gokul Gowri, Peng Yin, Jonathan Gootenberg, Omar Abudayyeh View a PDF of the paper titled Generative Distribution Embeddings: Lifting autoencoders to the space of distributions for multiscale representation learning, by Nic Fishman and 4 other authors View PDF HTML (experimental) Abstract:Many real-world problems require reasoning across multiple scales, demanding models which operate not on single data points, but on entire distributions. We introduce generative distribution embeddings (GDE), a framework that lifts autoencoders to the space of distributions. In GDEs, an encoder acts on sets of samples, and the decoder is replaced by a generator which aims to match the input distribution. This framework enables learning representations of distributions by coupling conditional generative models with encoder networks which satisfy a criterion we call distributional invariance. We show that GDEs learn predictive sufficient statistics embedded in the Wasserstein space, such that latent GDE distances approximately recover the $W_2$ distance, and latent interpolation approximately recovers optimal transport trajectories for Gaussian and Gaussian mixture distributions. We systematically benchmark ...

Related Articles

UMKC Announces New Master of Science in Artificial Intelligence
Ai Infrastructure

UMKC Announces New Master of Science in Artificial Intelligence

UMKC announces a new Master of Science in Artificial Intelligence program aimed at addressing workforce demand for AI expertise, set to l...

AI News - General · 4 min ·
Machine Learning

AI assistants are optimized to seem helpful. That is not the same thing as being helpful.

RLHF trains models on human feedback. Humans rate responses they like. And it turns out humans consistently rate confident, fluent, agree...

Reddit - Artificial Intelligence · 1 min ·
Llms

wtf bro did what? arc 3 2026

The Physarum Explorer is a high-speed, bio-inspired neural model designed specifically for ARC geometry. Here is the snapshot of its curr...

Reddit - Artificial Intelligence · 1 min ·
Meta Pauses Work With Mercor After Data Breach Puts AI Industry Secrets at Risk | WIRED
Machine Learning

Meta Pauses Work With Mercor After Data Breach Puts AI Industry Secrets at Risk | WIRED

Major AI labs are investigating a security incident that impacted Mercor, a leading data vendor. The incident could have exposed key data...

Wired - AI · 6 min ·
More in Machine Learning: This Week Guide Trending

No comments

No comments yet. Be the first to comment!

Stay updated with AI News

Get the latest news, tools, and insights delivered to your inbox.

Daily or weekly digest • Unsubscribe anytime