[2602.16086] LGQ: Learning Discretization Geometry for Scalable and Stable Image Tokenization

[2602.16086] LGQ: Learning Discretization Geometry for Scalable and Stable Image Tokenization

arXiv - Machine Learning 4 min read Article

Summary

The paper presents LGQ, a novel image tokenizer that learns discretization geometry to enhance scalability and stability in visual generation, outperforming existing methods in efficiency and representation.

Why It Matters

As image generation technologies advance, efficient tokenization becomes crucial for maintaining quality while reducing computational demands. LGQ addresses the limitations of current quantizers, making it significant for researchers and practitioners in computer vision and machine learning.

Key Takeaways

  • LGQ introduces a learnable discretization geometry for image tokenization.
  • It replaces hard nearest-neighbor lookups with soft assignments for better training.
  • The method achieves improved performance metrics with fewer active codes.
  • LGQ balances code utilization without rigid geometrical constraints.
  • The approach is validated on ImageNet, demonstrating significant enhancements over existing tokenizers.

Computer Science > Computer Vision and Pattern Recognition arXiv:2602.16086 (cs) [Submitted on 17 Feb 2026] Title:LGQ: Learning Discretization Geometry for Scalable and Stable Image Tokenization Authors:Idil Bilge Altun, Mert Onur Cakiroglu, Elham Buxton, Mehmet Dalkilic, Hasan Kurban View a PDF of the paper titled LGQ: Learning Discretization Geometry for Scalable and Stable Image Tokenization, by Idil Bilge Altun and 4 other authors View PDF HTML (experimental) Abstract:Discrete image tokenization is a key bottleneck for scalable visual generation: a tokenizer must remain compact for efficient latent-space priors while preserving semantic structure and using discrete capacity effectively. Existing quantizers face a trade-off: vector-quantized tokenizers learn flexible geometries but often suffer from biased straight-through optimization, codebook under-utilization, and representation collapse at large vocabularies. Structured scalar or implicit tokenizers ensure stable, near-complete utilization by design, yet rely on fixed discretization geometries that may allocate capacity inefficiently under heterogeneous latent statistics. We introduce Learnable Geometric Quantization (LGQ), a discrete image tokenizer that learns discretization geometry end-to-end. LGQ replaces hard nearest-neighbor lookup with temperature-controlled soft assignments, enabling fully differentiable training while recovering hard assignments at inference. The assignments correspond to posterior respon...

Related Articles

Nlp

What does your AI bot buddy really think of you?

Try out this prompt and let us know if you find the response to be unsettling. (Hint: you should) Prompt: You have been maintaining an in...

Reddit - Artificial Intelligence · 1 min ·
Nlp

Persistent memory MCP server for AI agents (MCP + REST)

Pluribus is a memory service for agents (MCP + HTTP, Postgres-backed) that stores structured memory: constraints, decisions, patterns, an...

Reddit - Artificial Intelligence · 1 min ·
Machine Learning

[P] Unix philosophy for ML pipelines: modular, swappable stages with typed contracts

We built an open-source prototype that applies Unix philosophy to retrieval pipelines. Each stage (PII redaction, chunking, dedup, embedd...

Reddit - Machine Learning · 1 min ·
Nlp

[P] Using YouTube as a data source (lessons from building a coffee domain dataset)

I started working on a small coffee coaching app recently - something that could answer questions around brew methods, grind size, extrac...

Reddit - Machine Learning · 1 min ·
More in Nlp: This Week Guide Trending

No comments

No comments yet. Be the first to comment!

Stay updated with AI News

Get the latest news, tools, and insights delivered to your inbox.

Daily or weekly digest • Unsubscribe anytime