Why isn’t LLM reasoning done in vector space instead of natural language?[D]
Why don’t LLMs use explicit vector-based reasoning instead of language-based chain-of-thought? What would happen if they did? Most LLM re...
GPT, Claude, Gemini, and other LLMs
Why don’t LLMs use explicit vector-based reasoning instead of language-based chain-of-thought? What would happen if they did? Most LLM re...
Abstract page for arXiv paper 2512.12072: VOYAGER: A Training Free Approach for Generating Diverse Datasets using LLMs
Abstract page for arXiv paper 2601.12248: AQUA-Bench: Beyond Finding Answers to Knowing When There Are None in Audio Question Answering
Abstract page for arXiv paper 2603.04045: Inference-Time Toxicity Mitigation in Protein Language Models
Abstract page for arXiv paper 2603.03590: Social Norm Reasoning in Multimodal Language Models: An Evaluation
Abstract page for arXiv paper 2603.03585: Belief-Sim: Towards Belief-Driven Simulation of Demographic Misinformation Susceptibility
Abstract page for arXiv paper 2603.04028: A Multi-Dimensional Quality Scoring Framework for Decentralized LLM Inference with Proof of Qua...
Abstract page for arXiv paper 2603.03555: Molt Dynamics: Emergent Social Phenomena in Autonomous AI Agent Populations
Abstract page for arXiv paper 2603.03543: Tucano 2 Cool: Better Open Source LLMs for Portuguese
Abstract page for arXiv paper 2603.03541: RAG-X: Systematic Diagnosis of Retrieval-Augmented Generation for Medical Question Answering
Abstract page for arXiv paper 2603.03536: SafeCRS: Personalized Safety Alignment for LLM-Based Conversational Recommender Systems
Abstract page for arXiv paper 2603.03946: Lang2Str: Two-Stage Crystal Structure Generation with LLMs and Continuous Flow Models
Abstract page for arXiv paper 2603.03512: Baseline Performance of AI Tools in Classifying Cognitive Demand of Mathematical Tasks
Abstract page for arXiv paper 2603.03508: Raising Bars, Not Parameters: LilMoo Compact Language Model for Hindi
Abstract page for arXiv paper 2603.03805: Relational In-Context Learning via Synthetic Pre-training with Structural Prior
Abstract page for arXiv paper 2603.03417: Parallel Test-Time Scaling with Multi-Sequence Verifiers
Abstract page for arXiv paper 2603.03415: Farther the Shift, Sparser the Representation: Analyzing OOD Mechanisms in LLMs
Abstract page for arXiv paper 2603.03756: MOOSE-Star: Unlocking Tractable Training for Scientific Discovery by Breaking the Complexity Ba...
Abstract page for arXiv paper 2603.03410: On Google's SynthID-Text LLM Watermarking System: Theoretical Analysis and Empirical Validation
Abstract page for arXiv paper 2603.03379: MemSifter: Offloading LLM Memory Retrieval via Outcome-Driven Proxy Reasoning
Abstract page for arXiv paper 2603.03612: Why Are Linear RNNs More Parallelizable?
Abstract page for arXiv paper 2603.03371: Sleeper Cell: Injecting Latent Malice Temporal Backdoors into Tool-Using LLMs
Abstract page for arXiv paper 2603.03597: NuMuon: Nuclear-Norm-Constrained Muon for Compressible LLM Training
Get the latest news, tools, and insights delivered to your inbox.
Daily or weekly digest • Unsubscribe anytime