Large Language Models

GPT, Claude, Gemini, and other LLMs

Top This Week

Llms

The Day AI Stopped Being a Tab You Switch To — Claude Is Now Inside Your Software

submitted by /u/monotvtv [link] [comments]

Reddit - Artificial Intelligence · 1 min ·
Llms

How are LLMs 'corrected' when users identify them spreading misinformation or saying something harmful?

I watched Last Week Tonight's piece on AI chatbots today, and it got me thinking about that old screenshot of a Google search in which Ge...

Reddit - Artificial Intelligence · 1 min ·
Llms

What is the scientific value of administering the standard Rorschach test to LLMs when the training data is almost certainly contaminated? (R) + [D]

A recent paper published in JMIR Mental Health (Csigó & Cserey, 2026) caught my attention. The researchers administered the 10 standa...

Reddit - Machine Learning · 1 min ·

All Content

[2603.03805] Relational In-Context Learning via Synthetic Pre-training with Structural Prior
Llms

[2603.03805] Relational In-Context Learning via Synthetic Pre-training with Structural Prior

Abstract page for arXiv paper 2603.03805: Relational In-Context Learning via Synthetic Pre-training with Structural Prior

arXiv - AI · 3 min ·
[2603.03417] Parallel Test-Time Scaling with Multi-Sequence Verifiers
Llms

[2603.03417] Parallel Test-Time Scaling with Multi-Sequence Verifiers

Abstract page for arXiv paper 2603.03417: Parallel Test-Time Scaling with Multi-Sequence Verifiers

arXiv - AI · 4 min ·
[2603.03415] Farther the Shift, Sparser the Representation: Analyzing OOD Mechanisms in LLMs
Llms

[2603.03415] Farther the Shift, Sparser the Representation: Analyzing OOD Mechanisms in LLMs

Abstract page for arXiv paper 2603.03415: Farther the Shift, Sparser the Representation: Analyzing OOD Mechanisms in LLMs

arXiv - AI · 4 min ·
[2603.03756] MOOSE-Star: Unlocking Tractable Training for Scientific Discovery by Breaking the Complexity Barrier
Llms

[2603.03756] MOOSE-Star: Unlocking Tractable Training for Scientific Discovery by Breaking the Complexity Barrier

Abstract page for arXiv paper 2603.03756: MOOSE-Star: Unlocking Tractable Training for Scientific Discovery by Breaking the Complexity Ba...

arXiv - Machine Learning · 3 min ·
[2603.03410] On Google's SynthID-Text LLM Watermarking System: Theoretical Analysis and Empirical Validation
Llms

[2603.03410] On Google's SynthID-Text LLM Watermarking System: Theoretical Analysis and Empirical Validation

Abstract page for arXiv paper 2603.03410: On Google's SynthID-Text LLM Watermarking System: Theoretical Analysis and Empirical Validation

arXiv - AI · 4 min ·
[2603.03379] MemSifter: Offloading LLM Memory Retrieval via Outcome-Driven Proxy Reasoning
Llms

[2603.03379] MemSifter: Offloading LLM Memory Retrieval via Outcome-Driven Proxy Reasoning

Abstract page for arXiv paper 2603.03379: MemSifter: Offloading LLM Memory Retrieval via Outcome-Driven Proxy Reasoning

arXiv - AI · 4 min ·
[2603.03612] Why Are Linear RNNs More Parallelizable?
Llms

[2603.03612] Why Are Linear RNNs More Parallelizable?

Abstract page for arXiv paper 2603.03612: Why Are Linear RNNs More Parallelizable?

arXiv - Machine Learning · 4 min ·
[2603.03371] Sleeper Cell: Injecting Latent Malice Temporal Backdoors into Tool-Using LLMs
Llms

[2603.03371] Sleeper Cell: Injecting Latent Malice Temporal Backdoors into Tool-Using LLMs

Abstract page for arXiv paper 2603.03371: Sleeper Cell: Injecting Latent Malice Temporal Backdoors into Tool-Using LLMs

arXiv - AI · 4 min ·
[2603.03597] NuMuon: Nuclear-Norm-Constrained Muon for Compressible LLM Training
Llms

[2603.03597] NuMuon: Nuclear-Norm-Constrained Muon for Compressible LLM Training

Abstract page for arXiv paper 2603.03597: NuMuon: Nuclear-Norm-Constrained Muon for Compressible LLM Training

arXiv - Machine Learning · 3 min ·
[2603.03538] Online Learnability of Chain-of-Thought Verifiers: Soundness and Completeness Trade-offs
Llms

[2603.03538] Online Learnability of Chain-of-Thought Verifiers: Soundness and Completeness Trade-offs

Abstract page for arXiv paper 2603.03538: Online Learnability of Chain-of-Thought Verifiers: Soundness and Completeness Trade-offs

arXiv - Machine Learning · 4 min ·
[2603.03535] Trade-offs in Ensembling, Merging and Routing Among Parameter-Efficient Experts
Llms

[2603.03535] Trade-offs in Ensembling, Merging and Routing Among Parameter-Efficient Experts

Abstract page for arXiv paper 2603.03535: Trade-offs in Ensembling, Merging and Routing Among Parameter-Efficient Experts

arXiv - Machine Learning · 3 min ·
[2603.03352] Perfect score on IPhO 2025 theory by Gemini agent
Llms

[2603.03352] Perfect score on IPhO 2025 theory by Gemini agent

Abstract page for arXiv paper 2603.03352: Perfect score on IPhO 2025 theory by Gemini agent

arXiv - AI · 3 min ·
[2603.03527] Logit-Level Uncertainty Quantification in Vision-Language Models for Histopathology Image Analysis
Llms

[2603.03527] Logit-Level Uncertainty Quantification in Vision-Language Models for Histopathology Image Analysis

Abstract page for arXiv paper 2603.03527: Logit-Level Uncertainty Quantification in Vision-Language Models for Histopathology Image Analysis

arXiv - Machine Learning · 4 min ·
[2603.03524] Test-Time Meta-Adaptation with Self-Synthesis
Llms

[2603.03524] Test-Time Meta-Adaptation with Self-Synthesis

Abstract page for arXiv paper 2603.03524: Test-Time Meta-Adaptation with Self-Synthesis

arXiv - AI · 3 min ·
[2603.03517] MMAI Gym for Science: Training Liquid Foundation Models for Drug Discovery
Llms

[2603.03517] MMAI Gym for Science: Training Liquid Foundation Models for Drug Discovery

Abstract page for arXiv paper 2603.03517: MMAI Gym for Science: Training Liquid Foundation Models for Drug Discovery

arXiv - AI · 4 min ·
[2603.03332] Fragile Thoughts: How Large Language Models Handle Chain-of-Thought Perturbations
Llms

[2603.03332] Fragile Thoughts: How Large Language Models Handle Chain-of-Thought Perturbations

Abstract page for arXiv paper 2603.03332: Fragile Thoughts: How Large Language Models Handle Chain-of-Thought Perturbations

arXiv - AI · 4 min ·
[2603.03330] Certainty robustness: Evaluating LLM stability under self-challenging prompts
Llms

[2603.03330] Certainty robustness: Evaluating LLM stability under self-challenging prompts

Abstract page for arXiv paper 2603.03330: Certainty robustness: Evaluating LLM stability under self-challenging prompts

arXiv - AI · 3 min ·
[2603.03329] AutoHarness: improving LLM agents by automatically synthesizing a code harness
Llms

[2603.03329] AutoHarness: improving LLM agents by automatically synthesizing a code harness

Abstract page for arXiv paper 2603.03329: AutoHarness: improving LLM agents by automatically synthesizing a code harness

arXiv - AI · 4 min ·
[2603.03328] StructLens: A Structural Lens for Language Models via Maximum Spanning Trees
Llms

[2603.03328] StructLens: A Structural Lens for Language Models via Maximum Spanning Trees

Abstract page for arXiv paper 2603.03328: StructLens: A Structural Lens for Language Models via Maximum Spanning Trees

arXiv - AI · 3 min ·
[2603.03326] Controllable and explainable personality sliders for LLMs at inference time
Llms

[2603.03326] Controllable and explainable personality sliders for LLMs at inference time

Abstract page for arXiv paper 2603.03326: Controllable and explainable personality sliders for LLMs at inference time

arXiv - AI · 3 min ·
Previous Page 257 Next

Related Topics

Stay updated with AI News

Get the latest news, tools, and insights delivered to your inbox.

Daily or weekly digest • Unsubscribe anytime