Large Language Models
GPT, Claude, Gemini, and other LLMs
Top This Week
All Content
[P] I cut my Claude Code token usage by using HDC as a context engine for my source tree
If you’re running Claude Code or Kiro regularly, you’re probably burning a few million tokens a week just on development. I’ve been build...
ChatGPT has experimented with watermarking AI text — 5 ways to use AI without sounding like it
ChatGPT has explored watermarking AI text — here are 5 simple ways to use AI without losing your voice or sounding like everyone else.
The Pentagon is making plans for AI companies to train on classified data, defense official says | MIT Technology Review
The generative AI models used in classified environments can answer questions, but don't currently learn from the data they see. Tha...
[2512.21323] Parallel Token Prediction for Language Models
Abstract page for arXiv paper 2512.21323: Parallel Token Prediction for Language Models
[2512.21039] Agentic Multi-Persona Framework for Evidence-Aware Fake News Detection
Abstract page for arXiv paper 2512.21039: Agentic Multi-Persona Framework for Evidence-Aware Fake News Detection
[2510.02282] VidGuard-R1: AI-Generated Video Detection and Explanation via Reasoning MLLMs and RL
Abstract page for arXiv paper 2510.02282: VidGuard-R1: AI-Generated Video Detection and Explanation via Reasoning MLLMs and RL
[2508.18088] How Quantization Shapes Bias in Large Language Models
Abstract page for arXiv paper 2508.18088: How Quantization Shapes Bias in Large Language Models
[2508.11847] Dropping Just a Handful of Preferences Can Change Top Large Language Model Rankings
Abstract page for arXiv paper 2508.11847: Dropping Just a Handful of Preferences Can Change Top Large Language Model Rankings
[2506.08762] EDINET-Bench: Evaluating LLMs on Complex Financial Tasks using Japanese Financial Statements
Abstract page for arXiv paper 2506.08762: EDINET-Bench: Evaluating LLMs on Complex Financial Tasks using Japanese Financial Statements
[2601.18734] Self-Distilled Reasoner: On-Policy Self-Distillation for Large Language Models
Abstract page for arXiv paper 2601.18734: Self-Distilled Reasoner: On-Policy Self-Distillation for Large Language Models
[2512.07419] Revolutionizing Mixed Precision Quantization: Towards Training-free Automatic Proxy Discovery via Large Language Models
Abstract page for arXiv paper 2512.07419: Revolutionizing Mixed Precision Quantization: Towards Training-free Automatic Proxy Discovery v...
[2510.17276] Breaking and Fixing Defenses Against Control-Flow Hijacking in Multi-Agent Systems
Abstract page for arXiv paper 2510.17276: Breaking and Fixing Defenses Against Control-Flow Hijacking in Multi-Agent Systems
[2509.25762] OPPO: Accelerating PPO-based RLHF via Pipeline Overlap
Abstract page for arXiv paper 2509.25762: OPPO: Accelerating PPO-based RLHF via Pipeline Overlap
[2508.02833] TIC-GRPO: Provable and Efficient Optimization for Reinforcement Learning from Human Feedback
Abstract page for arXiv paper 2508.02833: TIC-GRPO: Provable and Efficient Optimization for Reinforcement Learning from Human Feedback
[2506.09016] SPEED-RL: Faster Training of Reasoning Models via Online Curriculum Learning
Abstract page for arXiv paper 2506.09016: SPEED-RL: Faster Training of Reasoning Models via Online Curriculum Learning
[2505.23648] Continuous Chain of Thought Enables Parallel Exploration and Reasoning
Abstract page for arXiv paper 2505.23648: Continuous Chain of Thought Enables Parallel Exploration and Reasoning
[2603.05280] Layer by layer, module by module: Choose both for optimal OOD probing of ViT
Abstract page for arXiv paper 2603.05280: Layer by layer, module by module: Choose both for optimal OOD probing of ViT
[2603.05143] Feature Resemblance: On the Theoretical Understanding of Analogical Reasoning in Transformers
Abstract page for arXiv paper 2603.05143: Feature Resemblance: On the Theoretical Understanding of Analogical Reasoning in Transformers
[2603.05035] Good-Enough LLM Obfuscation (GELO)
Abstract page for arXiv paper 2603.05035: Good-Enough LLM Obfuscation (GELO)
[2603.05026] RepoLaunch: Automating Build&Test Pipeline of Code Repositories on ANY Language and ANY Platform
Abstract page for arXiv paper 2603.05026: RepoLaunch: Automating Build&Test Pipeline of Code Repositories on ANY Language and ANY Platform
Related Topics
Stay updated with AI News
Get the latest news, tools, and insights delivered to your inbox.
Daily or weekly digest • Unsubscribe anytime