Large Language Models

GPT, Claude, Gemini, and other LLMs

Top This Week

Llms

main skill in software engineering in 2026 is knowing what to ask Claude, not knowing how to code. and I can’t decide if that’s depressing or just the next abstraction layer.

Been writing code professionally for 8+ years. I’m now mass spending more time describing features in plain english than writing actual c...

Reddit - Artificial Intelligence · 1 min ·
Llms

Can we even achieve AGI with LLMs, why do AI bros still believe we can?

I've heard mixed discussions around this. Although not much evidence just rhetoric from the AGI will come from LLMs camp. submitted by /u...

Reddit - Artificial Intelligence · 1 min ·
Llms

You can now prompt OpenClaw into existence. fully 1st party on top of Claude Code

OpenClaw is basically banned from Claude ¯_(ツ)_/¯ Claude Code has Telegram support.. so what if we just, made it always stay on? turns ou...

Reddit - Artificial Intelligence · 1 min ·

All Content

[2512.21323] Parallel Token Prediction for Language Models
Llms

[2512.21323] Parallel Token Prediction for Language Models

Abstract page for arXiv paper 2512.21323: Parallel Token Prediction for Language Models

arXiv - Machine Learning · 3 min ·
[2512.21039] Agentic Multi-Persona Framework for Evidence-Aware Fake News Detection
Llms

[2512.21039] Agentic Multi-Persona Framework for Evidence-Aware Fake News Detection

Abstract page for arXiv paper 2512.21039: Agentic Multi-Persona Framework for Evidence-Aware Fake News Detection

arXiv - Machine Learning · 3 min ·
[2510.02282] VidGuard-R1: AI-Generated Video Detection and Explanation via Reasoning MLLMs and RL
Llms

[2510.02282] VidGuard-R1: AI-Generated Video Detection and Explanation via Reasoning MLLMs and RL

Abstract page for arXiv paper 2510.02282: VidGuard-R1: AI-Generated Video Detection and Explanation via Reasoning MLLMs and RL

arXiv - Machine Learning · 4 min ·
[2508.18088] How Quantization Shapes Bias in Large Language Models
Llms

[2508.18088] How Quantization Shapes Bias in Large Language Models

Abstract page for arXiv paper 2508.18088: How Quantization Shapes Bias in Large Language Models

arXiv - Machine Learning · 3 min ·
[2508.11847] Dropping Just a Handful of Preferences Can Change Top Large Language Model Rankings
Llms

[2508.11847] Dropping Just a Handful of Preferences Can Change Top Large Language Model Rankings

Abstract page for arXiv paper 2508.11847: Dropping Just a Handful of Preferences Can Change Top Large Language Model Rankings

arXiv - Machine Learning · 4 min ·
[2506.08762] EDINET-Bench: Evaluating LLMs on Complex Financial Tasks using Japanese Financial Statements
Llms

[2506.08762] EDINET-Bench: Evaluating LLMs on Complex Financial Tasks using Japanese Financial Statements

Abstract page for arXiv paper 2506.08762: EDINET-Bench: Evaluating LLMs on Complex Financial Tasks using Japanese Financial Statements

arXiv - Machine Learning · 4 min ·
[2601.18734] Self-Distilled Reasoner: On-Policy Self-Distillation for Large Language Models
Llms

[2601.18734] Self-Distilled Reasoner: On-Policy Self-Distillation for Large Language Models

Abstract page for arXiv paper 2601.18734: Self-Distilled Reasoner: On-Policy Self-Distillation for Large Language Models

arXiv - Machine Learning · 4 min ·
[2512.07419] Revolutionizing Mixed Precision Quantization: Towards Training-free Automatic Proxy Discovery via Large Language Models
Llms

[2512.07419] Revolutionizing Mixed Precision Quantization: Towards Training-free Automatic Proxy Discovery via Large Language Models

Abstract page for arXiv paper 2512.07419: Revolutionizing Mixed Precision Quantization: Towards Training-free Automatic Proxy Discovery v...

arXiv - Machine Learning · 4 min ·
[2510.17276] Breaking and Fixing Defenses Against Control-Flow Hijacking in Multi-Agent Systems
Llms

[2510.17276] Breaking and Fixing Defenses Against Control-Flow Hijacking in Multi-Agent Systems

Abstract page for arXiv paper 2510.17276: Breaking and Fixing Defenses Against Control-Flow Hijacking in Multi-Agent Systems

arXiv - Machine Learning · 4 min ·
[2509.25762] OPPO: Accelerating PPO-based RLHF via Pipeline Overlap
Llms

[2509.25762] OPPO: Accelerating PPO-based RLHF via Pipeline Overlap

Abstract page for arXiv paper 2509.25762: OPPO: Accelerating PPO-based RLHF via Pipeline Overlap

arXiv - Machine Learning · 3 min ·
[2508.02833] TIC-GRPO: Provable and Efficient Optimization for Reinforcement Learning from Human Feedback
Llms

[2508.02833] TIC-GRPO: Provable and Efficient Optimization for Reinforcement Learning from Human Feedback

Abstract page for arXiv paper 2508.02833: TIC-GRPO: Provable and Efficient Optimization for Reinforcement Learning from Human Feedback

arXiv - Machine Learning · 4 min ·
[2506.09016] SPEED-RL: Faster Training of Reasoning Models via Online Curriculum Learning
Llms

[2506.09016] SPEED-RL: Faster Training of Reasoning Models via Online Curriculum Learning

Abstract page for arXiv paper 2506.09016: SPEED-RL: Faster Training of Reasoning Models via Online Curriculum Learning

arXiv - Machine Learning · 3 min ·
[2505.23648] Continuous Chain of Thought Enables Parallel Exploration and Reasoning
Llms

[2505.23648] Continuous Chain of Thought Enables Parallel Exploration and Reasoning

Abstract page for arXiv paper 2505.23648: Continuous Chain of Thought Enables Parallel Exploration and Reasoning

arXiv - Machine Learning · 4 min ·
[2603.05280] Layer by layer, module by module: Choose both for optimal OOD probing of ViT
Llms

[2603.05280] Layer by layer, module by module: Choose both for optimal OOD probing of ViT

Abstract page for arXiv paper 2603.05280: Layer by layer, module by module: Choose both for optimal OOD probing of ViT

arXiv - Machine Learning · 3 min ·
[2603.05143] Feature Resemblance: On the Theoretical Understanding of Analogical Reasoning in Transformers
Llms

[2603.05143] Feature Resemblance: On the Theoretical Understanding of Analogical Reasoning in Transformers

Abstract page for arXiv paper 2603.05143: Feature Resemblance: On the Theoretical Understanding of Analogical Reasoning in Transformers

arXiv - Machine Learning · 3 min ·
[2603.05035] Good-Enough LLM Obfuscation (GELO)
Llms

[2603.05035] Good-Enough LLM Obfuscation (GELO)

Abstract page for arXiv paper 2603.05035: Good-Enough LLM Obfuscation (GELO)

arXiv - Machine Learning · 4 min ·
[2603.05026] RepoLaunch: Automating Build&Test Pipeline of Code Repositories on ANY Language and ANY Platform
Llms

[2603.05026] RepoLaunch: Automating Build&Test Pipeline of Code Repositories on ANY Language and ANY Platform

Abstract page for arXiv paper 2603.05026: RepoLaunch: Automating Build&Test Pipeline of Code Repositories on ANY Language and ANY Platform

arXiv - Machine Learning · 3 min ·
[2603.04964] Replaying pre-training data improves fine-tuning
Llms

[2603.04964] Replaying pre-training data improves fine-tuning

Abstract page for arXiv paper 2603.04964: Replaying pre-training data improves fine-tuning

arXiv - Machine Learning · 3 min ·
[2603.04716] SLO-Aware Compute Resource Allocation for Prefill-Decode Disaggregated LLM Inference
Llms

[2603.04716] SLO-Aware Compute Resource Allocation for Prefill-Decode Disaggregated LLM Inference

Abstract page for arXiv paper 2603.04716: SLO-Aware Compute Resource Allocation for Prefill-Decode Disaggregated LLM Inference

arXiv - Machine Learning · 4 min ·
[2603.04480] AbAffinity: A Large Language Model for Predicting Antibody Binding Affinity against SARS-CoV-2
Llms

[2603.04480] AbAffinity: A Large Language Model for Predicting Antibody Binding Affinity against SARS-CoV-2

Abstract page for arXiv paper 2603.04480: AbAffinity: A Large Language Model for Predicting Antibody Binding Affinity against SARS-CoV-2

arXiv - Machine Learning · 3 min ·
Previous Page 115 Next

Related Topics

Stay updated with AI News

Get the latest news, tools, and insights delivered to your inbox.

Daily or weekly digest • Unsubscribe anytime