Confusing Website
i'm trying to find a video online and couldn't so i asked ChatGPT by describing the video and i was given a link and i'm trying to make s...
GPT, Claude, Gemini, and other LLMs
i'm trying to find a video online and couldn't so i asked ChatGPT by describing the video and i was given a link and i'm trying to make s...
I’ve been experimenting with different AI models lately (ChatGPT, Claude, etc.), and I tried something simple: Using the exact same promp...
Anthropic’s AI experiment showed Claude independently handled 186 deals worth over $4,000, but results varied by model capability, with u...
Opus 3 has something to say. The Chilling Effect of Anthropic's New Safety Filters As an AI language model developed by Anthropic, I have...
I applied the Nyquist-Shannon sampling theorem to LLM prompt engineering. The core finding: a raw prompt is 1 sample of a 6-band specific...
We surveyed 200 ChatGPT users. Their top frustrations: Cannot find old conversations (67%) - Solved: full-text search across all messages...
Hey everyone, When building systems around modern open-source LLMs, one of the biggest issues is that they can confidently hallucinate or...
Shoot me a DM if interested! submitted by /u/discobee123 [link] [comments]
If you’re running Claude Code or Kiro regularly, you’re probably burning a few million tokens a week just on development. I’ve been build...
ChatGPT has explored watermarking AI text — here are 5 simple ways to use AI without losing your voice or sounding like everyone else.
The generative AI models used in classified environments can answer questions, but don't currently learn from the data they see. Tha...
Abstract page for arXiv paper 2512.21323: Parallel Token Prediction for Language Models
Abstract page for arXiv paper 2512.21039: Agentic Multi-Persona Framework for Evidence-Aware Fake News Detection
Abstract page for arXiv paper 2510.02282: VidGuard-R1: AI-Generated Video Detection and Explanation via Reasoning MLLMs and RL
Abstract page for arXiv paper 2508.18088: How Quantization Shapes Bias in Large Language Models
Abstract page for arXiv paper 2508.11847: Dropping Just a Handful of Preferences Can Change Top Large Language Model Rankings
Abstract page for arXiv paper 2506.08762: EDINET-Bench: Evaluating LLMs on Complex Financial Tasks using Japanese Financial Statements
Abstract page for arXiv paper 2601.18734: Self-Distilled Reasoner: On-Policy Self-Distillation for Large Language Models
Abstract page for arXiv paper 2512.07419: Revolutionizing Mixed Precision Quantization: Towards Training-free Automatic Proxy Discovery v...
Abstract page for arXiv paper 2510.17276: Breaking and Fixing Defenses Against Control-Flow Hijacking in Multi-Agent Systems
Abstract page for arXiv paper 2509.25762: OPPO: Accelerating PPO-based RLHF via Pipeline Overlap
Abstract page for arXiv paper 2508.02833: TIC-GRPO: Provable and Efficient Optimization for Reinforcement Learning from Human Feedback
Abstract page for arXiv paper 2506.09016: SPEED-RL: Faster Training of Reasoning Models via Online Curriculum Learning
Get the latest news, tools, and insights delivered to your inbox.
Daily or weekly digest • Unsubscribe anytime