OpenAI introduces new 'Trusted Contact' safeguard for cases of possible self-harm | TechCrunch
The company is expanding its efforts to protect ChatGPT users in cases where conversations may turn to self-harm.
GPT, Claude, Gemini, and other LLMs
The company is expanding its efforts to protect ChatGPT users in cases where conversations may turn to self-harm.
Thanks to Musk v. Altman, the public is getting a concrete look at details of Sam Altman’s ouster from OpenAI, much of it centered on for...
I’m not a machine learning expert or anything, but I do enjoy learning about how it all works. I’ve noticed that one of the main limitati...
Abstract page for arXiv paper 2506.20746: Dynamic Weight Grafting: Localizing Finetuned Factual Knowledge in Transformers
Abstract page for arXiv paper 2506.15872: Hidden Breakthroughs in Language Model Training
Abstract page for arXiv paper 2508.11999: MOON: Generative MLLM-based Multimodal Representation Learning for E-commerce Product Understan...
Abstract page for arXiv paper 2508.06526: PiKV: KV Cache Management System for Mixture of Experts
Abstract page for arXiv paper 2506.15307: SecP-Tuning: Efficient Privacy-Preserving Prompt Tuning for Large Language Models via MPC
Abstract page for arXiv paper 2506.14003: Unlearning Isn't Invisible: Detecting Unlearning Traces in LLMs from Model Outputs
Abstract page for arXiv paper 2507.15852: Advancing Complex Video Object Segmentation via Progressive Concept Construction
Abstract page for arXiv paper 2507.04219: Model Collapse Is Not a Bug but a Feature in Machine Unlearning for LLMs
Abstract page for arXiv paper 2506.02939: QKV Projections Require a Fraction of Their Memory
Abstract page for arXiv paper 2506.20666: Cognitive models can reveal interpretable value trade-offs in language models
Abstract page for arXiv paper 2506.18841: LongWriter-Zero: Mastering Ultra-Long Text Generation via Reinforcement Learning
Abstract page for arXiv paper 2505.19590: Learning to Reason without External Rewards
Abstract page for arXiv paper 2506.13474: Language Agents for Hypothesis-driven Clinical Decision Making with Reinforcement Learning
Abstract page for arXiv paper 2506.15498: SPARE: Single-Pass Annotation with Reference-Guided Evaluation for Automatic Process Supervisio...
Abstract page for arXiv paper 2505.18116: NFT: Bridging Supervised Learning and Reinforcement Learning in Math Reasoning
Abstract page for arXiv paper 2506.10085: VITA: Zero-Shot Value Functions via Test-Time Adaptation of Vision-Language Models
Abstract page for arXiv paper 2505.16122: Plan and Budget: Effective and Efficient Test-Time Scaling on Reasoning Large Language Models
Abstract page for arXiv paper 2505.14042: Adversarially Pretrained Transformers May Be Universally Robust In-Context Learners
Abstract page for arXiv paper 2506.08902: Intention-Conditioned Flow Occupancy Models
Abstract page for arXiv paper 2506.06683: RoboPARA: Dual-Arm Robot Planning with Parallel Allocation and Recomposition Across Tasks
Get the latest news, tools, and insights delivered to your inbox.
Daily or weekly digest • Unsubscribe anytime