Claude Mythos and the Global South’s AI Governance Dilemma
Claude Mythos Preview is an experimental AI model built by Anthropic, the company behind the Claude family of models.
GPT, Claude, Gemini, and other LLMs
Claude Mythos Preview is an experimental AI model built by Anthropic, the company behind the Claude family of models.
I enabled Personal Intelligence, connected my Google apps, and now Gemini guesses what I want without me saying it.
In the months before he committed a grisly mass shooting, Phoenix Ikner obsessively used Open AI's ChatGPT.
Abstract page for arXiv paper 2511.00177: Can SAEs reveal and mitigate racial biases of LLMs in healthcare?
Abstract page for arXiv paper 2511.00405: UME-R1: Exploring Reasoning-Driven Generative Multimodal Embeddings
Abstract page for arXiv paper 2510.18560: WebDevJudge: Evaluating (M)LLMs as Critiques for Web Development Quality
Abstract page for arXiv paper 2510.15905: Digital Companionship: Overlapping Uses of AI Companions and AI Assistants
Abstract page for arXiv paper 2510.21910: Adversarial Déjà Vu: Jailbreak Dictionary Learning for Stronger Generalization to Unseen Attacks
Abstract page for arXiv paper 2510.15863: PolySkill: Learning Generalizable Skills Through Polymorphic Abstraction
Abstract page for arXiv paper 2510.20264: Optimistic Task Inference for Behavior Foundation Models
Abstract page for arXiv paper 2510.21314: A Convergence Analysis of Adaptive Optimizers under Floating-point Quantization
Abstract page for arXiv paper 2510.13888: Reliable Fine-Grained Evaluation of Natural Language Math Proofs
Abstract page for arXiv paper 2510.17206: Soft-Masked Diffusion Language Models
Abstract page for arXiv paper 2510.18245: Scaling Laws Meet Model Architecture: Toward Inference-Efficient LLMs
Abstract page for arXiv paper 2510.10066: OBsmith: LLM-Powered JavaScript Obfuscator Testing
Abstract page for arXiv paper 2510.09462: Adaptive Attacks on Trusted Monitors Subvert AI Control Protocols
Abstract page for arXiv paper 2510.13117: On the Reasoning Abilities of Masked Diffusion Language Models
Abstract page for arXiv paper 2510.07940: TTOM: Test-Time Optimization and Memorization for Compositional Video Generation
Abstract page for arXiv paper 2510.06377: Relational Transformer: Toward Zero-Shot Foundation Models for Relational Data
Abstract page for arXiv paper 2510.06292: ChainMPQ: Interleaved Text-Image Reasoning Chains for Mitigating Relation Hallucinations
Abstract page for arXiv paper 2510.05064: Boomerang Distillation Enables Zero-Shot Model Size Interpolation
Abstract page for arXiv paper 2510.05174: Emergent Coordination in Multi-Agent Language Models
Abstract page for arXiv paper 2510.05132: Training Large Language Models To Reason In Parallel With Global Forking Tokens
Get the latest news, tools, and insights delivered to your inbox.
Daily or weekly digest • Unsubscribe anytime