Open-source diagnostic for AI misalignment. Model agnostic, industry agnostic. Free to Run.
We shipped iFixAi earlier this week. An open-source diagnostic for AI misalignment. 32 tests across fabrication, manipulation, deception,...
ML algorithms, training, and inference
We shipped iFixAi earlier this week. An open-source diagnostic for AI misalignment. 32 tests across fabrication, manipulation, deception,...
For the past several years I've been quietly assembling and processing what I believe is one of the larger privately held pretraining cor...
Watch as the Equity podcast team discusses what's actually at stake in the courtroom and what to watch for as Altman and others take the ...
Abstract page for arXiv paper 2509.23279: Vid-Freeze: Protecting Images from Malicious Image-to-Video Generation via Temporal Freezing
Abstract page for arXiv paper 2509.22258: Beyond Classification Accuracy: Neural-MedBench and the Need for Deeper Reasoning Benchmarks
Abstract page for arXiv paper 2509.05892: Challenges in Deep Learning-Based Small Organ Segmentation: A Benchmarking Perspective for Medi...
Abstract page for arXiv paper 2506.13130: ZINA: Multimodal Fine-grained Hallucination Detection and Editing
Abstract page for arXiv paper 2506.09749: Large Language Models for Combinatorial Optimization of Design Structure Matrix
Abstract page for arXiv paper 2505.15925: VERDI: VLM-Embedded Reasoning for Autonomous Driving
Abstract page for arXiv paper 2503.12575: BalancedDPO: Adaptive Multi-Metric Alignment
Abstract page for arXiv paper 2503.11572: Implicit Bias-Like Patterns in Reasoning Models
Abstract page for arXiv paper 2501.11782: Human-AI Collaborative Game Testing with Vision Language Models
Abstract page for arXiv paper 2501.07813: Talk to Right Specialists: Iterative Routing in Multi-agent Systems for Question Answering
Abstract page for arXiv paper 2408.11871: MegaFake: A Theory-Driven Dataset of Fake News Generated by Large Language Models
Abstract page for arXiv paper 2406.14194: VLBiasBench: A Comprehensive Benchmark for Evaluating Bias in Large Vision-Language Model
Abstract page for arXiv paper 2604.01438: ClawSafety: "Safe" LLMs, Unsafe Agents
Abstract page for arXiv paper 2603.18633: An Onto-Relational-Sophic Framework for Governing Synthetic Minds
Abstract page for arXiv paper 2603.09127: Collective AI can amplify tiny perturbations into divergent decisions
Abstract page for arXiv paper 2602.07943: IV Co-Scientist: Multi-Agent LLM Framework for Causal Instrumental Variable Discovery
Abstract page for arXiv paper 2602.03151: Enhancing Foundation VLM Robustness to Missing Modality: Scalable Diffusion for Bi-directional ...
Abstract page for arXiv paper 2601.22776: TSPO: Breaking the Double Homogenization Dilemma in Multi-turn Search Policy Optimization
Abstract page for arXiv paper 2601.21439: The Paradox of Robustness: Decoupling Rule-Based Logic from Affective Noise in High-Stakes Deci...
Abstract page for arXiv paper 2511.16383: An Agent-Based Framework for the Automatic Validation of Mathematical Optimization Models
Get the latest news, tools, and insights delivered to your inbox.
Daily or weekly digest • Unsubscribe anytime