[P] If you're building AI agents, logs aren't enough. You need evidence.
I have built a programmable governance layer for AI agents. I am considering to open source completely. Looking for feedback. Agent demos...
Alignment, bias, regulation, and responsible AI
I have built a programmable governance layer for AI agents. I am considering to open source completely. Looking for feedback. Agent demos...
Abstract page for arXiv paper 2510.14628: RLAIF-SPA: Structured AI Feedback for Semantic-Prosodic Alignment in Speech Synthesis
Abstract page for arXiv paper 2504.05995: NativQA Framework: Enabling LLMs and VLMs with Native, Local, and Everyday Knowledge
This paper presents a framework for bounding probabilities of causation using partial causal diagrams, addressing limitations of existing...
This technical report presents a comprehensive risk analysis framework for frontier AI, focusing on emerging threats and mitigation strat...
The paper introduces Precedent-Informed Reasoning (PIR) to enhance reasoning in Large Language Models (LLMs) by leveraging past cases, im...
This paper explores the stability of minimum-norm interpolating deep ReLU networks, identifying conditions under which these networks mai...
The paper presents sleep2vec, a model for aligning diverse nocturnal biosignals to improve sleep staging and clinical assessments, addres...
This paper explores how competition for attention in AI systems can lead to tipping points from beneficial to harmful outcomes, providing...
This article discusses the challenges of benchmarking Large Language Models (LLMs) as they reach new performance levels, introducing a fr...
The paper introduces GRAIL, a method for recognizing agent goals through imitation learning, enhancing goal recognition accuracy in AI sy...
The paper introduces MechPert, a framework that enhances unseen genetic perturbation prediction by leveraging mechanistic consensus among...
The paper presents the ForesightSafety Bench, a comprehensive framework for evaluating AI safety risks, addressing limitations in current...
The paper 'NEST: Nascent Encoded Steganographic Thoughts' explores the potential for large language models (LLMs) to conceal reasoning wi...
The paper presents GUI-GENESIS, a framework for automating the synthesis of efficient training environments for GUI agents, enhancing per...
The paper presents the REAL framework, which addresses knowledge conflicts in Knowledge-Intensive Visual Question Answering (KI-VQA) by i...
This paper introduces Head Entropy, a method for predicting answer correctness in large language models (LLMs) by analyzing attention ent...
This article discusses the integration of AI in clinical diagnostics, focusing on the use of abductive explanations to enhance AI's align...
This paper presents the Optimized Certainty Equivalent Risk-Controlling Prediction Sets (OCE-RCPS), a framework designed to enhance relia...
This paper presents a Physics-guided Causal Model for trajectory prediction in autonomous driving, focusing on zero-shot generalization a...
This paper introduces the concept of cumulative utility parity in federated learning, addressing fairness in client participation, partic...
This paper discusses the identification and diagnosis of pathological chain-of-thought reasoning in AI models, highlighting three specifi...
The paper discusses the need for claim-level auditability in deep research agents, highlighting the shift from factual errors to weak cla...
Get the latest news, tools, and insights delivered to your inbox.
Daily or weekly digest • Unsubscribe anytime