[2305.08175] ResidualPlanner+: a scalable matrix mechanism for marginals and beyond
Abstract page for arXiv paper 2305.08175: ResidualPlanner+: a scalable matrix mechanism for marginals and beyond
Alignment, bias, regulation, and responsible AI
Abstract page for arXiv paper 2305.08175: ResidualPlanner+: a scalable matrix mechanism for marginals and beyond
Abstract page for arXiv paper 2604.02610: Structure-Preserving Multi-View Embedding Using Gromov-Wasserstein Optimal Transport
Abstract page for arXiv paper 2604.02574: Understanding the Effects of Safety Unalignment on Large Language Models
The paper presents a novel framework for continual uncertainty learning in robust control of nonlinear dynamical systems, addressing chal...
This paper presents a novel approach to 3D scene rendering using multimodal Gaussian splatting, integrating RF sensing for improved accur...
The paper presents FLoRG, a federated fine-tuning framework that utilizes low-rank Gram matrices and Procrustes alignment to enhance the ...
This paper presents a delta method approach for sample size analysis in estimating probabilities of causation (PoCs), addressing the need...
The paper presents 'Wink', a system designed to recover coding agents from misbehaviors, enhancing their reliability in software developm...
This article explores the challenges and opportunities in overseeing AI agents without constant human oversight, focusing on user studies...
The paper presents HiVAE, a hierarchical variational architecture designed to enhance AI's theory of mind capabilities, enabling better i...
This paper explores how learning under noisy supervision is influenced by a feedback-truth gap, demonstrating its effects across various ...
This paper explores how reference-guided evaluators can enhance LLM alignment in non-verifiable domains, demonstrating significant improv...
This article discusses the use of large language models (LLMs) for deanonymizing online users, demonstrating high precision in identifyin...
LiveClin introduces a novel clinical benchmark for evaluating medical LLMs, addressing issues of data contamination and knowledge obsoles...
This study investigates whether adversarial code comments can mislead AI security reviewers during vulnerability detection in code, revea...
This article examines the stability of attention heads in transformer models, revealing insights into their representational robustness a...
The paper evaluates AI safety datasets, revealing they often misrepresent real-world attacks due to an overreliance on triggering cues, l...
This paper evaluates the reliability of Mamba, a state-space model, for medical imaging under various attack scenarios, highlighting vuln...
The paper introduces the AI Gamestore, a platform for evaluating machine general intelligence through human games, highlighting its poten...
This article presents a hybrid federated learning model that combines SWIN Transformer and CNN for diagnosing lung diseases, particularly...
The paper presents ODESteer, a novel ODE-based framework for aligning large language models (LLMs) by addressing limitations in existing ...
This article presents a benchmarking framework for optimizing AI models on ARM Cortex processors, focusing on energy efficiency and perfo...
This article proposes a Privacy by Design framework for AI applications targeting children, addressing privacy risks and compliance with ...
Get the latest news, tools, and insights delivered to your inbox.
Daily or weekly digest • Unsubscribe anytime