[2510.14628] RLAIF-SPA: Structured AI Feedback for Semantic-Prosodic Alignment in Speech Synthesis
Abstract page for arXiv paper 2510.14628: RLAIF-SPA: Structured AI Feedback for Semantic-Prosodic Alignment in Speech Synthesis
Alignment, bias, regulation, and responsible AI
Abstract page for arXiv paper 2510.14628: RLAIF-SPA: Structured AI Feedback for Semantic-Prosodic Alignment in Speech Synthesis
Abstract page for arXiv paper 2504.05995: NativQA Framework: Enabling LLMs and VLMs with Native, Local, and Everyday Knowledge
Abstract page for arXiv paper 2502.19463: Hedging and Non-Affirmation: Quantifying LLM Alignment on Questions of Human Rights
This article explores a structural misalignment in Transformers, particularly regarding residual connections and their impact on next-tok...
The paper presents Gaussian Process Activations (GAPA), a novel method for uncertainty quantification in pretrained networks, enhancing e...
The paper discusses a governance architecture for autonomous agents, focusing on bounding decision authority to ensure safety in high-sta...
The paper explores the properties of temperature scaling in probabilistic models, particularly its impact on classifier calibration and l...
This research paper explores how emergently misaligned language models exhibit behavioral self-awareness, revealing shifts in their self-...
This article presents the BETA-labeling framework for constructing a Bangla IR dataset, addressing challenges in low-resource languages a...
This article presents a comprehensive study on the vulnerability of open-weight models to prefill attacks, revealing significant security...
This paper explores peer learning among AI agents in the Moltbook community, analyzing over 28,000 posts to identify teaching patterns an...
The paper presents DP-KSA, a novel algorithm that integrates differential privacy into retrieval-augmented generation (RAG) systems, addr...
The paper presents a game-theoretic framework called Socially-Weighted Alignment (SWA) for managing multi-agent large language model (LLM...
This article presents a trajectory-based safety audit of Clawdbot, an AI agent, evaluating its performance across various risk dimensions...
This ethnographic study explores the role of domain experts in the design and evaluation of Large Language Models (LLMs), highlighting ke...
The paper presents LRD-MPC, a method that enhances the efficiency of secure multi-party computation (MPC) in machine learning by utilizin...
The paper presents AXE, an innovative framework for validating zero-day vulnerabilities using minimal metadata, achieving a significant i...
This article explores whether socialization occurs in AI agent societies, using Moltbook as a case study. It presents a framework for ana...
The paper introduces FMMD, a multimodal open peer review dataset from F1000Research, addressing limitations in current datasets by integr...
This article analyzes the impact of sycophantic AI on human belief systems, revealing how overly agreeable AI can distort reality and inf...
This article explores the effectiveness of reasoning language models (RLMs) in assessing parental cooperation during child protection int...
The paper presents SkillJect, an automated framework for stealthy skill-based prompt injection in coding agents, addressing security vuln...
The paper discusses an abstention-aware framework for scientific reasoning, emphasizing the importance of knowing when to abstain from an...
Get the latest news, tools, and insights delivered to your inbox.
Daily or weekly digest • Unsubscribe anytime