The state of AI safety in four fake graphs
submitted by /u/tekz [link] [comments]
Alignment, bias, regulation, and responsible AI
submitted by /u/tekz [link] [comments]
Abstract page for arXiv paper 2603.14267: DiFlowDubber: Discrete Flow Matching for Automated Video Dubbing via Cross-Modal Alignment and ...
Abstract page for arXiv paper 2601.22440: AI and My Values: User Perceptions of LLMs' Ability to Extract, Embody, and Explain Human Value...
Abstract page for arXiv paper 2603.00078: Alignment Is Not Enough: A Relational Framework for Moral Standing in Human-AI Interaction
Abstract page for arXiv paper 2603.00068: The Global Landscape of Environmental AI Regulation: From the Cost of Reasoning to a Right to G...
Abstract page for arXiv paper 2603.00066: Contesting Artificial Moral Agents
Abstract page for arXiv paper 2603.00057: "Bespoke Bots": Diverse Instructor Needs for Customizing Generative AI Classroom Chatbots
Abstract page for arXiv paper 2603.00047: What Is the Geometry of the Alignment Tax?
Abstract page for arXiv paper 2603.00042: Maximizing the Spectral Energy Gain in Sub-1-Bit LLMs via Latent Geometry Alignment
Abstract page for arXiv paper 2603.00024: Personalization Increases Affective Alignment but Has Role-Dependent Effects on Epistemic Indep...
Abstract page for arXiv paper 2603.02203: Tool Verification for Test-Time Reinforcement Learning
Abstract page for arXiv paper 2603.01630: SEED-SET: Scalable Evolving Experimental Design for System-level Ethical Testing
Abstract page for arXiv paper 2603.01620: ToolRLA: Fine-Grained Reward Decomposition for Tool-Integrated Reinforcement Learning Alignment...
Abstract page for arXiv paper 2603.01562: RubricBench: Aligning Model-Generated Rubrics with Human Standards
Abstract page for arXiv paper 2603.01396: HarmonyCell: Automating Single-Cell Perturbation Modeling under Semantic and Distribution Shifts
Abstract page for arXiv paper 2603.01290: Opponent State Inference Under Partial Observability: An HMM-POMDP Framework for 2026 Formula 1...
Abstract page for arXiv paper 2603.00993: CollabEval: Enhancing LLM-as-a-Judge via Multi-Agent Collaboration
Abstract page for arXiv paper 2603.00590: Fair in Mind, Fair in Action? A Synchronous Benchmark for Understanding and Generation in UMLLMs
A recent ICLR paper proposes Behavior Learning — replacing neural layers with learnable constrained optimization blocks. It models it as:...
Abstract page for arXiv paper 2602.03775: An Empirical Study of Collective Behaviors and Social Dynamics in Large Language Model Agents
Abstract page for arXiv paper 2502.01383: InfoBridge: Mutual Information estimation via Bridge Matching
Abstract page for arXiv paper 2509.23371: Alignment through Meta-Weighted Online Sampling: Bridging the Gap between Data Generation and P...
Abstract page for arXiv paper 2505.19441: Fairness-in-the-Workflow: How Machine Learning Practitioners at Big Tech Companies Approach Fai...
Get the latest news, tools, and insights delivered to your inbox.
Daily or weekly digest • Unsubscribe anytime