Newsom signs executive order requiring AI companies to have safety, privacy guardrails
submitted by /u/Fcking_Chuck [link] [comments]
Alignment, bias, regulation, and responsible AI
submitted by /u/Fcking_Chuck [link] [comments]
Abstract page for arXiv paper 2511.16417: Pharos-ESG: A Framework for Multimodal Parsing, Contextual Narration, and Hierarchical Labeling...
Abstract page for arXiv paper 2510.08847: What Is Your Agent's GPA? A Framework for Evaluating Agent Goal-Plan-Action Alignment
This article evaluates transfer learning models for IoT DDoS detection, focusing on explainability and resource constraints. It analyzes ...
This article explores the relationship between AI and humans through the lens of large language models (LLMs), focusing on the Sydney per...
The paper discusses the security risks posed by implicit prompt injection in large language model (LLM) agents, demonstrating how adversa...
This paper presents a novel framework combining BanglaBERT and a two-layer stacked LSTM for effective multi-label cyberbullying detection...
The paper presents HubScan, a tool designed to detect hubness poisoning in Retrieval-Augmented Generation systems, addressing a critical ...
This paper presents a novel approach to differentially private data truncation using public second moments, enhancing privacy without com...
This article presents a framework called DiSP (Diffusion Self-Purification) to mitigate backdoor attacks in Multimodal Diffusion Language...
This article discusses the application of foundation models in histopathology, highlighting a novel approach that improves robustness and...
The article presents CrossLLM-Mamba, a novel framework for RNA interaction prediction that utilizes multimodal state space fusion of larg...
The paper introduces SOTAlign, a semi-supervised framework for aligning unimodal vision and language models using minimal paired data and...
The paper 'Poisoned Acoustics' explores training-data poisoning attacks on deep neural networks, demonstrating significant vulnerabilitie...
This paper presents a novel differentiable approximation to the zero-one loss, enhancing gradient-based optimization in machine learning ...
This article presents FedWQ-CP, a novel approach to federated uncertainty quantification that addresses dual heterogeneity in data and mo...
This paper analyzes the vulnerabilities of Large Language Models (LLMs) to prompt injection and jailbreak attacks, evaluating various def...
The paper presents TT-SEAL, a selective encryption framework designed for Tensor-Train Decomposed (TTD) networks, enhancing security and ...
This article presents a novel approach for unsupervised denoising of diffusion-weighted images (dMRI) by addressing noise bias and varian...
This article evaluates misinformation exposure on the Chinese web by comparing traditional search engines, LLMs, and AI-generated overvie...
The paper presents MetaOthello, a study exploring how transformers manage multiple world models through a controlled suite of Othello var...
The paper presents new disagreement-based certificates for generalization bounds in deep learning models, addressing limitations of exist...
This paper explores contextual online Reinforcement Learning with Human Feedback (RLHF) using a Generalized Bilinear Preference Model to ...
Get the latest news, tools, and insights delivered to your inbox.
Daily or weekly digest • Unsubscribe anytime