AI Agents

Autonomous agents, tool use, and agentic systems

Top This Week

Ai Agents

Microsoft's newest open-source project: Runtime security for AI agents

submitted by /u/Fcking_Chuck [link] [comments]

Reddit - Artificial Intelligence · 1 min ·
[2510.16609] Prior Knowledge Makes It Possible: From Sublinear Graph Algorithms to LLM Test-Time Methods
Llms

[2510.16609] Prior Knowledge Makes It Possible: From Sublinear Graph Algorithms to LLM Test-Time Methods

Abstract page for arXiv paper 2510.16609: Prior Knowledge Makes It Possible: From Sublinear Graph Algorithms to LLM Test-Time Methods

arXiv - Machine Learning · 4 min ·
[2604.02131] Intelligent Cloud Orchestration: A Hybrid Predictive and Heuristic Framework for Cost Optimization
Machine Learning

[2604.02131] Intelligent Cloud Orchestration: A Hybrid Predictive and Heuristic Framework for Cost Optimization

Abstract page for arXiv paper 2604.02131: Intelligent Cloud Orchestration: A Hybrid Predictive and Heuristic Framework for Cost Optimization

arXiv - Machine Learning · 3 min ·

All Content

[2602.21420] Overconfident Errors Need Stronger Correction: Asymmetric Confidence Penalties for Reinforcement Learning
Llms

[2602.21420] Overconfident Errors Need Stronger Correction: Asymmetric Confidence Penalties for Reinforcement Learning

This paper introduces the Asymmetric Confidence-aware Error Penalty (ACE) to enhance reinforcement learning by addressing overconfident e...

arXiv - Machine Learning · 4 min ·
[2602.21401] The Headless Firm: How AI Reshapes Enterprise Boundaries
Ai Agents

[2602.21401] The Headless Firm: How AI Reshapes Enterprise Boundaries

The paper discusses how AI is transforming enterprise structures by reducing coordination costs, leading to the emergence of the 'Headles...

arXiv - AI · 4 min ·
[2602.21368] Black-Box Reliability Certification for AI Agents via Self-Consistency Sampling and Conformal Calibration
Ai Infrastructure

[2602.21368] Black-Box Reliability Certification for AI Agents via Self-Consistency Sampling and Conformal Calibration

This paper presents a method for certifying the reliability of black-box AI systems using self-consistency sampling and conformal calibra...

arXiv - Machine Learning · 3 min ·
[2602.21360] Representation Theorems for Cumulative Propositional Dependence Logics
Machine Learning

[2602.21360] Representation Theorems for Cumulative Propositional Dependence Logics

This paper presents representation theorems for cumulative propositional dependence logic and team semantics, establishing key equivalenc...

arXiv - AI · 3 min ·
[2602.21327] Equitable Evaluation via Elicitation
Ai Startups

[2602.21327] Equitable Evaluation via Elicitation

The paper discusses an AI-driven approach for equitable skill evaluation, addressing biases in self-presentation among job seekers. It pr...

arXiv - Machine Learning · 3 min ·
[2602.21269] Group Orthogonalized Policy Optimization:Group Policy Optimization as Orthogonal Projection in Hilbert Space
Llms

[2602.21269] Group Orthogonalized Policy Optimization:Group Policy Optimization as Orthogonal Projection in Hilbert Space

The paper introduces Group Orthogonalized Policy Optimization (GOPO), a novel algorithm for aligning large language models using Hilbert ...

arXiv - Machine Learning · 4 min ·
[2602.21255] A General Equilibrium Theory of Orchestrated AI Agent Systems
Llms

[2602.21255] A General Equilibrium Theory of Orchestrated AI Agent Systems

This paper presents a general equilibrium theory for orchestrated AI agent systems, modeling large language model (LLM) agents within a p...

arXiv - AI · 4 min ·
[2602.21251] AgenticTyper: Automated Typing of Legacy Software Projects Using Agentic AI
Llms

[2602.21251] AgenticTyper: Automated Typing of Legacy Software Projects Using Agentic AI

AgenticTyper is a novel AI-driven tool that automates the typing of legacy JavaScript projects, significantly reducing manual effort and ...

arXiv - AI · 3 min ·
[2602.21231] ACAR: Adaptive Complexity Routing for Multi-Model Ensembles with Auditable Decision Traces
Llms

[2602.21231] ACAR: Adaptive Complexity Routing for Multi-Model Ensembles with Auditable Decision Traces

The paper presents ACAR, a framework for adaptive complexity routing in multi-model ensembles, demonstrating improved task routing accura...

arXiv - Machine Learning · 4 min ·
[2602.21228] ImpRIF: Stronger Implicit Reasoning Leads to Better Complex Instruction Following
Llms

[2602.21228] ImpRIF: Stronger Implicit Reasoning Leads to Better Complex Instruction Following

The paper presents ImpRIF, a method to enhance large language models' implicit reasoning capabilities, improving their performance in fol...

arXiv - AI · 4 min ·
[2602.21227] Budget-Aware Agentic Routing via Boundary-Guided Training
Llms

[2602.21227] Budget-Aware Agentic Routing via Boundary-Guided Training

The paper presents Budget-Aware Agentic Routing, a method for optimizing the use of large language models in autonomous agents by balanci...

arXiv - AI · 4 min ·
[2602.21220] Field-Theoretic Memory for AI Agents: Continuous Dynamics for Context Preservation
Ai Agents

[2602.21220] Field-Theoretic Memory for AI Agents: Continuous Dynamics for Context Preservation

The paper presents a novel memory system for AI agents, utilizing continuous fields governed by partial differential equations to enhance...

arXiv - Machine Learning · 3 min ·
[2602.22070] Language Models Exhibit Inconsistent Biases Towards Algorithmic Agents and Human Experts
Llms

[2602.22070] Language Models Exhibit Inconsistent Biases Towards Algorithmic Agents and Human Experts

This study explores how large language models (LLMs) exhibit inconsistent biases towards algorithmic agents and human experts in decision...

arXiv - AI · 4 min ·
[2602.22094] Petri Net Relaxation for Infeasibility Explanation and Sequential Task Planning
Ai Agents

[2602.22094] Petri Net Relaxation for Infeasibility Explanation and Sequential Task Planning

This paper presents a novel approach using Petri nets to identify infeasibilities in sequential task planning, enhancing robustness and e...

arXiv - AI · 3 min ·
[2602.21858] ProactiveMobile: A Comprehensive Benchmark for Boosting Proactive Intelligence on Mobile Devices
Llms

[2602.21858] ProactiveMobile: A Comprehensive Benchmark for Boosting Proactive Intelligence on Mobile Devices

The paper introduces ProactiveMobile, a benchmark aimed at enhancing proactive intelligence in mobile devices, addressing the limitations...

arXiv - AI · 4 min ·
[2602.22067] Semantic Partial Grounding via LLMs
Llms

[2602.22067] Semantic Partial Grounding via LLMs

The paper proposes SPG-LLM, a novel approach for semantic partial grounding in AI planning that utilizes large language models (LLMs) to ...

arXiv - AI · 3 min ·
[2602.21889] 2-Step Agent: A Framework for the Interaction of a Decision Maker with AI Decision Support
Machine Learning

[2602.21889] 2-Step Agent: A Framework for the Interaction of a Decision Maker with AI Decision Support

The paper presents the 2-Step Agent framework, which models the interaction between decision makers and AI decision support systems, high...

arXiv - Machine Learning · 3 min ·
[2602.21746] fEDM+: A Risk-Based Fuzzy Ethical Decision Making Framework with Principle-Level Explainability and Pluralistic Validation
Machine Learning

[2602.21746] fEDM+: A Risk-Based Fuzzy Ethical Decision Making Framework with Principle-Level Explainability and Pluralistic Validation

The paper presents fEDM+, an enhanced fuzzy ethical decision-making framework that improves explainability and validation by integrating ...

arXiv - AI · 4 min ·
[2602.21814] Prompt Architecture Determines Reasoning Quality: A Variable Isolation Study on the Car Wash Problem
Llms

[2602.21814] Prompt Architecture Determines Reasoning Quality: A Variable Isolation Study on the Car Wash Problem

This study investigates how different prompt architectures affect reasoning quality in large language models, specifically addressing the...

arXiv - AI · 3 min ·
[2602.21745] The ASIR Courage Model: A Phase-Dynamic Framework for Truth Transitions in Human and AI Systems
Machine Learning

[2602.21745] The ASIR Courage Model: A Phase-Dynamic Framework for Truth Transitions in Human and AI Systems

The ASIR Courage Model presents a phase-dynamic framework for understanding truth transitions in both human and AI systems, emphasizing t...

arXiv - AI · 4 min ·
Previous Page 54 Next

Related Topics

Stay updated with AI News

Get the latest news, tools, and insights delivered to your inbox.

Daily or weekly digest • Unsubscribe anytime