AI Infrastructure

GPUs, training clusters, MLOps, and deployment

Top This Week

[2601.16933] Reward-Forcing: Autoregressive Video Generation with Reward Feedback
Machine Learning

[2601.16933] Reward-Forcing: Autoregressive Video Generation with Reward Feedback

Abstract page for arXiv paper 2601.16933: Reward-Forcing: Autoregressive Video Generation with Reward Feedback

arXiv - Machine Learning · 3 min ·
[2511.03909] Tensor Computation of Euler Characteristic Functions and Transforms
Ai Infrastructure

[2511.03909] Tensor Computation of Euler Characteristic Functions and Transforms

Abstract page for arXiv paper 2511.03909: Tensor Computation of Euler Characteristic Functions and Transforms

arXiv - Machine Learning · 3 min ·
[2510.05497] Patterns behind Chaos: Forecasting Data Movement for Efficient Large-Scale MoE LLM Inference
Llms

[2510.05497] Patterns behind Chaos: Forecasting Data Movement for Efficient Large-Scale MoE LLM Inference

Abstract page for arXiv paper 2510.05497: Patterns behind Chaos: Forecasting Data Movement for Efficient Large-Scale MoE LLM Inference

arXiv - Machine Learning · 4 min ·

All Content

[2602.20497] LESA: Learnable Stage-Aware Predictors for Diffusion Model Acceleration
Machine Learning

[2602.20497] LESA: Learnable Stage-Aware Predictors for Diffusion Model Acceleration

The paper introduces LESA, a framework for accelerating diffusion models using learnable stage-aware predictors, achieving significant sp...

arXiv - AI · 4 min ·
[2602.20467] Elimination-compensation pruning for fully-connected neural networks
Machine Learning

[2602.20467] Elimination-compensation pruning for fully-connected neural networks

This paper introduces a novel pruning method for fully-connected neural networks, which compensates for the removal of weights by adjusti...

arXiv - Machine Learning · 4 min ·
[2602.20449] Protein Language Models Diverge from Natural Language: Comparative Analysis and Improved Inference
Llms

[2602.20449] Protein Language Models Diverge from Natural Language: Comparative Analysis and Improved Inference

This article explores the differences between protein language models (PLMs) and natural language models, highlighting how these distinct...

arXiv - Machine Learning · 4 min ·
[2602.20442] Imputation of Unknown Missingness in Sparse Electronic Health Records
Machine Learning

[2602.20442] Imputation of Unknown Missingness in Sparse Electronic Health Records

The paper presents a novel algorithm for imputing unknown missing values in sparse electronic health records (EHRs) using a transformer-b...

arXiv - Machine Learning · 4 min ·
[2602.20400] Three Concrete Challenges and Two Hopes for the Safety of Unsupervised Elicitation
Llms

[2602.20400] Three Concrete Challenges and Two Hopes for the Safety of Unsupervised Elicitation

This article discusses three significant challenges and two potential solutions for improving the safety of unsupervised elicitation in l...

arXiv - Machine Learning · 4 min ·
[2602.20361] Learning During Detection: Continual Learning for Neural OFDM Receivers via DMRS
Machine Learning

[2602.20361] Learning During Detection: Continual Learning for Neural OFDM Receivers via DMRS

This paper presents a continual learning framework for neural OFDM receivers that allows for real-time adaptation to changing communicati...

arXiv - Machine Learning · 3 min ·
[2602.20332] No One Size Fits All: QueryBandits for Hallucination Mitigation
Llms

[2602.20332] No One Size Fits All: QueryBandits for Hallucination Mitigation

The paper introduces QueryBandits, a model-agnostic framework designed to mitigate hallucinations in large language models (LLMs) by opti...

arXiv - Machine Learning · 4 min ·
[2602.20292] Quantifying the Expectation-Realisation Gap for Agentic AI Systems
Ai Infrastructure

[2602.20292] Quantifying the Expectation-Realisation Gap for Agentic AI Systems

This article examines the expectation-realisation gap in agentic AI systems, revealing discrepancies between anticipated productivity gai...

arXiv - AI · 3 min ·
[2602.20271] Uncertainty-Aware Delivery Delay Duration Prediction via Multi-Task Deep Learning
Machine Learning

[2602.20271] Uncertainty-Aware Delivery Delay Duration Prediction via Multi-Task Deep Learning

This paper presents a multi-task deep learning model for predicting delivery delay durations in logistics, addressing challenges posed by...

arXiv - Machine Learning · 4 min ·
[2602.20217] KnapSpec: Self-Speculative Decoding via Adaptive Layer Selection as a Knapsack Problem
Llms

[2602.20217] KnapSpec: Self-Speculative Decoding via Adaptive Layer Selection as a Knapsack Problem

The paper introduces KnapSpec, a framework for self-speculative decoding that optimizes layer selection in LLMs as a knapsack problem, en...

arXiv - Machine Learning · 4 min ·
[2602.20214] Right to History: A Sovereignty Kernel for Verifiable AI Agent Execution
Ai Safety

[2602.20214] Right to History: A Sovereignty Kernel for Verifiable AI Agent Execution

This paper proposes the 'Right to History,' a principle ensuring individuals have a verifiable record of AI agent actions on personal har...

arXiv - AI · 3 min ·
[2602.20208] Model Merging in the Essential Subspace
Machine Learning

[2602.20208] Model Merging in the Essential Subspace

This paper presents ESM, a novel framework for merging multiple task-specific models into a single multi-task model, addressing inter-tas...

arXiv - Machine Learning · 3 min ·
[2602.20207] Golden Layers and Where to Find Them: Improved Knowledge Editing for Large Language Models Via Layer Gradient Analysis
Llms

[2602.20207] Golden Layers and Where to Find Them: Improved Knowledge Editing for Large Language Models Via Layer Gradient Analysis

This article discusses the concept of 'golden layers' in large language models (LLMs) and presents a novel method, Layer Gradient Analysi...

arXiv - AI · 4 min ·
[2602.20204] Analyzing Latency Hiding and Parallelism in an MLIR-based AI Kernel Compiler
Machine Learning

[2602.20204] Analyzing Latency Hiding and Parallelism in an MLIR-based AI Kernel Compiler

This paper analyzes the effectiveness of latency hiding and parallelism techniques in an MLIR-based AI kernel compiler, focusing on vecto...

arXiv - AI · 3 min ·
[2602.20200] Global Prior Meets Local Consistency: Dual-Memory Augmented Vision-Language-Action Model for Efficient Robotic Manipulation
Machine Learning

[2602.20200] Global Prior Meets Local Consistency: Dual-Memory Augmented Vision-Language-Action Model for Efficient Robotic Manipulation

The paper presents OptimusVLA, a dual-memory framework for robotic manipulation that enhances efficiency and robustness in action generat...

arXiv - AI · 4 min ·
[2602.20196] OpenPort Protocol: A Security Governance Specification for AI Agent Tool Access
Ai Safety

[2602.20196] OpenPort Protocol: A Security Governance Specification for AI Agent Tool Access

The OpenPort Protocol introduces a governance-first approach for AI agents, ensuring secure access to application tools while addressing ...

arXiv - AI · 4 min ·
[2602.20191] MoBiQuant: Mixture-of-Bits Quantization for Token-Adaptive Elastic LLMs
Llms

[2602.20191] MoBiQuant: Mixture-of-Bits Quantization for Token-Adaptive Elastic LLMs

The paper presents MoBiQuant, a novel quantization framework for elastic large language models (LLMs) that adapts weight precision based ...

arXiv - Machine Learning · 4 min ·
[2602.20169] Autonomous AI and Ownership Rules
Robotics

[2602.20169] Autonomous AI and Ownership Rules

This article explores the ownership rules surrounding AI-generated outputs, examining how they are linked to their creators and the impli...

arXiv - AI · 3 min ·
[2601.12815] Multimodal Multi-Agent Empowered Legal Judgment Prediction
Ai Infrastructure

[2601.12815] Multimodal Multi-Agent Empowered Legal Judgment Prediction

This paper presents JurisMMA, a novel framework for Legal Judgment Prediction (LJP) that utilizes multimodal data to enhance the accuracy...

arXiv - AI · 4 min ·
[2602.21061] Tool Building as a Path to "Superintelligence"
Llms

[2602.21061] Tool Building as a Path to "Superintelligence"

The paper explores how Large Language Models (LLMs) can achieve superintelligence through the Diligent Learner framework, emphasizing the...

arXiv - AI · 3 min ·
Previous Page 86 Next

Related Topics

Stay updated with AI News

Get the latest news, tools, and insights delivered to your inbox.

Daily or weekly digest • Unsubscribe anytime