Natural Language Processing

Text understanding and language tasks

Top This Week

Nlp

[D] Is lossy compression acceptable for conversational agent memory? Every system today uses knowledge graph triples — here's why I think that's wrong.

Been thinking about this and want to know if others have hit the same issue. The dominant approach for agent memory (Mem0, Zep, most RAG ...

Reddit - Machine Learning · 1 min ·
[2601.11016] Contextual Distributionally Robust Optimization with Causal and Continuous Structure: An Interpretable and Tractable Approach
Nlp

[2601.11016] Contextual Distributionally Robust Optimization with Causal and Continuous Structure: An Interpretable and Tractable Approach

Abstract page for arXiv paper 2601.11016: Contextual Distributionally Robust Optimization with Causal and Continuous Structure: An Interp...

arXiv - Machine Learning · 4 min ·
[2511.22294] Structure is Supervision: Multiview Masked Autoencoders for Radiology
Machine Learning

[2511.22294] Structure is Supervision: Multiview Masked Autoencoders for Radiology

Abstract page for arXiv paper 2511.22294: Structure is Supervision: Multiview Masked Autoencoders for Radiology

arXiv - Machine Learning · 4 min ·

All Content

Llms

[R] Understanding targeted LLM fine-tuning

This article discusses a preprint on targeted instruction selection for fine-tuning large language models (LLMs), emphasizing systematic ...

Reddit - Machine Learning · 1 min ·
Machine Learning

[P] Whisper Accent — Accent-Aware English Speech Recognition

Whisper-Accent is a project aimed at enhancing Whisper's performance in recognizing accented English speech, providing tools for research...

Reddit - Machine Learning · 1 min ·
Machine Learning

Remote ML Intern – NLP, Diffusion Models, End-to-End Deployment (2nd Year AI/ML Student)

A second-year AI/ML student seeks a remote internship in machine learning, showcasing skills in NLP, diffusion models, and end-to-end pro...

Reddit - ML Jobs · 1 min ·
Anthropic Slams China for AI Theft, But Critics Say the Outrage Is Hypocritical
Nlp

Anthropic Slams China for AI Theft, But Critics Say the Outrage Is Hypocritical

Anthropic accuses Chinese developers of stealing AI secrets from its Claude chatbot, sparking criticism over its own data scraping practi...

AI Tools & Products · 7 min ·
[2602.07774] Generative Reasoning Re-ranker
Llms

[2602.07774] Generative Reasoning Re-ranker

The paper presents the Generative Reasoning Re-ranker (GR2), an innovative framework for enhancing recommendation systems using Large Lan...

arXiv - AI · 4 min ·
[2601.16449] Emotion-LLaMAv2 and MMEVerse: A New Framework and Benchmark for Multimodal Emotion Understanding
Llms

[2601.16449] Emotion-LLaMAv2 and MMEVerse: A New Framework and Benchmark for Multimodal Emotion Understanding

The paper introduces Emotion-LLaMAv2 and MMEVerse, a new framework and benchmark aimed at enhancing multimodal emotion understanding thro...

arXiv - AI · 4 min ·
[2601.04205] STaRR: Spatial-Temporal Token-Dynamics-Aware Responsive Remasking for Diffusion Language Models
Llms

[2601.04205] STaRR: Spatial-Temporal Token-Dynamics-Aware Responsive Remasking for Diffusion Language Models

The paper presents STaRR, a novel framework for responsive remasking in diffusion language models that adapts remasking decisions based o...

arXiv - AI · 3 min ·
[2601.00671] Fast-weight Product Key Memory
Llms

[2601.00671] Fast-weight Product Key Memory

The paper introduces Fast-weight Product Key Memory (FwPKM), a novel memory layer designed to enhance sequence modeling in language model...

arXiv - AI · 3 min ·
[2512.09730] Interpreto: An Explainability Library for Transformers
Llms

[2512.09730] Interpreto: An Explainability Library for Transformers

Interpreto is an open-source library designed for interpreting HuggingFace transformers, offering both attribution methods and concept-ba...

arXiv - Machine Learning · 3 min ·
[2511.20629] MapReduce LoRA: Advancing the Pareto Front in Multi-Preference Optimization for Generative Models
Machine Learning

[2511.20629] MapReduce LoRA: Advancing the Pareto Front in Multi-Preference Optimization for Generative Models

The paper presents MapReduce LoRA, a novel framework for optimizing generative models by addressing multi-preference alignment issues. It...

arXiv - Machine Learning · 4 min ·
[2510.06820] Efficient Discriminative Joint Encoders for Large Scale Vision-Language Reranking
Machine Learning

[2510.06820] Efficient Discriminative Joint Encoders for Large Scale Vision-Language Reranking

The paper presents EDJE, an Efficient Discriminative Joint Encoder designed to enhance vision-language reranking by precomputing visual t...

arXiv - Machine Learning · 3 min ·
[2510.27246] Beyond a Million Tokens: Benchmarking and Enhancing Long-Term Memory in LLMs
Llms

[2510.27246] Beyond a Million Tokens: Benchmarking and Enhancing Long-Term Memory in LLMs

This paper presents a new framework for evaluating and enhancing long-term memory in large language models (LLMs), introducing the BEAM b...

arXiv - AI · 4 min ·
[2509.26287] Flower: A Flow-Matching Solver for Inverse Problems
Machine Learning

[2509.26287] Flower: A Flow-Matching Solver for Inverse Problems

The paper introduces Flower, a novel solver for linear inverse problems that utilizes a pre-trained flow model to enhance reconstruction ...

arXiv - Machine Learning · 3 min ·
[2510.14979] From Pixels to Words -- Towards Native Vision-Language Primitives at Scale
Llms

[2510.14979] From Pixels to Words -- Towards Native Vision-Language Primitives at Scale

The paper discusses the development of native Vision-Language Models (VLMs) that integrate vision and language capabilities more effectiv...

arXiv - AI · 4 min ·
[2510.13632] Closing the Gap Between Text and Speech Understanding in LLMs
Llms

[2510.13632] Closing the Gap Between Text and Speech Understanding in LLMs

This paper addresses the performance gap between text and speech understanding in large language models (LLMs), proposing a new method, S...

arXiv - AI · 4 min ·
[2508.12674] Unfolded Laplacian Spectral Embedding: A Theoretically Grounded Approach to Dynamic Network Representation
Machine Learning

[2508.12674] Unfolded Laplacian Spectral Embedding: A Theoretically Grounded Approach to Dynamic Network Representation

The paper introduces Unfolded Laplacian Spectral Embedding (ULSE), a novel method for dynamic network representation that ensures stabili...

arXiv - Machine Learning · 3 min ·
[2510.05598] AgentDR: Dynamic Recommendation with Implicit Item-Item Relations via LLM-based Agents
Llms

[2510.05598] AgentDR: Dynamic Recommendation with Implicit Item-Item Relations via LLM-based Agents

The paper presents AgentDR, a novel framework that enhances recommendation systems by leveraging LLMs to understand implicit item relatio...

arXiv - AI · 4 min ·
[2506.16224] Malware Classification Leveraging NLP & Machine Learning for Enhanced Accuracy
Machine Learning

[2506.16224] Malware Classification Leveraging NLP & Machine Learning for Enhanced Accuracy

This article explores the use of NLP and machine learning techniques for enhancing malware classification accuracy, achieving a notable 9...

arXiv - Machine Learning · 3 min ·
[2506.01928] Esoteric Language Models: Bridging Autoregressive and Masked Diffusion LLMs
Llms

[2506.01928] Esoteric Language Models: Bridging Autoregressive and Masked Diffusion LLMs

The paper introduces Eso-LMs, a novel language model that integrates autoregressive and masked diffusion paradigms, enhancing inference e...

arXiv - Machine Learning · 4 min ·
[2505.22842] Bayesian Attention Mechanism: A Probabilistic Framework for Positional Encoding and Context Length Extrapolation
Llms

[2505.22842] Bayesian Attention Mechanism: A Probabilistic Framework for Positional Encoding and Context Length Extrapolation

The paper introduces the Bayesian Attention Mechanism (BAM), a novel framework for positional encoding in transformer models that enhance...

arXiv - Machine Learning · 3 min ·
Previous Page 84 Next

Related Topics

Stay updated with AI News

Get the latest news, tools, and insights delivered to your inbox.

Daily or weekly digest • Unsubscribe anytime