Natural Language Processing

Text understanding and language tasks

Top This Week

Llms

[R] Is autoresearch really better than classic hyperparameter tuning?

We did experiments comparing Optuna & autoresearch. Autoresearch converges faster, is more cost-efficient, and even generalizes bette...

Reddit - Machine Learning · 1 min ·
Nlp

Automate IOS devices through XCUITest with droidrun.

Automate iOS apps with XCUITest and Droidrun using just natural language. You send the command to Droidrun, and the agent starts the task...

Reddit - Artificial Intelligence · 1 min ·
Machine Learning

[P] Trained a small BERT on 276K Kubernetes YAMLs using tree positional encoding instead of sequential

I trained a BERT-style transformer on 276K Kubernetes YAML files, replacing standard positional encoding with learned tree coordinates (d...

Reddit - Machine Learning · 1 min ·

All Content

[2504.12007] Diffusion Generative Recommendation with Continuous Tokens
Llms

[2504.12007] Diffusion Generative Recommendation with Continuous Tokens

The paper presents ContRec, a novel framework that integrates continuous tokens into LLM-based recommender systems, enhancing user prefer...

arXiv - AI · 4 min ·
[2502.17364] Bridging Gaps in Natural Language Processing for Yorùbá: A Systematic Review of a Decade of Progress and Prospects
Nlp

[2502.17364] Bridging Gaps in Natural Language Processing for Yorùbá: A Systematic Review of a Decade of Progress and Prospects

This systematic review analyzes a decade of progress in Natural Language Processing (NLP) for the Yorùbá language, highlighting challenge...

arXiv - AI · 4 min ·
[2502.17028] Distributional Vision-Language Alignment by Cauchy-Schwarz Divergence
Nlp

[2502.17028] Distributional Vision-Language Alignment by Cauchy-Schwarz Divergence

The paper presents CS-Aligner, a novel framework for vision-language alignment that integrates Cauchy-Schwarz divergence with mutual info...

arXiv - Machine Learning · 4 min ·
[2405.10385] Augmenting Lateral Thinking in Language Models with Humor and Riddle Data for the BRAINTEASER Task
Llms

[2405.10385] Augmenting Lateral Thinking in Language Models with Humor and Riddle Data for the BRAINTEASER Task

This paper explores enhancing language models' lateral thinking abilities by integrating humor and riddle datasets for the BRAINTEASER ta...

arXiv - Machine Learning · 4 min ·
[2508.19113] Hybrid Deep Searcher: Scalable Parallel and Sequential Search Reasoning
Machine Learning

[2508.19113] Hybrid Deep Searcher: Scalable Parallel and Sequential Search Reasoning

The paper presents HybridDeepSearcher, a novel approach that enhances search reasoning by integrating parallel query expansion with struc...

arXiv - AI · 4 min ·
[2508.13404] TASER: Table Agents for Schema-guided Extraction and Recommendation
Nlp

[2508.13404] TASER: Table Agents for Schema-guided Extraction and Recommendation

The paper presents TASER, a system designed for schema-guided extraction and recommendation from complex financial tables, improving data...

arXiv - Machine Learning · 4 min ·
[2508.01012] AutoEDA: Enabling EDA Flow Automation through Microservice-Based LLM Agents
Llms

[2508.01012] AutoEDA: Enabling EDA Flow Automation through Microservice-Based LLM Agents

The article presents AutoEDA, a framework that utilizes microservice-based LLM agents to automate Electronic Design Automation (EDA) proc...

arXiv - AI · 4 min ·
[2506.04500] "Don't Do That!": Guiding Embodied Systems through Large Language Model-based Constraint Generation
Llms

[2506.04500] "Don't Do That!": Guiding Embodied Systems through Large Language Model-based Constraint Generation

This paper presents STPR, a framework that utilizes large language models to convert complex natural language constraints into executable...

arXiv - AI · 4 min ·
[2503.12434] A Survey on the Optimization of Large Language Model-based Agents
Llms

[2503.12434] A Survey on the Optimization of Large Language Model-based Agents

This survey reviews optimization techniques for Large Language Model (LLM)-based agents, categorizing methods into parameter-driven and p...

arXiv - AI · 4 min ·
[2602.21204] Test-Time Training with KV Binding Is Secretly Linear Attention
Machine Learning

[2602.21204] Test-Time Training with KV Binding Is Secretly Linear Attention

This paper explores the concept of Test-Time Training (TTT) with KV binding, revealing that it functions as learned linear attention rath...

arXiv - Machine Learning · 3 min ·
[2602.21189] Why Pass@k Optimization Can Degrade Pass@1: Prompt Interference in LLM Post-training
Llms

[2602.21189] Why Pass@k Optimization Can Degrade Pass@1: Prompt Interference in LLM Post-training

The paper explores the trade-off between Pass@k and Pass@1 performance metrics in large language models, revealing how optimizing for Pas...

arXiv - Machine Learning · 4 min ·
[2602.21165] PVminer: A Domain-Specific Tool to Detect the Patient Voice in Patient Generated Data
Machine Learning

[2602.21165] PVminer: A Domain-Specific Tool to Detect the Patient Voice in Patient Generated Data

PVminer is a novel NLP framework designed to detect the patient voice in patient-generated data, improving the analysis of patient-provid...

arXiv - AI · 4 min ·
[2602.21136] SparkMe: Adaptive Semi-Structured Interviewing for Qualitative Insight Discovery
Llms

[2602.21136] SparkMe: Adaptive Semi-Structured Interviewing for Qualitative Insight Discovery

The paper presents SparkMe, a multi-agent LLM system designed for adaptive semi-structured interviewing, enhancing qualitative data colle...

arXiv - AI · 4 min ·
[2602.21052] Position-Aware Sequential Attention for Accurate Next Item Recommendations
Machine Learning

[2602.21052] Position-Aware Sequential Attention for Accurate Next Item Recommendations

The paper presents a novel kernelized self-attention mechanism designed to enhance next-item recommendations by improving the representat...

arXiv - Machine Learning · 3 min ·
[2602.20979] Toward an Agentic Infused Software Ecosystem
Nlp

[2602.20979] Toward an Agentic Infused Software Ecosystem

This article discusses the concept of an Agentic Infused Software Ecosystem (AISE), emphasizing the need for a holistic approach to integ...

arXiv - AI · 3 min ·
[2602.20967] Training-Free Intelligibility-Guided Observation Addition for Noisy ASR
Machine Learning

[2602.20967] Training-Free Intelligibility-Guided Observation Addition for Noisy ASR

This paper presents a novel training-free method for improving automatic speech recognition (ASR) in noisy environments by using intellig...

arXiv - AI · 3 min ·
[2602.20877] E-MMKGR: A Unified Multimodal Knowledge Graph Framework for E-commerce Applications
Nlp

[2602.20877] E-MMKGR: A Unified Multimodal Knowledge Graph Framework for E-commerce Applications

The paper presents E-MMKGR, a unified framework for multimodal knowledge graphs tailored for e-commerce, enhancing recommendation systems...

arXiv - AI · 3 min ·
[2602.20735] RMIT-ADM+S at the MMU-RAG NeurIPS 2025 Competition
Llms

[2602.20735] RMIT-ADM+S at the MMU-RAG NeurIPS 2025 Competition

The paper presents RMIT-ADM+S, an award-winning system for the Text-to-Text track at the NeurIPS 2025 Competition, featuring a novel retr...

arXiv - AI · 3 min ·
[2602.20751] SibylSense: Adaptive Rubric Learning via Memory Tuning and Adversarial Probing
Machine Learning

[2602.20751] SibylSense: Adaptive Rubric Learning via Memory Tuning and Adversarial Probing

The paper presents SibylSense, a novel approach to adaptive rubric learning that enhances reward mechanisms in reinforcement learning thr...

arXiv - Machine Learning · 3 min ·
[2602.20731] Communication-Inspired Tokenization for Structured Image Representations
Machine Learning

[2602.20731] Communication-Inspired Tokenization for Structured Image Representations

The paper presents COMmunication inspired Tokenization (COMiT), a novel framework for structured image representations that enhances obje...

arXiv - Machine Learning · 4 min ·
Previous Page 80 Next

Related Topics

Stay updated with AI News

Get the latest news, tools, and insights delivered to your inbox.

Daily or weekly digest • Unsubscribe anytime