Natural Language Processing

Text understanding and language tasks

Top This Week

Llms

[R] Is autoresearch really better than classic hyperparameter tuning?

We did experiments comparing Optuna & autoresearch. Autoresearch converges faster, is more cost-efficient, and even generalizes bette...

Reddit - Machine Learning · 1 min ·
Nlp

Automate IOS devices through XCUITest with droidrun.

Automate iOS apps with XCUITest and Droidrun using just natural language. You send the command to Droidrun, and the agent starts the task...

Reddit - Artificial Intelligence · 1 min ·
Machine Learning

[P] Trained a small BERT on 276K Kubernetes YAMLs using tree positional encoding instead of sequential

I trained a BERT-style transformer on 276K Kubernetes YAML files, replacing standard positional encoding with learned tree coordinates (d...

Reddit - Machine Learning · 1 min ·

All Content

[2510.15425] TeamFormer: Shallow Parallel Transformers with Progressive Approximation
Machine Learning

[2510.15425] TeamFormer: Shallow Parallel Transformers with Progressive Approximation

The paper introduces TeamFormer, a shallow Transformer architecture that enhances parallelism and reduces training time while maintaining...

arXiv - Machine Learning · 4 min ·
[2509.19975] From Samples to Scenarios: A New Paradigm for Probabilistic Forecasting
Machine Learning

[2509.19975] From Samples to Scenarios: A New Paradigm for Probabilistic Forecasting

This article introduces a new paradigm for probabilistic forecasting, proposing 'Probabilistic Scenarios' as an alternative to traditiona...

arXiv - Machine Learning · 3 min ·
[2508.21785] Learning Unified Representations from Heterogeneous Data for Robust Heart Rate Modeling
Machine Learning

[2508.21785] Learning Unified Representations from Heterogeneous Data for Robust Heart Rate Modeling

This paper presents a novel framework for heart rate modeling that addresses data heterogeneity by learning unified representations from ...

arXiv - Machine Learning · 4 min ·
[2508.13904] One-Step Flow Q-Learning: Addressing the Diffusion Policy Bottleneck in Offline Reinforcement Learning
Machine Learning

[2508.13904] One-Step Flow Q-Learning: Addressing the Diffusion Policy Bottleneck in Offline Reinforcement Learning

The paper introduces One-Step Flow Q-Learning (OFQL), a novel framework that improves offline reinforcement learning by enabling one-step...

arXiv - Machine Learning · 4 min ·
[2505.18877] RefLoRA: Refactored Low-Rank Adaptation for Efficient Fine-Tuning of Large Models
Machine Learning

[2505.18877] RefLoRA: Refactored Low-Rank Adaptation for Efficient Fine-Tuning of Large Models

The paper presents RefLoRA, a novel method for fine-tuning large models by optimizing low-rank adaptations, leading to faster convergence...

arXiv - Machine Learning · 3 min ·
[2501.08219] Characterizing LLM Inference Energy-Performance Tradeoffs across Workloads and GPU Scaling
Llms

[2501.08219] Characterizing LLM Inference Energy-Performance Tradeoffs across Workloads and GPU Scaling

The paper explores the energy-performance tradeoffs in LLM inference across various workloads and GPU scaling, revealing significant insi...

arXiv - Machine Learning · 4 min ·
[2411.00759] Minibatch Optimal Transport and Perplexity Bound Estimation in Discrete Flow Matching
Machine Learning

[2411.00759] Minibatch Optimal Transport and Perplexity Bound Estimation in Discrete Flow Matching

This paper introduces a novel approach to discrete flow matching using minibatch optimal transport, enhancing generative performance whil...

arXiv - Machine Learning · 4 min ·
[2602.21142] LUMEN: Longitudinal Multi-Modal Radiology Model for Prognosis and Diagnosis
Llms

[2602.21142] LUMEN: Longitudinal Multi-Modal Radiology Model for Prognosis and Diagnosis

The LUMEN model enhances radiological diagnosis by leveraging longitudinal imaging data and multi-modal training, improving prognostic ca...

arXiv - Machine Learning · 4 min ·
[2602.20901] SpatiaLQA: A Benchmark for Evaluating Spatial Logical Reasoning in Vision-Language Models
Llms

[2602.20901] SpatiaLQA: A Benchmark for Evaluating Spatial Logical Reasoning in Vision-Language Models

The paper introduces SpatiaLQA, a benchmark for evaluating spatial logical reasoning in Vision-Language Models (VLMs), highlighting their...

arXiv - Machine Learning · 4 min ·
[2602.20816] Don't Ignore the Tail: Decoupling top-K Probabilities for Efficient Language Model Distillation
Llms

[2602.20816] Don't Ignore the Tail: Decoupling top-K Probabilities for Efficient Language Model Distillation

The paper presents a novel approach to language model distillation by introducing a tail-aware divergence that enhances the influence of ...

arXiv - Machine Learning · 3 min ·
[2602.20805] Assessing the Impact of Speaker Identity in Speech Spoofing Detection
Machine Learning

[2602.20805] Assessing the Impact of Speaker Identity in Speech Spoofing Detection

This paper investigates the influence of speaker identity on speech spoofing detection systems, proposing a framework that integrates spe...

arXiv - Machine Learning · 3 min ·
[2602.20528] Stop-Think-AutoRegress: Language Modeling with Latent Diffusion Planning
Llms

[2602.20528] Stop-Think-AutoRegress: Language Modeling with Latent Diffusion Planning

The paper presents STAR-LDM, a novel language model that integrates latent diffusion planning with autoregressive generation, enhancing n...

arXiv - Machine Learning · 3 min ·
[2602.20198] KEMP-PIP: A Feature-Fusion Based Approach for Pro-inflammatory Peptide Prediction
Llms

[2602.20198] KEMP-PIP: A Feature-Fusion Based Approach for Pro-inflammatory Peptide Prediction

The article presents KEMP-PIP, a novel hybrid machine learning framework designed for predicting pro-inflammatory peptides by integrating...

arXiv - Machine Learning · 3 min ·
[2602.20164] Benchmarking Distilled Language Models: Performance and Efficiency in Resource-Constrained Settings
Llms

[2602.20164] Benchmarking Distilled Language Models: Performance and Efficiency in Resource-Constrained Settings

This paper benchmarks distilled language models, demonstrating their superior performance and efficiency in resource-constrained environm...

arXiv - Machine Learning · 3 min ·
[2602.20165] VISION-ICE: Video-based Interpretation and Spatial Identification of Arrhythmia Origins via Neural Networks in Intracardiac Echocardiography
Machine Learning

[2602.20165] VISION-ICE: Video-based Interpretation and Spatial Identification of Arrhythmia Origins via Neural Networks in Intracardiac Echocardiography

The paper presents VISION-ICE, an AI framework utilizing intracardiac echocardiography to identify arrhythmia origins, achieving 66.2% ac...

arXiv - Machine Learning · 4 min ·
[2602.21196] Untied Ulysses: Memory-Efficient Context Parallelism via Headwise Chunking
Machine Learning

[2602.21196] Untied Ulysses: Memory-Efficient Context Parallelism via Headwise Chunking

The paper presents UPipe, a novel technique for memory-efficient context parallelism in Transformer models, achieving significant reducti...

arXiv - Machine Learning · 4 min ·
[2602.21133] SOM-VQ: Topology-Aware Tokenization for Interactive Generative Models
Machine Learning

[2602.21133] SOM-VQ: Topology-Aware Tokenization for Interactive Generative Models

The paper presents SOM-VQ, a novel tokenization method that enhances interactive generative models by integrating vector quantization wit...

arXiv - Machine Learning · 3 min ·
[2602.21078] ProxyFL: A Proxy-Guided Framework for Federated Semi-Supervised Learning
Machine Learning

[2602.21078] ProxyFL: A Proxy-Guided Framework for Federated Semi-Supervised Learning

The article presents ProxyFL, a novel framework for Federated Semi-Supervised Learning (FSSL) that addresses data heterogeneity issues by...

arXiv - Machine Learning · 4 min ·
[2602.21081] Scaling Vision Transformers: Evaluating DeepSpeed for Image-Centric Workloads
Llms

[2602.21081] Scaling Vision Transformers: Evaluating DeepSpeed for Image-Centric Workloads

This article evaluates the use of DeepSpeed to enhance the scalability of Vision Transformers (ViTs) for image-centric workloads, focusin...

arXiv - Machine Learning · 3 min ·
[2602.20932] Hierarchic-EEG2Text: Assessing EEG-To-Text Decoding across Hierarchical Abstraction Levels
Nlp

[2602.20932] Hierarchic-EEG2Text: Assessing EEG-To-Text Decoding across Hierarchical Abstraction Levels

This article presents a novel approach to EEG-to-text decoding, exploring how hierarchical abstraction levels affect classification perfo...

arXiv - Machine Learning · 4 min ·
Previous Page 78 Next

Related Topics

Stay updated with AI News

Get the latest news, tools, and insights delivered to your inbox.

Daily or weekly digest • Unsubscribe anytime