[2602.22107] Don't stop me now: Rethinking Validation Criteria for Model Parameter Selection

[2602.22107] Don't stop me now: Rethinking Validation Criteria for Model Parameter Selection

arXiv - Machine Learning 4 min read Article

Summary

This paper examines how different validation criteria for model parameter selection impact test performance in neural classifiers, revealing that loss-based criteria outperform accuracy-based methods.

Why It Matters

Understanding the effectiveness of validation criteria is crucial for improving model performance in machine learning. This study challenges traditional reliance on validation accuracy, suggesting that loss-based metrics may yield better outcomes, which can significantly influence model training practices in the field.

Key Takeaways

  • Early stopping based on validation accuracy often results in lower test accuracy.
  • Loss-based validation criteria provide more stable and comparable test accuracy.
  • Single validation rules frequently underperform compared to the best performance across all epochs.

Computer Science > Machine Learning arXiv:2602.22107 (cs) [Submitted on 25 Feb 2026] Title:Don't stop me now: Rethinking Validation Criteria for Model Parameter Selection Authors:Andrea Apicella, Francesco Isgrò, Andrea Pollastro, Roberto Prevete View a PDF of the paper titled Don't stop me now: Rethinking Validation Criteria for Model Parameter Selection, by Andrea Apicella and 3 other authors View PDF Abstract:Despite the extensive literature on training loss functions, the evaluation of generalization on the validation set remains underexplored. In this work, we conduct a systematic empirical and statistical study of how the validation criterion used for model selection affects test performance in neural classifiers, with attention to early stopping. Using fully connected networks on standard benchmarks under $k$-fold evaluation, we compare: (i) early stopping with patience and (ii) post-hoc selection over all epochs (i.e. no early stopping). Models are trained with cross-entropy, C-Loss, or PolyLoss; the model parameter selection on the validation set is made using accuracy or one of the three loss functions, each considered independently. Three main findings emerge. (1) Early stopping based on validation accuracy performs worst, consistently selecting checkpoints with lower test accuracy than both loss-based early stopping and post-hoc selection. (2) Loss-based validation criteria yield comparable and more stable test accuracy. (3) Across datasets and folds, any singl...

Related Articles

Llms

Is the Mirage Effect a bug, or is it Geometric Reconstruction in action? A framework for why VLMs perform better "hallucinating" than guessing, and what that may tell us about what's really inside these models

Last week, a team from Stanford and UCSF (Asadi, O'Sullivan, Fei-Fei Li, Euan Ashley et al.) dropped two companion papers. The first, MAR...

Reddit - Artificial Intelligence · 1 min ·
Yupp shuts down after raising $33M from a16z crypto's Chris Dixon | TechCrunch
Machine Learning

Yupp shuts down after raising $33M from a16z crypto's Chris Dixon | TechCrunch

Less than a year after launching, with checks from some of the biggest names in Silicon Valley, crowdsourced AI model feedback startup Yu...

TechCrunch - AI · 4 min ·
Machine Learning

[R] Fine-tuning services report

If you have some data and want to train or run a small custom model but don't have powerful enough hardware for training, fine-tuning ser...

Reddit - Machine Learning · 1 min ·
Machine Learning

[D] Does ML have a "bible"/reference textbook at the Intermediate/Advanced level?

Hello, everyone! This is my first time posting here and I apologise if the question is, perhaps, a bit too basic for this sub-reddit. A b...

Reddit - Machine Learning · 1 min ·
More in Machine Learning: This Week Guide Trending

No comments

No comments yet. Be the first to comment!

Stay updated with AI News

Get the latest news, tools, and insights delivered to your inbox.

Daily or weekly digest • Unsubscribe anytime