[2309.10370] Geometric structure of shallow neural networks and constructive ${\mathcal L}^2$ cost minimization

[2309.10370] Geometric structure of shallow neural networks and constructive ${\mathcal L}^2$ cost minimization

arXiv - Machine Learning 4 min read

About this article

Abstract page for arXiv paper 2309.10370: Geometric structure of shallow neural networks and constructive ${\mathcal L}^2$ cost minimization

Computer Science > Machine Learning arXiv:2309.10370 (cs) [Submitted on 19 Sep 2023 (v1), last revised 26 Feb 2026 (this version, v4)] Title:Geometric structure of shallow neural networks and constructive ${\mathcal L}^2$ cost minimization Authors:Thomas Chen, Patrícia Muñoz Ewald View a PDF of the paper titled Geometric structure of shallow neural networks and constructive ${\mathcal L}^2$ cost minimization, by Thomas Chen and 1 other authors View PDF HTML (experimental) Abstract:In this paper, we approach the problem of cost (loss) minimization in underparametrized shallow ReLU networks through the explicit construction of upper bounds which appeal to the structure of classification data, without use of gradient descent. A key focus is on elucidating the geometric structure of approximate and precise minimizers. We consider an $L^2$ cost function, input space $\mathbb{R}^M$, output space ${\mathbb R}^Q$ with $Q\leq M$, and training input sample size that can be arbitrarily large. We prove an upper bound on the minimum of the cost function of order $O(\delta_P)$ where $\delta_P$ measures the signal-to-noise ratio of training data. In the special case $M=Q$, we explicitly determine an exact degenerate local minimum of the cost function, and show that the sharp value differs from the upper bound obtained for $Q\leq M$ by a relative error $O(\delta_P^2)$. The proof of the upper bound yields a constructively trained network; we show that it metrizes a particular $Q$-dimension...

Originally published on March 02, 2026. Curated by AI News.

Related Articles

Machine Learning

[R], 31 MILLIONS High frequency data, Light GBM worked perfectly

We just published a paper on predicting adverse selection in high-frequency crypto markets using LightGBM, and I wanted to share it here ...

Reddit - Machine Learning · 1 min ·
Machine Learning

[D] Those of you with 10+ years in ML — what is the public completely wrong about?

For those of you who've been in ML/AI research or applied ML for 10+ years — what's the gap between what the public thinks AI is doing vs...

Reddit - Machine Learning · 1 min ·
UMKC Announces New Master of Science in Artificial Intelligence
Ai Infrastructure

UMKC Announces New Master of Science in Artificial Intelligence

UMKC announces a new Master of Science in Artificial Intelligence program aimed at addressing workforce demand for AI expertise, set to l...

AI News - General · 4 min ·
Machine Learning

AI assistants are optimized to seem helpful. That is not the same thing as being helpful.

RLHF trains models on human feedback. Humans rate responses they like. And it turns out humans consistently rate confident, fluent, agree...

Reddit - Artificial Intelligence · 1 min ·
More in Machine Learning: This Week Guide Trending

No comments

No comments yet. Be the first to comment!

Stay updated with AI News

Get the latest news, tools, and insights delivered to your inbox.

Daily or weekly digest • Unsubscribe anytime