[2604.01484] The topological gap at criticality: scaling exponent d + η, universality, and scope

[2604.01484] The topological gap at criticality: scaling exponent d + η, universality, and scope

arXiv - Machine Learning 4 min read

About this article

Abstract page for arXiv paper 2604.01484: The topological gap at criticality: scaling exponent d + η, universality, and scope

Condensed Matter > Statistical Mechanics arXiv:2604.01484 (cond-mat) [Submitted on 1 Apr 2026] Title:The topological gap at criticality: scaling exponent d + η, universality, and scope Authors:Matthew Loftus View a PDF of the paper titled The topological gap at criticality: scaling exponent d + {\eta}, universality, and scope, by Matthew Loftus View PDF HTML (experimental) Abstract:The topological gap $\Delta = TP_{H_1}^{real} - TP_{H_1}^{shuf}$ -- the excess $H_1$ total persistence of the majority-spin alpha complex over a density-matched null -- encodes critical correlations in spin models. We establish finite-size scaling: $\Delta(L,T) = A L^{d+\eta} G_-(L|t/T_c|)$, with $G_-(x) \sim (1+x/x_0)^{-(1+\beta/\nu)}$. For 2D Ising, $\alpha = 2.249 \pm 0.038$, matching $d+\eta = 9/4$ to $0.03\sigma$; the $G_-$ exponent $\gamma = 1.089 \pm 0.077$ is consistent with $1+\beta/\nu = 9/8$ ($\Delta R^2 < 10^{-5}$). For 2D Potts $q=3$ with $L$ up to 1024, $\alpha = 2.272 \pm 0.024$ ($0.2\sigma$ from $d+\eta = 2.267$), with two-term corrections to scaling ($R^2 = 0.9999$). The $G_-$ exponent $\gamma = 1.114$ (68% CI $[1.053, 1.173]$) matches $1+\beta/\nu = 17/15$. Scope boundaries: the law fails for 2D Potts $q=4$ ($\alpha = 2.347 \pm 0.017$, $9.3\sigma$ from $d+\eta = 5/2$) where logarithmic corrections prevent convergence, and for raw 3D Ising ($4\sigma$ from $d+\eta$), but density normalization $\Delta/|M|^{1/2}$ recovers $\alpha = 3.06 \pm 0.04$ ($0.6\sigma$). The framework fails ...

Originally published on April 03, 2026. Curated by AI News.

Related Articles

Machine Learning

HydraLM: 22× faster decoding and 16× smaller state memory in long-context inference experiments [P]

I’ve been experimenting with HydraLM, a long-context model for inference, and the numbers are getting a bit wild: the repo’s benchmark su...

Reddit - Machine Learning · 1 min ·
Machine Learning

How to know if a research-oriented role is for you? [D]

I’m currently a first-year Master’s student in Data Science &amp; AI, and I’m trying to figure out whether a research-oriented career is ...

Reddit - Machine Learning · 1 min ·
Machine Learning

GPU Compass – open-source, real-time GPU pricing across 20+ clouds [P]

We maintain an open-source catalog of cloud GPU offerings (skypilot-catalog, Apache 2.0). It auto-fetches pricing from 20+ cloud APIs eve...

Reddit - Machine Learning · 1 min ·
5 AI Models Tried to Scam Me. Some of Them Were Scary Good | WIRED
Machine Learning

5 AI Models Tried to Scam Me. Some of Them Were Scary Good | WIRED

The cyber capabilities of AI models have experts rattled. AI’s social skills may be just as dangerous.

Wired - AI · 8 min ·
More in Machine Learning: This Week Guide Trending

No comments

No comments yet. Be the first to comment!

Stay updated with AI News

Get the latest news, tools, and insights delivered to your inbox.

Daily or weekly digest • Unsubscribe anytime