[2602.16316] A Graph Meta-Network for Learning on Kolmogorov-Arnold Networks

[2602.16316] A Graph Meta-Network for Learning on Kolmogorov-Arnold Networks

arXiv - Machine Learning 4 min read Article

Summary

This paper introduces WS-KAN, a novel weight-space architecture for Kolmogorov-Arnold Networks (KANs), demonstrating its superior performance over traditional methods in various tasks.

Why It Matters

The development of WS-KAN addresses a significant gap in the design of architectures for KANs, which have unique permutation symmetries. This advancement can enhance machine learning applications, making it crucial for researchers and practitioners in the field.

Key Takeaways

  • WS-KAN is the first architecture specifically designed for Kolmogorov-Arnold Networks.
  • It leverages the permutation symmetries shared with MLPs to improve learning efficiency.
  • Empirical evaluations show WS-KAN outperforms structure-agnostic baselines across diverse tasks.
  • A comprehensive zoo of trained KANs serves as benchmarks for assessing WS-KAN's performance.
  • The code for WS-KAN is publicly available, promoting further research and development.

Computer Science > Machine Learning arXiv:2602.16316 (cs) [Submitted on 18 Feb 2026] Title:A Graph Meta-Network for Learning on Kolmogorov-Arnold Networks Authors:Guy Bar-Shalom, Ami Tavory, Itay Evron, Maya Bechler-Speicher, Ido Guy, Haggai Maron View a PDF of the paper titled A Graph Meta-Network for Learning on Kolmogorov-Arnold Networks, by Guy Bar-Shalom and 5 other authors View PDF HTML (experimental) Abstract:Weight-space models learn directly from the parameters of neural networks, enabling tasks such as predicting their accuracy on new datasets. Naive methods -- like applying MLPs to flattened parameters -- perform poorly, making the design of better weight-space architectures a central challenge. While prior work leveraged permutation symmetries in standard networks to guide such designs, no analogous analysis or tailored architecture yet exists for Kolmogorov-Arnold Networks (KANs). In this work, we show that KANs share the same permutation symmetries as MLPs, and propose the KAN-graph, a graph representation of their computation. Building on this, we develop WS-KAN, the first weight-space architecture that learns on KANs, which naturally accounts for their symmetry. We analyze WS-KAN's expressive power, showing it can replicate an input KAN's forward pass - a standard approach for assessing expressiveness in weight-space architectures. We construct a comprehensive ``zoo'' of trained KANs spanning diverse tasks, which we use as benchmarks to empirically evaluate...

Related Articles

Machine Learning

[P] Fused MoE Dispatch in Pure Triton: Beating CUDA-Optimized Megablocks at Inference Batch Sizes

I built a fused MoE dispatch kernel in pure Triton that handles the full forward pass for Mixture-of-Experts models. No CUDA, no vendor-s...

Reddit - Machine Learning · 1 min ·
Machine Learning

[D] ICML Rebuttal Question

I am currently working on my response on the rebuttal acknowledgments for ICML and I doubting how to handle the strawman argument of that...

Reddit - Machine Learning · 1 min ·
Machine Learning

[D] ML researcher looking to switch to a product company.

Hey, I am an AI researcher currently working in a deep tech company as a data scientist. Prior to this, I was doing my PhD. My current ro...

Reddit - Machine Learning · 1 min ·
Machine Learning

Building behavioural response models of public figures using Brain scan data (Predict their next move using psychological modelling) [P]

Hey guys, I’m the same creator of Netryx V2, the geolocation tool. I’ve been working on something new called COGNEX. It learns how a pers...

Reddit - Machine Learning · 1 min ·
More in Machine Learning: This Week Guide Trending

No comments

No comments yet. Be the first to comment!

Stay updated with AI News

Get the latest news, tools, and insights delivered to your inbox.

Daily or weekly digest • Unsubscribe anytime