[2404.16890] Layer Collapse Can be Induced by Unstructured Pruning

[2404.16890] Layer Collapse Can be Induced by Unstructured Pruning

arXiv - AI 3 min read Article

Summary

This paper explores how unstructured pruning can lead to layer collapse in neural networks, demonstrating that it can effectively reduce model complexity without significant performance loss.

Why It Matters

Understanding the implications of unstructured pruning is crucial for optimizing neural network architectures. This research reveals that pruning can not only reduce parameters but also eliminate entire layers, enhancing model efficiency and performance in various applications.

Key Takeaways

  • Unstructured pruning can induce layer collapse in neural networks.
  • Magnitude-based pruning reduces neuron entropy, allowing for layer removal.
  • The method is validated across CNNs, Vision Transformers, and NLP models.
  • Effective layer removal can occur with minimal performance degradation.
  • This research contributes to more efficient model design in AI.

Computer Science > Machine Learning arXiv:2404.16890 (cs) [Submitted on 24 Apr 2024 (v1), last revised 23 Feb 2026 (this version, v2)] Title:Layer Collapse Can be Induced by Unstructured Pruning Authors:Zhu Liao, Victor Quétu, Van-Tam Nguyen, Enzo Tartaglione View a PDF of the paper titled Layer Collapse Can be Induced by Unstructured Pruning, by Zhu Liao and 3 other authors View PDF Abstract:Unstructured pruning is a popular compression method for efficiently reducing model parameters. However, while it effectively decreases the number of parameters, it is commonly believed that unstructured pruning cannot shorten the computational critical path, i.e., the maximum number of layers traversed during forward propagation. In this paper, we study when and how unstructured pruning can yield structural effects. For rectifier-activated networks, we introduce the notion of neuron entropy, which quantifies the degree of nonlinearity utilization. We show that magnitude-based pruning naturally lowers this entropy, sometimes down to zero-entropy layers that become linearizable and can thus be removed. Building on this insight, we propose a method that leverages "unstructured" pruning to favor sparsity in low-entropy layers, enabling their complete removal. We validate the phenomenon across CNNs, Vision Transformers, and NLP models: unstructured pruning can induce effective layer removal with little or no performance degradation in over-parameterized networks. Subjects: Machine Learnin...

Related Articles

Machine Learning

[D] ICML Rebuttal Question

I am currently working on my response on the rebuttal acknowledgments for ICML and I doubting how to handle the strawman argument of that...

Reddit - Machine Learning · 1 min ·
Machine Learning

[D] ML researcher looking to switch to a product company.

Hey, I am an AI researcher currently working in a deep tech company as a data scientist. Prior to this, I was doing my PhD. My current ro...

Reddit - Machine Learning · 1 min ·
Machine Learning

Building behavioural response models of public figures using Brain scan data (Predict their next move using psychological modelling) [P]

Hey guys, I’m the same creator of Netryx V2, the geolocation tool. I’ve been working on something new called COGNEX. It learns how a pers...

Reddit - Machine Learning · 1 min ·
Machine Learning

[P] bitnet-edge: Ternary-weight CNNs ({-1,0,+1}) on MNIST and CIFAR-10, deployed to ESP32-S3 with zero multiplications

I built a pipeline that takes ternary-quantized CNNs from PyTorch training all the way to bare-metal inference on an ESP32-S3 microcontro...

Reddit - Machine Learning · 1 min ·
More in Machine Learning: This Week Guide Trending

No comments

No comments yet. Be the first to comment!

Stay updated with AI News

Get the latest news, tools, and insights delivered to your inbox.

Daily or weekly digest • Unsubscribe anytime