[2512.04189] BEP: A Binary Error Propagation Algorithm for Binary Neural Networks Training
Summary
The paper presents BEP, a novel Binary Error Propagation algorithm for training Binary Neural Networks (BNNs) that enables efficient backpropagation using only binary operations.
Why It Matters
BEP addresses the challenges of training BNNs, which are crucial for resource-constrained environments. By allowing end-to-end binary training, it enhances the performance of BNNs, making them more viable for practical applications in AI and machine learning.
Key Takeaways
- BEP is the first algorithm to enable backpropagation in Binary Neural Networks using only binary operations.
- The algorithm improves test accuracy by up to +10.57% in recurrent neural networks.
- BEP facilitates end-to-end training for multi-layer architectures, enhancing the efficiency of BNNs.
- The approach eliminates the need for floating-point arithmetic during training, preserving computational efficiency.
- BEP is released as an open-source repository, promoting accessibility and further research.
Computer Science > Machine Learning arXiv:2512.04189 (cs) [Submitted on 3 Dec 2025 (v1), last revised 17 Feb 2026 (this version, v2)] Title:BEP: A Binary Error Propagation Algorithm for Binary Neural Networks Training Authors:Luca Colombo, Fabrizio Pittorino, Daniele Zambon, Carlo Baldassi, Manuel Roveri, Cesare Alippi View a PDF of the paper titled BEP: A Binary Error Propagation Algorithm for Binary Neural Networks Training, by Luca Colombo and 5 other authors View PDF HTML (experimental) Abstract:Binary Neural Networks (BNNs), which constrain both weights and activations to binary values, offer substantial reductions in computational complexity, memory footprint, and energy consumption. These advantages make them particularly well suited for deployment on resource-constrained devices. However, training BNNs via gradient-based optimization remains challenging due to the discrete nature of their variables. The dominant approach, quantization-aware training, circumvents this issue by employing surrogate gradients. Yet, this method requires maintaining latent full-precision parameters and performing the backward pass with floating-point arithmetic, thereby forfeiting the efficiency of binary operations during training. While alternative approaches based on local learning rules exist, they are unsuitable for global credit assignment and for back-propagating errors in multi-layer architectures. This paper introduces Binary Error Propagation (BEP), the first learning algorithm...