We present a new method for structured pruning of neural networks, based on the recently proposed neuron merging trick in which following a pruning operation, the weights of the next layer are suitably modified. By a rigorous mathematical analysis of the neuron merging technique we prove an upper bound on the reconstruction error. This bound defines a new objective function for pruning-and-merging.
Articles
Related Articles
January 25, 2024
Rotation Invariant Quantization for Model Compression
Post-training Neural Network (NN) model compression is an attractive approach for deploying large, memory-consuming models on...
Read More >
1 MIN READING
July 7, 2023
Near-Memory Processing Offload to Remote (Persistent) Memory.
Read More >
0 MIN READING
November 5, 2015
Hardware Transactions in Nonvolatile Memory
Hardware transactional memory (HTM) implementations already provide a transactional abstraction at HW speed in multi-core systems....
Read More >
1 MIN READING