@article{doi:10.1137/20M1383239, author = {Baykal, Cenk and Liebenwein, Lucas and Gilitschenski, Igor and Feldman, Dan and Rus, Daniela}, title = {Sensitivity-Informed Provable Pruning of Neural Networks}, journal = {SIAM Journal on Mathematics of Data Science}, volume = {4}, number = {1}, pages = {26-45}, year = {2022}, doi = {10.1137/20M1383239}, URL = { https://doi.org/10.1137/20M1383239 }, eprint = { https://doi.org/10.1137/20M1383239 } , abstract = { We introduce a family of pruning algorithms that sparsifies the parameters of a trained model in a way that approximately preserves the model's predictive accuracy. Our algorithms use a small batch of input points to construct a data-informed importance sampling distribution over the network's parameters and use either a sampling-based or a deterministic pruning procedure, or an adaptive mixture of both, to discard redundant weights. Our methods are simultaneously computationally efficient, provably accurate, and broadly applicable to various network architectures and data distributions. The presented approaches are simple to implement and can be easily integrated into standard prune-retrain pipelines. We present empirical comparisons showing that our algorithms reliably generate highly compressed networks that incur minimal loss in performance, regardless of whether the original network is fully trained or randomly initialized. } }