作者: Max Welling , Karen Ullrich , Christos Louizos
DOI:
关键词:
摘要: Compression and computational efficiency in deep learning have become a problem of great significance. In this work, we argue that the most principled effective way to attack is by adopting Bayesian point view, where through sparsity inducing priors prune large parts network. We introduce two novelties paper: 1) use hierarchical nodes instead individual weights, 2) posterior uncertainties determine optimal fixed precision encode weights. Both factors significantly contribute achieving state art terms compression rates, while still staying competitive with methods designed optimize for speed or energy efficiency.