Abstract:
Looking for sparsity is nowadays crucial to speed up the training of large-scale neural networks. Projections onto the \ell_{1} and \ell_{1,\propto} are among the mos...Show MoreMetadata
Abstract:
Looking for sparsity is nowadays crucial to speed up the training of large-scale neural networks. Projections onto the \ell_{1} and \ell_{1,\propto} are among the most efficient techniques to sparsify and reduce the overall cost of neural networks. In this paper, we introduce a new projection algorithm for the \ell_{1,\infty} norm ball. Its worst-case time complexity is \mathcal{O}(nm+J\log(nm)) for a matrix in \mathbb{R}^{n\times m}. J is a term that tends to 0 when the sparsity is high, and to n\times m in the worst case. The algorithm is easy to implement and it is guaranteed to converge to the exact solution in finite time. Moreover, we propose to incorporate the \ell_{1,\infty} ball projection while training an auto encoder to enforce feature selection and sparsity of the weights. Sparsification appears in the encoder to primarily do feature selection due to our application in biology, where only a very small part (< 2%) of the data is relevant. We show that in both the biological and general cases of sparsity, our method is the fastest.
Date of Conference: 28-30 October 2024
Date Added to IEEE Xplore: 28 January 2025
ISBN Information: