Abstract
Large neural networks are very successful in various tasks. However, with limited data, the generalization capabilities of deep neural networks are also very limited. In this paper, we empirically start showing that intrinsically sparse neural networks with adaptive sparse connectivity, which by design have a strict parameter budget during the training phase, have better generalization capabilities than their fully-connected counterparts. Besides this, we propose a new technique to train these sparse models by combining the Sparse Evolutionary Training (SET) procedure with neurons pruning. Operated on MultiLayer Perceptron (MLP) and tested on 15 datasets, our proposed technique zeros out around 50% of the hidden neurons during training, while having a linear number of parameters to optimize with respect to the number of neurons. The results show a competitive classification and generalization performance.
Original language | English |
---|---|
Title of host publication | ICML 2019 Workshop on Understanding and Improving General-ization in Deep Learning |
Publication status | Published - 14 Jun 2019 |
Externally published | Yes |
Event | 36th International Conference on Machine Learning, ICML 2019 - Long Beach, United States Duration: 10 Jun 2019 → 15 Jun 2019 Conference number: 36 |
Conference
Conference | 36th International Conference on Machine Learning, ICML 2019 |
---|---|
Abbreviated title | ICML 2019 |
Country/Territory | United States |
City | Long Beach |
Period | 10/06/19 → 15/06/19 |