Abstract
The early layers of a deep neural net have the fewest parameters, but take up the
most computation. In this extended abstract, we propose to only train the hidden layers for a set portion of the training run, freezing them out one-by-one and excluding them from the backward pass. We empirically demonstrate that FreezeOut yields savings of up to 20% wall-clock time during training with 3% loss in accuracy for DenseNets on CIFAR.
most computation. In this extended abstract, we propose to only train the hidden layers for a set portion of the training run, freezing them out one-by-one and excluding them from the backward pass. We empirically demonstrate that FreezeOut yields savings of up to 20% wall-clock time during training with 3% loss in accuracy for DenseNets on CIFAR.
Original language | English |
---|---|
Number of pages | 7 |
Publication status | Published - 8 Dec 2017 |
Event | NIPS 2017 Workshop on Optimization: 10th NIPS Workshop on Optimization for Machine Learning - Long Beach, Long Beach, United States Duration: 8 Dec 2017 → … Conference number: 10 |
Workshop
Workshop | NIPS 2017 Workshop on Optimization |
---|---|
Abbreviated title | NIPS |
Country/Territory | United States |
City | Long Beach |
Period | 8/12/17 → … |
Keywords
- Machine learning
- Neural network
ASJC Scopus subject areas
- Artificial Intelligence