Abstract
The early layers of a deep neural net have the fewest parameters, but take up the
most computation. In this extended abstract, we propose to only train the hidden layers for a set portion of the training run, freezing them out one-by-one and excluding them from the backward pass. We empirically demonstrate that FreezeOut yields savings of up to 20% wall-clock time during training with 3% loss in accuracy for DenseNets on CIFAR.
most computation. In this extended abstract, we propose to only train the hidden layers for a set portion of the training run, freezing them out one-by-one and excluding them from the backward pass. We empirically demonstrate that FreezeOut yields savings of up to 20% wall-clock time during training with 3% loss in accuracy for DenseNets on CIFAR.
| Original language | English |
|---|---|
| Number of pages | 7 |
| Publication status | Published - 8 Dec 2017 |
| Event | NIPS 2017 Workshop on Optimization: 10th NIPS Workshop on Optimization for Machine Learning - Long Beach, Long Beach, United States Duration: 8 Dec 2017 → … Conference number: 10 |
Workshop
| Workshop | NIPS 2017 Workshop on Optimization |
|---|---|
| Abbreviated title | NIPS |
| Country/Territory | United States |
| City | Long Beach |
| Period | 8/12/17 → … |
Keywords
- Machine learning
- Neural network
ASJC Scopus subject areas
- Artificial Intelligence
Fingerprint
Dive into the research topics of 'FreezeOut: Accelerate Training by Progressively Freezing Layers'. Together they form a unique fingerprint.Cite this
- APA
- Author
- BIBTEX
- Harvard
- Standard
- RIS
- Vancouver