Closing the generalization gap in large batch training of neural networks
Train longer, generalize better – Big batch training This is a code repository used to generate the results appearing in “Train longer, generalize better: closing the generalization gap in large batch training of neural networks” By Elad Hoffer, Itay Hubara and Daniel Soudry. It is based off convNet.pytorch with some helpful options such as: Training on several datasets Complete logging of trained experiment Graph visualization of the training/validation loss and accuracy Definition of preprocessing and optimization regime for each model […]
Read more