Some Experiments with Neural Networks and Deep Learning.
Based on Neural Networks and Deep Learning by Michael Nielsen.
- armadillo
- LAPACK and BLAS have to be installed on your system.
On Debian based systems you can use
sudo apt-get install libblas-dev liblapack-dev
to install them.
- use simpler version of problem and network first -> speed up
- find threshold of immediate (in first few epochs) decrease, no rise or oscillation of training cost
- order of magnitude: start with 0.01, multiply or divide by 10
- fine tune: compare multiple values directly above found magnitude
- iterate a few times
- use early stopping or half threshold as constant value
- validation accuracy
- best improvement
- start with 0 and get good eta
- start at 1
- order of magnitude as with eta
- fine tune as with eta
- re-optimize eta
- bouncing with eta
- same as lambda
- but order of magnitude already known, [0; 1]
- bounce with eta and lambda
- roughly optimize other hyper-parameters
- plot validation accuracy versus time -> try many different
- scale eta anti-proportional to mini-batch size
- proceed by optimizing other hyper-parameters