Deep learning guide line¶
If the input vector is fixed use an fully connected NN
If we have a topological structure than CNN
If there is a sequence than a gated recurrent net
Batch normalization can have a drastic effect on optimization especially for CNN and networks with sigmoidal nonlinearities.
We can ommit dropout if we use batch normalization
Enough data¶
If we the training error is high than there is no reason to collect more data since we should improve the model.
We should gather more data only when the validation set is much worse than the test set. And we already regularizing have dropout and have tuned hyper parameters.