Make dropout default to deterministic inference with option for stochastic
Currently the dropout layer applies dropout during training and inference, giving a stochastic result (like Gal et al.). It should instead default to a deterministic result (standard dropout).
Ideally this would be controllable from the configuration file.
Fixes to this should be applied to the activation layer and also to the dropout applied by channel sparse convolution.