__Training__ code of pre-activation ResNet on ImageNet. It follows the setup in
[fb.resnet.torch](https://github.com/facebook/fb.resnet.torch)(except for the weight decay) and gets similar performance (with much fewer lines of code).
__Training__ code of ResNet on ImageNet, with pre-activation and squeeze-and-excitation.
The pre-act ResNet follows the setup in [fb.resnet.torch](https://github.com/facebook/fb.resnet.torch)(except for the weight decay)
and gets similar performance (with much fewer lines of code).
Models can be [downloaded here](https://goo.gl/6XjK9V).
The speed is 1310 image/s on 4 Tesla M40, if your data is fast enough.
You should be able to see good GPU utilization (around 95%), if your data is fast enough.
See the [tutorial](http://tensorpack.readthedocs.io/en/latest/tutorial/efficient-dataflow.html) on how to speed up your data.

...
...
@@ -46,9 +47,6 @@ The per-pixel mean used here is slightly different from the original.
Reproduce pre-activation ResNet on CIFAR10.
The train error shown here is a moving average of the error rate of each batch in training.
The validation error here is computed on test set.

Also see a [DenseNet implementation](https://github.com/YixuanLi/densenet-tensorflow) of the paper [Densely Connected Convolutional Networks](https://arxiv.org/abs/1608.06993).