Commit b17c8e73 authored by Yuxin Wu's avatar Yuxin Wu

update README

parent 7d40e049
......@@ -37,11 +37,11 @@ It's Yet Another TF wrapper, but different in:
2. Focus on __training speed__.
+ Speed comes for free with tensorpack -- it uses TensorFlow in the __correct way__.
Even on a tiny CNN example, the training runs [1.6x faster](https://gist.github.com/ppwwyyxx/8d95da79f8d97036a7d67c2416c851b6) than the equivalent Keras code.
On various CNNs, it runs 1.5~1.7x faster than the equivalent Keras code.
+ Data-parallel multi-GPU training is off-the-shelf to use. It is as fast as Google's [official benchmark](https://www.tensorflow.org/performance/benchmarks).
+ Data-parallel multi-GPU/distributed training is off-the-shelf to use. It is as fast as Google's [official benchmark](https://www.tensorflow.org/performance/benchmarks).
+ Data-parallel distributed training is off-the-shelf to use. It is as slow as Google's official benchmark.
+ See [tensorpack/benchmarks](https://github.com/tensorpack/benchmarks) for some benchmark scripts.
3. Focus on __large datasets__.
+ It's painful to read/preprocess data through TF.
......@@ -69,4 +69,4 @@ Dependencies:
pip install -U git+https://github.com/ppwwyyxx/tensorpack.git
# or add `--user` to avoid system-wide installation.
```
Besides, if you only want to use `tensorpack.dataflow` alone as a data processing library, TensorFlow is also optional.
If you only want to use `tensorpack.dataflow` alone as a data processing library, TensorFlow is also optional.
......@@ -18,6 +18,7 @@ Training examples with __reproducible__ and meaningful performance.
+ [Spatial Transformer Networks on MNIST addition](SpatialTransformer)
+ [Visualize CNN saliency maps](Saliency)
+ [Similarity learning on MNIST](SimilarityLearning)
+ Learn steering filters with [Dynamic Filter Networks](DynamicFilterNetwork)
+ Load a pre-trained [AlexNet](load-alexnet.py) or [VGG16](load-vgg16.py) model.
+ Load a pre-trained [Convolutional Pose Machines](ConvolutionalPoseMachines/).
......
Markdown is supported
0% or
You are about to add 0 people to the discussion. Proceed with caution.
Finish editing this message first!
Please register or to comment