Commit acae3fe5 authored by Yuxin Wu's avatar Yuxin Wu

add missing files

parent 40711389
...@@ -32,12 +32,12 @@ matrix: ...@@ -32,12 +32,12 @@ matrix:
env: TF_VERSION=1.6.0 TF_TYPE=release PYPI=true env: TF_VERSION=1.6.0 TF_TYPE=release PYPI=true
- os: linux - os: linux
python: 2.7 python: 2.7
env: TF_VERSION=1.head TF_TYPE=nightly env: TF_TYPE=nightly
- os: linux - os: linux
python: 3.5 python: 3.5
env: TF_VERSION=1.head TF_TYPE=nightly env: TF_TYPE=nightly
allow_failures: allow_failures:
- env: TF_VERSION=1.head TF_TYPE=nightly - env: TF_TYPE=nightly
install: install:
- pip install -U pip # the pip version on travis is too old - pip install -U pip # the pip version on travis is too old
......
...@@ -37,7 +37,7 @@ Instead of showing you 10 random networks with random accuracy, ...@@ -37,7 +37,7 @@ Instead of showing you 10 random networks with random accuracy,
And everything runs on multiple GPUs. Some highlights: And everything runs on multiple GPUs. Some highlights:
### Vision: ### Vision:
+ [Train ResNet on ImageNet](examples/ResNet) + [Train ResNet](examples/ResNet) and [other models](examples/ImageNetModels) on ImageNet.
+ [Train Faster-RCNN / Mask-RCNN on COCO object detection](examples/FasterRCNN) + [Train Faster-RCNN / Mask-RCNN on COCO object detection](examples/FasterRCNN)
+ [Generative Adversarial Network(GAN) variants](examples/GAN), including DCGAN, InfoGAN, Conditional GAN, WGAN, BEGAN, DiscoGAN, Image to Image, CycleGAN. + [Generative Adversarial Network(GAN) variants](examples/GAN), including DCGAN, InfoGAN, Conditional GAN, WGAN, BEGAN, DiscoGAN, Image to Image, CycleGAN.
+ [DoReFa-Net: train binary / low-bitwidth CNN on ImageNet](examples/DoReFa-Net) + [DoReFa-Net: train binary / low-bitwidth CNN on ImageNet](examples/DoReFa-Net)
......
...@@ -4,23 +4,22 @@ ...@@ -4,23 +4,22 @@
Training examples with __reproducible performance__. Training examples with __reproducible performance__.
__The word "reproduce" should always means reproduce performance__. __The word "reproduce" should always means reproduce performance__.
Reproducing a method is usually easy, but you don't know whether you've made mistakes, because wrong code will often appear to work. With the magic of SGD, wrong code often appears to still work, unless you check its performance number.
Reproducing __performance__ results is what really matters, and is something that's hardly seen on github.
See [Unawareness of Deep Learning Mistakes](https://medium.com/@ppwwyyxx/unawareness-of-deep-learning-mistakes-d5b5774da0ba). See [Unawareness of Deep Learning Mistakes](https://medium.com/@ppwwyyxx/unawareness-of-deep-learning-mistakes-d5b5774da0ba).
## Getting Started: ## Getting Started:
These examples don't have meaningful performance numbers. They are supposed to be just demos. These examples don't have meaningful performance numbers. They are supposed to be just demos.
+ [An illustrative MNIST example with explanation of the framework](basics/mnist-convnet.py) + [An illustrative MNIST example with explanation of the framework](basics/mnist-convnet.py)
+ A tiny [Cifar ConvNet](basics/cifar-convnet.py) and [SVHN ConvNet](basics/svhn-digit-convnet.py)
+ Tensorpack supports any symbolic libraries. See the same MNIST example written with [tf.layers](basics/mnist-tflayers.py), [tf-slim](basics/mnist-tfslim.py), and [with weights visualizations](basics/mnist-visualizations.py) + Tensorpack supports any symbolic libraries. See the same MNIST example written with [tf.layers](basics/mnist-tflayers.py), [tf-slim](basics/mnist-tfslim.py), and [with weights visualizations](basics/mnist-visualizations.py)
+ A tiny [Cifar ConvNet](basics/cifar-convnet.py) and [SVHN ConvNet](basics/svhn-digit-convnet.py)
+ If you've used Keras, check out [Keras examples](keras) + If you've used Keras, check out [Keras examples](keras)
+ [A boilerplate file to start with, for your own tasks](boilerplate.py) + [A boilerplate file to start with, for your own tasks](boilerplate.py)
## Vision: ## Vision:
| Name | Performance | | Name | Performance |
| --- | --- | | --- | --- |
| Train [ResNet](ResNet) and [ShuffleNet](ImageNetModels) on ImageNet | reproduce paper | | Train [ResNet](ResNet), [ShuffleNet and other models](ImageNetModels) on ImageNet | reproduce paper |
| [Train Faster-RCNN / Mask-RCNN on COCO](FasterRCNN) | reproduce paper | | [Train Faster-RCNN / Mask-RCNN on COCO](FasterRCNN) | reproduce paper |
| [DoReFa-Net: training binary / low-bitwidth CNN on ImageNet](DoReFa-Net) | reproduce paper | | [DoReFa-Net: training binary / low-bitwidth CNN on ImageNet](DoReFa-Net) | reproduce paper |
| [Generative Adversarial Network(GAN) variants](GAN), including DCGAN, InfoGAN, <br/> Conditional GAN, WGAN, BEGAN, DiscoGAN, Image to Image, CycleGAN | visually reproduce | | [Generative Adversarial Network(GAN) variants](GAN), including DCGAN, InfoGAN, <br/> Conditional GAN, WGAN, BEGAN, DiscoGAN, Image to Image, CycleGAN | visually reproduce |
......
...@@ -15,16 +15,7 @@ if [ $TF_TYPE == "release" ]; then ...@@ -15,16 +15,7 @@ if [ $TF_TYPE == "release" ]; then
fi fi
fi fi
if [ $TF_TYPE == "nightly" ]; then if [ $TF_TYPE == "nightly" ]; then
if [[ $TRAVIS_PYTHON_VERSION == 2* ]]; then TF_BINARY_URL="tf-nightly"
TF_BINARY_URL=https://ci.tensorflow.org/view/Nightly/job/nightly-matrix-cpu/TF_BUILD_IS_OPT=OPT,TF_BUILD_IS_PIP=PIP,TF_BUILD_PYTHON_VERSION=PYTHON2,label=cpu-slave/lastSuccessfulBuild/artifact/pip_test/whl/tensorflow-${TF_VERSION}-cp27-none-linux_x86_64.whl
fi
if [[ $TRAVIS_PYTHON_VERSION == 3.4* ]]; then
TF_BINARY_URL=https://ci.tensorflow.org/view/Nightly/job/nightly-matrix-cpu/TF_BUILD_IS_OPT=OPT,TF_BUILD_IS_PIP=PIP,TF_BUILD_PYTHON_VERSION=PYTHON3,label=cpu-slave/lastSuccessfulBuild/artifact/pip_test/whl/tensorflow-${TF_VERSION}-cp34-cp34m-linux_x86_64.whl
fi
if [[ $TRAVIS_PYTHON_VERSION == 3.5* ]]; then
TF_BINARY_URL=https://ci.tensorflow.org/view/Nightly/job/nightly-python35-linux-cpu/lastSuccessfulBuild/artifact/pip_test/whl/tensorflow-${TF_VERSION}-cp35-cp35m-linux_x86_64.whl
fi
fi fi
......
Markdown is supported
0% or
You are about to add 0 people to the discussion. Proceed with caution.
Finish editing this message first!
Please register or to comment