Skip to content
Projects
Groups
Snippets
Help
Loading...
Help
Support
Keyboard shortcuts
?
Submit feedback
Contribute to GitLab
Sign in
Toggle navigation
S
seminar-breakout
Project overview
Project overview
Details
Activity
Releases
Repository
Repository
Files
Commits
Branches
Tags
Contributors
Graph
Compare
Issues
0
Issues
0
List
Boards
Labels
Milestones
Merge Requests
0
Merge Requests
0
CI / CD
CI / CD
Pipelines
Jobs
Schedules
Analytics
Analytics
CI / CD
Repository
Value Stream
Wiki
Wiki
Members
Members
Collapse sidebar
Close sidebar
Activity
Graph
Create a new issue
Jobs
Commits
Issue Boards
Open sidebar
Shashank Suhas
seminar-breakout
Commits
0fe3f52b
You need to sign in or sign up before continuing.
Commit
0fe3f52b
authored
Dec 18, 2016
by
Yuxin Wu
Browse files
Options
Browse Files
Download
Email Patches
Plain Diff
update readme
parent
9b1d0907
Changes
3
Hide whitespace changes
Inline
Side-by-side
Showing
3 changed files
with
31 additions
and
14 deletions
+31
-14
README.md
README.md
+21
-13
examples/README.md
examples/README.md
+9
-1
tensorpack/dataflow/dataset/ilsvrc.py
tensorpack/dataflow/dataset/ilsvrc.py
+1
-0
No files found.
README.md
View file @
0fe3f52b
# tensorpack
# tensorpack
Neural Network Toolbox on TensorFlow
Neural Network Toolbox on TensorFlow
See some
[
examples
](
examples
)
to learn about the framework.
See some
[
examples
](
examples
)
to learn about the framework:
They're not only for demonstration of the framework -- you can train them and reproduce the results in papers.
### Vision:
+
[
DoReFa-Net: training binary / low bitwidth CNN on ImageNet
](
examples/DoReFa-Net
)
+
[
DoReFa-Net: training binary / low-bitwidth CNN on ImageNet
](
DoReFa-Net
)
+
[
ResNet for ImageNet/Cifar10/SVHN classification
](
examples/ResNet
)
+
[
ResNet for ImageNet/Cifar10/SVHN
](
ResNet
)
+
[
InceptionV3 on ImageNet
](
examples/Inception/inceptionv3.py
)
+
[
InceptionV3 on ImageNet
](
Inception/inceptionv3.py
)
+
[
Fully-convolutional Network for Holistically-Nested Edge Detection(HED)
](
examples/HED
)
+
[
Fully-convolutional Network for Holistically-Nested Edge Detection(HED)
](
HED
)
+
[
Spatial Transformer Network on MNIST addition
](
examples/SpatialTransformer
)
+
[
Spatial Transformer Networks on MNIST addition
](
SpatialTransformer
)
+
[
Generative Adversarial Network(GAN) variants, including DCGAN, Image2Image, InfoGAN
](
examples/GAN
)
+
[
Deep Q-Network(DQN) variants on Atari games
](
examples/Atari2600
)
### Reinforcement Learning:
+
[
Asynchronous Advantage Actor-Critic(A3C) with demos on OpenAI Gym
](
examples/OpenAIGym
)
+
[
Deep Q-Network(DQN) variants on Atari games
](
Atari2600
)
+
[
LSTM-CTC for speech recognition
](
examples/TIMIT
)
+
[
Asynchronous Advantage Actor-Critic(A3C) with demos on OpenAI Gym
](
OpenAIGym
)
+
[
char-RNN language model
](
examples/char-rnn
)
### Unsupervised Learning:
+
[
Several Generative Adversarial Network(GAN) variants, including DCGAN, Image2Image, InfoGAN
](
examples/GAN
)
### Speech / NLP:
+
[
LSTM-CTC for speech recognition
](
TIMIT
)
+
[
char-rnn for fun
](
char-rnn
)
The examples are not only for demonstration of the framework -- you can train them and reproduce the results in papers.
## Features:
## Features:
...
...
examples/README.md
View file @
0fe3f52b
...
@@ -6,7 +6,7 @@ Training examples with __reproducible__ and meaningful performance.
...
@@ -6,7 +6,7 @@ Training examples with __reproducible__ and meaningful performance.
## Vision:
## Vision:
+
[
An illustrative mnist example with explanation of the framework
](
mnist-convnet.py
)
+
[
An illustrative mnist example with explanation of the framework
](
mnist-convnet.py
)
+
[
A tiny SVHN ConvNet with 97.8% accuracy
](
svhn-digit-convnet.py
)
+
[
A tiny SVHN ConvNet with 97.8% accuracy
](
svhn-digit-convnet.py
)
+
[
DoReFa-Net: binary / low-bitwidth CNN on ImageNet
](
DoReFa-Net
)
+
[
DoReFa-Net:
training
binary / low-bitwidth CNN on ImageNet
](
DoReFa-Net
)
+
[
ResNet for ImageNet/Cifar10/SVHN
](
ResNet
)
+
[
ResNet for ImageNet/Cifar10/SVHN
](
ResNet
)
+
[
Inception-BN with 71% accuracy
](
Inception/inception-bn.py
)
+
[
Inception-BN with 71% accuracy
](
Inception/inception-bn.py
)
+
[
InceptionV3 with 74% accuracy (similar to the official code)
](
Inception/inceptionv3.py
)
+
[
InceptionV3 with 74% accuracy (similar to the official code)
](
Inception/inceptionv3.py
)
...
@@ -24,3 +24,11 @@ Training examples with __reproducible__ and meaningful performance.
...
@@ -24,3 +24,11 @@ Training examples with __reproducible__ and meaningful performance.
## Speech / NLP:
## Speech / NLP:
+
[
LSTM-CTC for speech recognition
](
TIMIT
)
+
[
LSTM-CTC for speech recognition
](
TIMIT
)
+
[
char-rnn for fun
](
char-rnn
)
+
[
char-rnn for fun
](
char-rnn
)
Note to contributors:
We have a high bar for examples. It needs to satisfy one of the following:
+
Reproduce performance of a published or well-known paper.
+
Get state-of-the-art performance on some task.
+
Illustrate a new way of using the library that are currently not covered.
tensorpack/dataflow/dataset/ilsvrc.py
View file @
0fe3f52b
...
@@ -68,6 +68,7 @@ class ILSVRCMeta(object):
...
@@ -68,6 +68,7 @@ class ILSVRCMeta(object):
for
line
in
f
.
readlines
():
for
line
in
f
.
readlines
():
name
,
cls
=
line
.
strip
()
.
split
()
name
,
cls
=
line
.
strip
()
.
split
()
ret
.
append
((
name
,
int
(
cls
)))
ret
.
append
((
name
,
int
(
cls
)))
assert
len
(
ret
)
return
ret
return
ret
def
get_per_pixel_mean
(
self
,
size
=
None
):
def
get_per_pixel_mean
(
self
,
size
=
None
):
...
...
Write
Preview
Markdown
is supported
0%
Try again
or
attach a new file
Attach a file
Cancel
You are about to add
0
people
to the discussion. Proceed with caution.
Finish editing this message first!
Cancel
Please
register
or
sign in
to comment