Skip to content
Projects
Groups
Snippets
Help
Loading...
Help
Support
Keyboard shortcuts
?
Submit feedback
Contribute to GitLab
Sign in
Toggle navigation
S
seminar-breakout
Project overview
Project overview
Details
Activity
Releases
Repository
Repository
Files
Commits
Branches
Tags
Contributors
Graph
Compare
Issues
0
Issues
0
List
Boards
Labels
Milestones
Merge Requests
0
Merge Requests
0
CI / CD
CI / CD
Pipelines
Jobs
Schedules
Analytics
Analytics
CI / CD
Repository
Value Stream
Wiki
Wiki
Members
Members
Collapse sidebar
Close sidebar
Activity
Graph
Create a new issue
Jobs
Commits
Issue Boards
Open sidebar
Shashank Suhas
seminar-breakout
Commits
3036e824
Commit
3036e824
authored
May 20, 2019
by
Yuxin Wu
Browse files
Options
Browse Files
Download
Email Patches
Plain Diff
update docs
parent
88796373
Changes
4
Hide whitespace changes
Inline
Side-by-side
Showing
4 changed files
with
25 additions
and
12 deletions
+25
-12
README.md
README.md
+2
-2
docs/tutorial/inference.md
docs/tutorial/inference.md
+8
-4
tensorpack/models/conv2d.py
tensorpack/models/conv2d.py
+2
-3
tensorpack/models/nonlin.py
tensorpack/models/nonlin.py
+13
-3
No files found.
README.md
View file @
3036e824
...
@@ -35,7 +35,7 @@ See [tutorials and documentations](http://tensorpack.readthedocs.io/tutorial/ind
...
@@ -35,7 +35,7 @@ See [tutorials and documentations](http://tensorpack.readthedocs.io/tutorial/ind
## Examples:
## Examples:
We refuse toy examples.
We refuse toy examples.
Instead of showing tiny CNNs trained on MNIST/Cifar10,
Instead of showing tiny CNNs trained on MNIST/Cifar10,
we provide training scripts that reproduce well-known papers.
we provide training scripts that reproduce well-known papers.
We refuse low-quality implementations.
We refuse low-quality implementations.
...
@@ -68,7 +68,7 @@ Dependencies:
...
@@ -68,7 +68,7 @@ Dependencies:
+
Python 2.7 or 3.3+. Python 2.7 is supported until
[
it retires in 2020
](
https://pythonclock.org/
)
.
+
Python 2.7 or 3.3+. Python 2.7 is supported until
[
it retires in 2020
](
https://pythonclock.org/
)
.
+
Python bindings for OpenCV. (Optional, but required by a lot of features)
+
Python bindings for OpenCV. (Optional, but required by a lot of features)
+
TensorFlow ≥ 1.3, < 2. (
Optional,
if you only want to use
`tensorpack.dataflow`
alone as a data processing library)
+
TensorFlow ≥ 1.3, < 2. (
Not required
if you only want to use
`tensorpack.dataflow`
alone as a data processing library)
```
```
pip install --upgrade git+https://github.com/tensorpack/tensorpack.git
pip install --upgrade git+https://github.com/tensorpack/tensorpack.git
# or add `--user` to install to user's local directories
# or add `--user` to install to user's local directories
...
...
docs/tutorial/inference.md
View file @
3036e824
...
@@ -40,15 +40,15 @@ for inference demo after training.
...
@@ -40,15 +40,15 @@ for inference demo after training.
It has functionailities to build the graph, load the checkpoint, and
It has functionailities to build the graph, load the checkpoint, and
return a callable for you for simple prediction. Refer to its docs for details.
return a callable for you for simple prediction. Refer to its docs for details.
OfflinePredictor is only for quick demo purposes.
To use it, you need to provide your model, checkpoint, and define what are the
It runs inference on numpy arrays, therefore may not be the most efficient way.
input & output tensors to infer with. You can obtain names of tensors by
It also has very limited functionalities
.
`print()`
, or assign a name to a tensor by
`tf.identity(x, name=)`
.
A simple example of how it works:
A simple example of how it works:
```
python
```
python
pred_config
=
PredictConfig
(
pred_config
=
PredictConfig
(
session_init
=
get_model_loader
(
model_path
),
model
=
YourModel
(),
model
=
YourModel
(),
session_init
=
get_model_loader
(
model_path
),
input_names
=
[
'input1'
,
'input2'
],
# tensor names in the graph, or name of the declared inputs
input_names
=
[
'input1'
,
'input2'
],
# tensor names in the graph, or name of the declared inputs
output_names
=
[
'output1'
,
'output2'
])
# tensor names in the graph
output_names
=
[
'output1'
,
'output2'
])
# tensor names in the graph
predictor
=
OfflinePredictor
(
pred_config
)
predictor
=
OfflinePredictor
(
pred_config
)
...
@@ -60,6 +60,10 @@ e.g., use NHWC format, support encoded image format, etc.
...
@@ -60,6 +60,10 @@ e.g., use NHWC format, support encoded image format, etc.
You can make these changes inside the
`model`
or
`tower_func`
in your
`PredictConfig`
.
You can make these changes inside the
`model`
or
`tower_func`
in your
`PredictConfig`
.
The example in
[
examples/basics/export-model.py
](
../examples/basics/export-model.py
)
demonstrates such an altered inference graph.
The example in
[
examples/basics/export-model.py
](
../examples/basics/export-model.py
)
demonstrates such an altered inference graph.
OfflinePredictor is only for quick demo purposes.
It runs inference on numpy arrays, therefore may not be the most efficient way.
It also has very limited functionalities.
### Exporter
### Exporter
In addition to the standard checkpoint format tensorpack saved for you during training,
In addition to the standard checkpoint format tensorpack saved for you during training,
...
...
tensorpack/models/conv2d.py
View file @
3036e824
...
@@ -37,12 +37,11 @@ def Conv2D(
...
@@ -37,12 +37,11 @@ def Conv2D(
activity_regularizer
=
None
,
activity_regularizer
=
None
,
split
=
1
):
split
=
1
):
"""
"""
A wrapper around `tf.layers.Conv2D`.
Similar to `tf.layers.Conv2D`, but with some differences:
Some differences to maintain backward-compatibility:
1. Default kernel initializer is variance_scaling_initializer(2.0).
1. Default kernel initializer is variance_scaling_initializer(2.0).
2. Default padding is 'same'.
2. Default padding is 'same'.
3. Support 'split' argument to do group conv
. Note that this is not efficient
.
3. Support 'split' argument to do group conv
olution
.
Variable Names:
Variable Names:
...
...
tensorpack/models/nonlin.py
View file @
3036e824
...
@@ -4,6 +4,7 @@
...
@@ -4,6 +4,7 @@
import
tensorflow
as
tf
import
tensorflow
as
tf
from
..utils.develop
import
log_deprecated
from
..compat
import
tfv1
from
..compat
import
tfv1
from
.batch_norm
import
BatchNorm
from
.batch_norm
import
BatchNorm
from
.common
import
VariableHolder
,
layer_register
from
.common
import
VariableHolder
,
layer_register
...
@@ -36,7 +37,7 @@ def Maxout(x, num_unit):
...
@@ -36,7 +37,7 @@ def Maxout(x, num_unit):
@
layer_register
()
@
layer_register
()
def
PReLU
(
x
,
init
=
0.001
,
name
=
'output'
):
def
PReLU
(
x
,
init
=
0.001
,
name
=
None
):
"""
"""
Parameterized ReLU as in the paper `Delving Deep into Rectifiers: Surpassing
Parameterized ReLU as in the paper `Delving Deep into Rectifiers: Surpassing
Human-Level Performance on ImageNet Classification
Human-Level Performance on ImageNet Classification
...
@@ -45,16 +46,18 @@ def PReLU(x, init=0.001, name='output'):
...
@@ -45,16 +46,18 @@ def PReLU(x, init=0.001, name='output'):
Args:
Args:
x (tf.Tensor): input
x (tf.Tensor): input
init (float): initial value for the learnable slope.
init (float): initial value for the learnable slope.
name (str):
name of the output.
name (str):
deprecated argument. Don't use
Variable Names:
Variable Names:
* ``alpha``: learnable slope.
* ``alpha``: learnable slope.
"""
"""
if
name
is
not
None
:
log_deprecated
(
"PReLU(name=...) is deprecated! The output tensor will be named `output`."
)
init
=
tfv1
.
constant_initializer
(
init
)
init
=
tfv1
.
constant_initializer
(
init
)
alpha
=
tfv1
.
get_variable
(
'alpha'
,
[],
initializer
=
init
)
alpha
=
tfv1
.
get_variable
(
'alpha'
,
[],
initializer
=
init
)
x
=
((
1
+
alpha
)
*
x
+
(
1
-
alpha
)
*
tf
.
abs
(
x
))
x
=
((
1
+
alpha
)
*
x
+
(
1
-
alpha
)
*
tf
.
abs
(
x
))
ret
=
tf
.
multiply
(
x
,
0.5
,
name
=
name
)
ret
=
tf
.
multiply
(
x
,
0.5
,
name
=
name
or
None
)
ret
.
variables
=
VariableHolder
(
alpha
=
alpha
)
ret
.
variables
=
VariableHolder
(
alpha
=
alpha
)
return
ret
return
ret
...
@@ -64,7 +67,14 @@ def PReLU(x, init=0.001, name='output'):
...
@@ -64,7 +67,14 @@ def PReLU(x, init=0.001, name='output'):
def
BNReLU
(
x
,
name
=
None
):
def
BNReLU
(
x
,
name
=
None
):
"""
"""
A shorthand of BatchNormalization + ReLU.
A shorthand of BatchNormalization + ReLU.
Args:
x (tf.Tensor): the input
name: deprecated, don't use.
"""
"""
if
name
is
not
None
:
log_deprecated
(
"BNReLU(name=...) is deprecated! The output tensor will be named `output`."
)
x
=
BatchNorm
(
'bn'
,
x
)
x
=
BatchNorm
(
'bn'
,
x
)
x
=
tf
.
nn
.
relu
(
x
,
name
=
name
)
x
=
tf
.
nn
.
relu
(
x
,
name
=
name
)
return
x
return
x
Write
Preview
Markdown
is supported
0%
Try again
or
attach a new file
Attach a file
Cancel
You are about to add
0
people
to the discussion. Proceed with caution.
Finish editing this message first!
Cancel
Please
register
or
sign in
to comment