Skip to content
Projects
Groups
Snippets
Help
Loading...
Help
Support
Keyboard shortcuts
?
Submit feedback
Contribute to GitLab
Sign in
Toggle navigation
S
seminar-breakout
Project overview
Project overview
Details
Activity
Releases
Repository
Repository
Files
Commits
Branches
Tags
Contributors
Graph
Compare
Issues
0
Issues
0
List
Boards
Labels
Milestones
Merge Requests
0
Merge Requests
0
CI / CD
CI / CD
Pipelines
Jobs
Schedules
Analytics
Analytics
CI / CD
Repository
Value Stream
Wiki
Wiki
Members
Members
Collapse sidebar
Close sidebar
Activity
Graph
Create a new issue
Jobs
Commits
Issue Boards
Open sidebar
Shashank Suhas
seminar-breakout
Commits
f9f1e437
Commit
f9f1e437
authored
Nov 12, 2017
by
Yuxin Wu
Browse files
Options
Browse Files
Download
Email Patches
Plain Diff
update docs
parent
e0c1ee77
Changes
1
Hide whitespace changes
Inline
Side-by-side
Showing
1 changed file
with
7 additions
and
3 deletions
+7
-3
docs/tutorial/trainer.md
docs/tutorial/trainer.md
+7
-3
No files found.
docs/tutorial/trainer.md
View file @
f9f1e437
...
...
@@ -14,8 +14,10 @@ You'll only need to __select__ what trainer to use.
Following the terminology in TensorFlow,
a "tower" function is something that takes input tensors and adds __one replicate__ of the model to the graph.
Most types of neural-network training could fall into this category.
This concept is used mainly to support:
All non-base trainers in tensorpack is a subclass of
[
TowerTrainer
](
../modules/train.html#tensorpack.train.TowerTrainer
)
.
The concept of tower is used mainly to support:
1.
Data-parallel multi-GPU training, where a replicate is built on each GPU.
2.
Automatically building the graph for inference, where a replicate is built under inference mode.
...
...
@@ -24,8 +26,10 @@ This concept is used mainly to support:
### MultiGPU Trainers
For data-parallel multi-GPU training, different
[
multi-GPU trainers
](
http://tensorpack.readthedocs.io/en/latest/modules/train.html
)
implement different parallel logic, all reaching the same performance as the
[
official TF benchmark
](
https://www.tensorflow.org/performance/benchmarks
)
.
implement different parallel logic.
They take care of device placement, gradient averaging and synchronoization
in the efficient way and all reach the same performance as the
[
official TF benchmarks
](
https://www.tensorflow.org/performance/benchmarks
)
.
It takes only one line of code change to use them.
Note some common problems when using these trainers:
...
...
Write
Preview
Markdown
is supported
0%
Try again
or
attach a new file
Attach a file
Cancel
You are about to add
0
people
to the discussion. Proceed with caution.
Finish editing this message first!
Cancel
Please
register
or
sign in
to comment