Skip to content
Projects
Groups
Snippets
Help
Loading...
Help
Support
Keyboard shortcuts
?
Submit feedback
Contribute to GitLab
Sign in
Toggle navigation
S
seminar-breakout
Project overview
Project overview
Details
Activity
Releases
Repository
Repository
Files
Commits
Branches
Tags
Contributors
Graph
Compare
Issues
0
Issues
0
List
Boards
Labels
Milestones
Merge Requests
0
Merge Requests
0
CI / CD
CI / CD
Pipelines
Jobs
Schedules
Analytics
Analytics
CI / CD
Repository
Value Stream
Wiki
Wiki
Members
Members
Collapse sidebar
Close sidebar
Activity
Graph
Create a new issue
Jobs
Commits
Issue Boards
Open sidebar
Shashank Suhas
seminar-breakout
Commits
584e9cd4
Commit
584e9cd4
authored
Sep 26, 2017
by
Yuxin Wu
Browse files
Options
Browse Files
Download
Email Patches
Plain Diff
Fix hyperparam and optimizer issue in distributed trainer (#431)
parent
c8028236
Changes
2
Hide whitespace changes
Inline
Side-by-side
Showing
2 changed files
with
3 additions
and
1 deletion
+3
-1
tensorpack/callbacks/param.py
tensorpack/callbacks/param.py
+1
-1
tensorpack/train/distributed.py
tensorpack/train/distributed.py
+2
-0
No files found.
tensorpack/callbacks/param.py
View file @
584e9cd4
...
@@ -66,7 +66,7 @@ class GraphVarParam(HyperParam):
...
@@ -66,7 +66,7 @@ class GraphVarParam(HyperParam):
def
setup_graph
(
self
):
def
setup_graph
(
self
):
""" Will setup the assign operator for that variable. """
""" Will setup the assign operator for that variable. """
all_vars
=
tf
.
globa
l_variables
()
all_vars
=
tf
.
al
l_variables
()
for
v
in
all_vars
:
for
v
in
all_vars
:
if
v
.
name
==
self
.
var_name
:
if
v
.
name
==
self
.
var_name
:
self
.
var
=
v
self
.
var
=
v
...
...
tensorpack/train/distributed.py
View file @
584e9cd4
...
@@ -204,6 +204,8 @@ class DistributedTrainerReplicated(MultiGPUTrainerBase):
...
@@ -204,6 +204,8 @@ class DistributedTrainerReplicated(MultiGPUTrainerBase):
cbs
=
self
.
_input_source
.
setup
(
self
.
model
.
get_inputs_desc
())
cbs
=
self
.
_input_source
.
setup
(
self
.
model
.
get_inputs_desc
())
self
.
config
.
callbacks
.
extend
(
cbs
)
self
.
config
.
callbacks
.
extend
(
cbs
)
# build the optimizer first, before entering any tower
self
.
model
.
get_optimizer
()
# Ngpu * Nvar * 2
# Ngpu * Nvar * 2
grad_list
=
MultiGPUTrainerBase
.
build_on_multi_tower
(
grad_list
=
MultiGPUTrainerBase
.
build_on_multi_tower
(
self
.
config
.
tower
,
self
.
config
.
tower
,
...
...
Write
Preview
Markdown
is supported
0%
Try again
or
attach a new file
Attach a file
Cancel
You are about to add
0
people
to the discussion. Proceed with caution.
Finish editing this message first!
Cancel
Please
register
or
sign in
to comment