Commit 55f640f7 authored by Yuxin Wu's avatar Yuxin Wu

fix #1046

parent 8566797f
...@@ -99,6 +99,7 @@ If you're unable to scale to multiple GPUs almost linearly: ...@@ -99,6 +99,7 @@ If you're unable to scale to multiple GPUs almost linearly:
If not, it's a bug or an environment setup problem. If not, it's a bug or an environment setup problem.
2. Then note that your model may have a different communication-computation pattern that affects efficiency. 2. Then note that your model may have a different communication-computation pattern that affects efficiency.
There isn't a simple answer to this. There isn't a simple answer to this.
You may try a different multi-GPU trainer; the speed can vary a lot in rare cases. You may try a different multi-GPU trainer; the speed can vary a lot between
trainers in rare cases.
Note that scalibility is always measured by keeping "batch size per GPU" constant. Note that scalibility is always measured by keeping "batch size per GPU" constant.
...@@ -118,7 +118,7 @@ def add_tensor_summary(x, types, name=None, collections=None, ...@@ -118,7 +118,7 @@ def add_tensor_summary(x, types, name=None, collections=None,
if name is None: if name is None:
name = x.op.name name = x.op.name
ctx = get_current_tower_context() ctx = get_current_tower_context()
if ctx is not None and not ctx.is_main_training_tower: if main_tower_only and ctx is not None and not ctx.is_main_training_tower:
return return
SUMMARY_TYPES_DIC = { SUMMARY_TYPES_DIC = {
......
Markdown is supported
0% or
You are about to add 0 people to the discussion. Proceed with caution.
Finish editing this message first!
Please register or to comment