Commit 55f640f7 authored by Yuxin Wu's avatar Yuxin Wu

fix #1046

parent 8566797f
......@@ -99,6 +99,7 @@ If you're unable to scale to multiple GPUs almost linearly:
If not, it's a bug or an environment setup problem.
2. Then note that your model may have a different communication-computation pattern that affects efficiency.
There isn't a simple answer to this.
You may try a different multi-GPU trainer; the speed can vary a lot in rare cases.
You may try a different multi-GPU trainer; the speed can vary a lot between
trainers in rare cases.
Note that scalibility is always measured by keeping "batch size per GPU" constant.
......@@ -118,7 +118,7 @@ def add_tensor_summary(x, types, name=None, collections=None,
if name is None:
name = x.op.name
ctx = get_current_tower_context()
if ctx is not None and not ctx.is_main_training_tower:
if main_tower_only and ctx is not None and not ctx.is_main_training_tower:
return
SUMMARY_TYPES_DIC = {
......
Markdown is supported
0% or
You are about to add 0 people to the discussion. Proceed with caution.
Finish editing this message first!
Please register or to comment