Commit 2010c43d authored by Yuxin Wu's avatar Yuxin Wu

trigger rtfd

parent 045f3352
...@@ -5,4 +5,4 @@ Sphinx>=1.6 ...@@ -5,4 +5,4 @@ Sphinx>=1.6
recommonmark==0.4.0 recommonmark==0.4.0
sphinx_rtd_theme sphinx_rtd_theme
mock mock
tensorflow==1.5.0 tensorflow
...@@ -31,18 +31,17 @@ The tower function needs to follow some conventions: ...@@ -31,18 +31,17 @@ The tower function needs to follow some conventions:
* Only put variables __trainable by gradient descent__ into `TRAINABLE_VARIABLES`. * Only put variables __trainable by gradient descent__ into `TRAINABLE_VARIABLES`.
* Put variables that need to be saved into `MODEL_VARIABLES`. * Put variables that need to be saved into `MODEL_VARIABLES`.
3. It has to respect variable scopes: 3. It has to respect variable scopes:
* The name of any trainable variables created in the function must be like "variable_scope_name/variable/name". * The name of any trainable variables created in the function must be like "variable_scope_name/custom/name".
Don't depend on name_scope's name. Don't use variable_scope's name twice. Don't depend on name_scope's name. Don't use variable_scope's name twice.
* The creation of any trainable variables must respect variable reuse. * The creation of any trainable variables must respect __reuse__ variable scope.
To respect variable reuse, use `tf.get_variable` instead of To respect variable reuse, use `tf.get_variable` instead of `tf.Variable` in the function.
`tf.Variable` in the function. On the other hand, for non-trainable variables, it's OK to use `tf.Variable` to force creation of new variables in each tower.
For non-trainable variables, it's OK to use `tf.Variable` to force creation of new variables in each tower. 4. It will always be called under a `TowerContext`, which can be accessed by `get_current_tower_contxt()`.
4. It will always be called under a `TowerContext`. The context contains information about training/inference mode, reuse, etc.
which will contain information about training/inference mode, reuse, etc.
These conventions are easy to follow, and most layer wrappers (e.g., These conventions are easy to follow, and most layer wrappers (e.g.,
tf.layers/slim/tensorlayer) do follow them. Note that certain Keras layers do not tf.layers/slim/tensorlayer) do follow them. Note that certain Keras layers do not
follow these conventions and may crash if used within tensorpack. follow these conventions and will need some workarounds if used within tensorpack.
It's possible to write ones that are not, but all existing trainers in It's possible to write ones that are not, but all existing trainers in
tensorpack are subclass of [TowerTrainer](../modules/train.html#tensorpack.train.TowerTrainer). tensorpack are subclass of [TowerTrainer](../modules/train.html#tensorpack.train.TowerTrainer).
......
...@@ -173,9 +173,11 @@ class ModelDesc(ModelDescBase): ...@@ -173,9 +173,11 @@ class ModelDesc(ModelDescBase):
It has the following constraints in addition to :class:`ModelDescBase`: It has the following constraints in addition to :class:`ModelDescBase`:
1. :meth:`build_graph(...)` method should return a cost when called under a training context. 1. :meth:`build_graph(...)` method should return a cost when called under a training context.
The cost will be the final cost to be optimized by the optimizer. The cost will be the final cost to be optimized by the optimizer.
Therefore it should include necessary regularization. Therefore it should include necessary regularization.
2. Subclass is expected to implement :meth:`optimizer()` method. 2. Subclass is expected to implement :meth:`optimizer()` method.
""" """
def get_cost(self): def get_cost(self):
......
Markdown is supported
0% or
You are about to add 0 people to the discussion. Proceed with caution.
Finish editing this message first!
Please register or to comment