Skip to content
Projects
Groups
Snippets
Help
Loading...
Help
Support
Keyboard shortcuts
?
Submit feedback
Contribute to GitLab
Sign in
Toggle navigation
S
seminar-breakout
Project overview
Project overview
Details
Activity
Releases
Repository
Repository
Files
Commits
Branches
Tags
Contributors
Graph
Compare
Issues
0
Issues
0
List
Boards
Labels
Milestones
Merge Requests
0
Merge Requests
0
CI / CD
CI / CD
Pipelines
Jobs
Schedules
Analytics
Analytics
CI / CD
Repository
Value Stream
Wiki
Wiki
Members
Members
Collapse sidebar
Close sidebar
Activity
Graph
Create a new issue
Jobs
Commits
Issue Boards
Open sidebar
Shashank Suhas
seminar-breakout
Commits
608ad4a9
Commit
608ad4a9
authored
Nov 17, 2016
by
Yuxin Wu
Browse files
Options
Browse Files
Download
Email Patches
Plain Diff
use get_scalar_var for learning rate
parent
3facd518
Changes
16
Hide whitespace changes
Inline
Side-by-side
Showing
16 changed files
with
38 additions
and
54 deletions
+38
-54
examples/Atari2600/DQN.py
examples/Atari2600/DQN.py
+1
-2
examples/DoReFa-Net/alexnet-dorefa.py
examples/DoReFa-Net/alexnet-dorefa.py
+1
-2
examples/HED/hed.py
examples/HED/hed.py
+1
-3
examples/Inception/inception-bn.py
examples/Inception/inception-bn.py
+1
-3
examples/Inception/inceptionv3.py
examples/Inception/inceptionv3.py
+1
-3
examples/OpenAIGym/train-atari.py
examples/OpenAIGym/train-atari.py
+1
-6
examples/ResNet/cifar10-resnet.py
examples/ResNet/cifar10-resnet.py
+1
-4
examples/ResNet/imagenet-resnet.py
examples/ResNet/imagenet-resnet.py
+3
-9
examples/ResNet/svhn-resnet.py
examples/ResNet/svhn-resnet.py
+1
-6
examples/SpatialTransformer/mnist-addition.py
examples/SpatialTransformer/mnist-addition.py
+1
-2
examples/char-rnn/char-rnn.py
examples/char-rnn/char-rnn.py
+1
-2
examples/cifar-convnet.py
examples/cifar-convnet.py
+1
-3
tensorpack/predict/base.py
tensorpack/predict/base.py
+2
-2
tensorpack/predict/dataset.py
tensorpack/predict/dataset.py
+10
-2
tensorpack/tfutils/common.py
tensorpack/tfutils/common.py
+2
-2
tensorpack/tfutils/symbolic_functions.py
tensorpack/tfutils/symbolic_functions.py
+10
-3
No files found.
examples/Atari2600/DQN.py
View file @
608ad4a9
...
...
@@ -158,8 +158,7 @@ def get_config():
reward_clip
=
(
-
1
,
1
),
history_len
=
FRAME_HISTORY
)
lr
=
tf
.
Variable
(
0.001
,
trainable
=
False
,
name
=
'learning_rate'
)
tf
.
scalar_summary
(
'learning_rate'
,
lr
)
lr
=
symbf
.
get_scalar_var
(
'learning_rate'
,
1e-3
,
summary
=
True
)
return
TrainConfig
(
dataset
=
dataset_train
,
...
...
examples/DoReFa-Net/alexnet-dorefa.py
View file @
608ad4a9
...
...
@@ -219,8 +219,7 @@ def get_config():
data_train
=
get_data
(
'train'
)
data_test
=
get_data
(
'val'
)
lr
=
tf
.
Variable
(
1e-4
,
trainable
=
False
,
name
=
'learning_rate'
)
tf
.
scalar_summary
(
'learning_rate'
,
lr
)
lr
=
get_scalar_var
(
'learning_rate'
,
1e-4
,
summary
=
True
)
return
TrainConfig
(
dataset
=
data_train
,
...
...
examples/HED/hed.py
View file @
608ad4a9
...
...
@@ -162,9 +162,7 @@ def get_config():
step_per_epoch
=
dataset_train
.
size
()
*
40
dataset_val
=
get_data
(
'val'
)
lr
=
tf
.
Variable
(
3e-5
,
trainable
=
False
,
name
=
'learning_rate'
)
tf
.
scalar_summary
(
'learning_rate'
,
lr
)
lr
=
get_scalar_var
(
'learning_rate'
,
3e-5
,
summary
=
True
)
return
TrainConfig
(
dataset
=
dataset_train
,
optimizer
=
tf
.
train
.
AdamOptimizer
(
lr
,
epsilon
=
1e-3
),
...
...
examples/Inception/inception-bn.py
View file @
608ad4a9
...
...
@@ -152,9 +152,7 @@ def get_config():
step_per_epoch
=
5000
dataset_val
=
get_data
(
'val'
)
lr
=
tf
.
Variable
(
0.045
,
trainable
=
False
,
name
=
'learning_rate'
)
tf
.
scalar_summary
(
'learning_rate'
,
lr
)
lr
=
get_scalar_var
(
'learning_rate'
,
0.045
,
summary
=
True
)
return
TrainConfig
(
dataset
=
dataset_train
,
optimizer
=
tf
.
train
.
MomentumOptimizer
(
lr
,
0.9
),
...
...
examples/Inception/inceptionv3.py
View file @
608ad4a9
...
...
@@ -258,9 +258,7 @@ def get_config():
dataset_train
=
get_data
(
'train'
)
dataset_val
=
get_data
(
'val'
)
lr
=
tf
.
Variable
(
0.045
,
trainable
=
False
,
name
=
'learning_rate'
)
tf
.
scalar_summary
(
'learning_rate'
,
lr
)
lr
=
get_scalar_var
(
'learning_rate'
,
0.045
,
summary
=
True
)
return
TrainConfig
(
dataset
=
dataset_train
,
optimizer
=
tf
.
train
.
AdamOptimizer
(
lr
,
epsilon
=
1e-3
),
...
...
examples/OpenAIGym/train-atari.py
View file @
608ad4a9
...
...
@@ -188,9 +188,7 @@ def get_config():
master
=
MySimulatorMaster
(
namec2s
,
names2c
,
M
)
dataflow
=
BatchData
(
DataFromQueue
(
master
.
queue
),
BATCH_SIZE
)
lr
=
tf
.
Variable
(
0.001
,
trainable
=
False
,
name
=
'learning_rate'
)
tf
.
scalar_summary
(
'learning_rate'
,
lr
)
lr
=
symbf
.
get_scalar_var
(
'learning_rate'
,
0.001
,
summary
=
True
)
return
TrainConfig
(
dataset
=
dataflow
,
optimizer
=
tf
.
train
.
AdamOptimizer
(
lr
,
epsilon
=
1e-3
),
...
...
@@ -200,9 +198,6 @@ def get_config():
ScheduledHyperParamSetter
(
'entropy_beta'
,
[(
80
,
0.005
)]),
ScheduledHyperParamSetter
(
'explore_factor'
,
[(
80
,
2
),
(
100
,
3
),
(
120
,
4
),
(
140
,
5
)]),
HumanHyperParamSetter
(
'learning_rate'
),
HumanHyperParamSetter
(
'entropy_beta'
),
HumanHyperParamSetter
(
'explore_factor'
),
master
,
StartProcOrThread
(
master
),
PeriodicCallback
(
Evaluator
(
EVAL_EPISODE
,
[
'state'
],
[
'logits'
]),
2
),
...
...
examples/ResNet/cifar10-resnet.py
View file @
608ad4a9
...
...
@@ -142,9 +142,7 @@ def get_config():
step_per_epoch
=
dataset_train
.
size
()
dataset_test
=
get_data
(
'test'
)
lr
=
tf
.
Variable
(
0.01
,
trainable
=
False
,
name
=
'learning_rate'
)
tf
.
scalar_summary
(
'learning_rate'
,
lr
)
lr
=
get_scalar_var
(
'learning_rate'
,
0.01
,
summary
=
True
)
return
TrainConfig
(
dataset
=
dataset_train
,
optimizer
=
tf
.
train
.
MomentumOptimizer
(
lr
,
0.9
),
...
...
@@ -155,7 +153,6 @@ def get_config():
ScheduledHyperParamSetter
(
'learning_rate'
,
[(
1
,
0.1
),
(
82
,
0.01
),
(
123
,
0.001
),
(
300
,
0.0002
)])
]),
session_config
=
get_default_sess_config
(
0.9
),
model
=
Model
(
n
=
18
),
step_per_epoch
=
step_per_epoch
,
max_epoch
=
400
,
...
...
examples/ResNet/imagenet-resnet.py
View file @
608ad4a9
...
...
@@ -19,7 +19,7 @@ from tensorpack.tfutils.summary import *
"""
Training code of Pre-Activation version of ResNet on ImageNet.
Mainly follow the setup in fb.resnet.torch
It mainly follows the setup in fb.resnet.torch, and get similar performance.
"""
TOTAL_BATCH_SIZE
=
256
...
...
@@ -116,8 +116,7 @@ class Model(ModelDesc):
wrong
=
prediction_incorrect
(
logits
,
label
,
5
,
name
=
'wrong-top5'
)
add_moving_summary
(
tf
.
reduce_mean
(
wrong
,
name
=
'train-error-top5'
))
wd_w
=
1e-4
wd_cost
=
tf
.
mul
(
wd_w
,
regularize_cost
(
'.*/W'
,
tf
.
nn
.
l2_loss
),
name
=
'l2_regularize_loss'
)
wd_cost
=
tf
.
mul
(
1e-4
,
regularize_cost
(
'.*/W'
,
tf
.
nn
.
l2_loss
),
name
=
'l2_regularize_loss'
)
add_moving_summary
(
loss
,
wd_cost
)
self
.
cost
=
tf
.
add_n
([
loss
,
wd_cost
],
name
=
'cost'
)
...
...
@@ -186,11 +185,7 @@ def get_config():
dataset_train
=
get_data
(
'train'
)
dataset_val
=
get_data
(
'val'
)
sess_config
=
get_default_sess_config
(
0.99
)
lr
=
tf
.
Variable
(
0.1
,
trainable
=
False
,
name
=
'learning_rate'
)
tf
.
scalar_summary
(
'learning_rate'
,
lr
)
lr
=
get_scalar_var
(
'learning_rate'
,
0.1
,
summary
=
True
)
return
TrainConfig
(
dataset
=
dataset_train
,
optimizer
=
tf
.
train
.
MomentumOptimizer
(
lr
,
0.9
,
use_nesterov
=
True
),
...
...
@@ -203,7 +198,6 @@ def get_config():
[(
30
,
1e-2
),
(
60
,
1e-3
),
(
85
,
1e-4
),
(
95
,
1e-5
)]),
HumanHyperParamSetter
(
'learning_rate'
),
]),
session_config
=
sess_config
,
model
=
Model
(),
step_per_epoch
=
5000
,
max_epoch
=
110
,
...
...
examples/ResNet/svhn-resnet.py
View file @
608ad4a9
...
...
@@ -64,11 +64,7 @@ def get_config():
step_per_epoch
=
dataset_train
.
size
()
dataset_test
=
get_data
(
'test'
)
sess_config
=
get_default_sess_config
(
0.9
)
lr
=
tf
.
Variable
(
0.1
,
trainable
=
False
,
name
=
'learning_rate'
)
tf
.
scalar_summary
(
'learning_rate'
,
lr
)
lr
=
get_scalar_var
(
'learning_rate'
,
0.01
,
summary
=
True
)
return
TrainConfig
(
dataset
=
dataset_train
,
optimizer
=
tf
.
train
.
MomentumOptimizer
(
lr
,
0.9
),
...
...
@@ -80,7 +76,6 @@ def get_config():
ScheduledHyperParamSetter
(
'learning_rate'
,
[(
1
,
0.1
),
(
20
,
0.01
),
(
28
,
0.001
),
(
50
,
0.0001
)])
]),
session_config
=
sess_config
,
model
=
Model
(
n
=
18
),
step_per_epoch
=
step_per_epoch
,
max_epoch
=
500
,
...
...
examples/SpatialTransformer/mnist-addition.py
View file @
608ad4a9
...
...
@@ -144,8 +144,7 @@ def get_config():
dataset_train
,
dataset_test
=
get_data
(
True
),
get_data
(
False
)
step_per_epoch
=
dataset_train
.
size
()
*
5
lr
=
symbolic_functions
.
get_scalar_var
(
'learning_rate'
,
5e-4
)
tf
.
scalar_summary
(
'learning_rate'
,
lr
)
lr
=
symbf
.
get_scalar_var
(
'learning_rate'
,
5e-4
,
summary
=
True
)
return
TrainConfig
(
dataset
=
dataset_train
,
...
...
examples/char-rnn/char-rnn.py
View file @
608ad4a9
...
...
@@ -112,8 +112,7 @@ def get_config():
ds
=
BatchData
(
ds
,
param
.
batch_size
)
step_per_epoch
=
ds
.
size
()
lr
=
tf
.
Variable
(
2e-3
,
trainable
=
False
,
name
=
'learning_rate'
)
tf
.
scalar_summary
(
'learning_rate'
,
lr
)
lr
=
symbolic_functions
.
get_scalar_var
(
'learning_rate'
,
2e-3
,
summary
=
True
)
return
TrainConfig
(
dataset
=
ds
,
...
...
examples/cifar-convnet.py
View file @
608ad4a9
...
...
@@ -110,9 +110,7 @@ def get_config(cifar_classnum):
sess_config
=
get_default_sess_config
(
0.5
)
lr
=
tf
.
Variable
(
1e-2
,
name
=
'learning_rate'
,
dtype
=
tf
.
float32
,
trainable
=
False
)
tf
.
scalar_summary
(
'learning_rate'
,
lr
)
lr
=
symbf
.
get_scalar_var
(
'learning_rate'
,
1e-2
,
summary
=
True
)
def
lr_func
(
lr
):
if
lr
<
3e-5
:
raise
StopTraining
()
...
...
tensorpack/predict/base.py
View file @
608ad4a9
...
...
@@ -87,9 +87,9 @@ class OfflinePredictor(OnlinePredictor):
def
__init__
(
self
,
config
):
self
.
graph
=
tf
.
Graph
()
with
self
.
graph
.
as_default
():
input_
va
rs
=
config
.
model
.
get_input_vars
()
input_
placehd
rs
=
config
.
model
.
get_input_vars
()
with
TowerContext
(
''
,
False
):
config
.
model
.
build_graph
(
input_
va
rs
)
config
.
model
.
build_graph
(
input_
placehd
rs
)
input_vars
=
get_tensors_by_names
(
config
.
input_names
)
output_vars
=
get_tensors_by_names
(
config
.
output_names
)
...
...
tensorpack/predict/dataset.py
View file @
608ad4a9
...
...
@@ -56,7 +56,11 @@ class SimpleDatasetPredictor(DatasetPredictorBase):
def
get_result
(
self
):
""" A generator to produce prediction for each data"""
with
tqdm
(
total
=
self
.
dataset
.
size
())
as
pbar
:
try
:
sz
=
self
.
dataset
.
size
()
except
NotImplementedError
:
sz
=
0
with
tqdm
(
total
=
sz
)
as
pbar
:
for
dp
in
self
.
dataset
.
get_data
():
res
=
self
.
predictor
(
dp
)
yield
res
...
...
@@ -111,7 +115,11 @@ class MultiProcessDatasetPredictor(DatasetPredictorBase):
ensure_proc_terminate
(
self
.
workers
+
[
self
.
result_queue
,
self
.
inqueue_proc
])
def
get_result
(
self
):
with
tqdm
(
total
=
self
.
dataset
.
size
())
as
pbar
:
try
:
sz
=
self
.
dataset
.
size
()
except
NotImplementedError
:
sz
=
0
with
tqdm
(
total
=
sz
)
as
pbar
:
die_cnt
=
0
while
True
:
res
=
self
.
result_queue
.
get
()
...
...
tensorpack/tfutils/common.py
View file @
608ad4a9
...
...
@@ -21,12 +21,12 @@ __all__ = ['get_default_sess_config',
'clear_collection'
,
'freeze_collection'
]
def
get_default_sess_config
(
mem_fraction
=
0.9
):
def
get_default_sess_config
(
mem_fraction
=
0.9
9
):
"""
Return a better session config to use as default.
Tensorflow default session config consume too much resources.
:param mem_fraction: fraction of memory to use.
:param mem_fraction: fraction of memory to use.
default to 0.99
:returns: a `tf.ConfigProto` object.
"""
conf
=
tf
.
ConfigProto
()
...
...
tensorpack/tfutils/symbolic_functions.py
View file @
608ad4a9
...
...
@@ -104,7 +104,14 @@ def huber_loss(x, delta=1, name='huber_loss'):
abscost
*
delta
-
0.5
*
delta
**
2
),
name
=
name
)
def
get_scalar_var
(
name
,
init_value
):
return
tf
.
get_variable
(
name
,
shape
=
[],
def
get_scalar_var
(
name
,
init_value
,
summary
=
False
,
trainable
=
False
):
"""
get a scalar variable with certain initial value
:param summary: summary this variable
"""
ret
=
tf
.
get_variable
(
name
,
shape
=
[],
initializer
=
tf
.
constant_initializer
(
init_value
),
trainable
=
False
)
trainable
=
trainable
)
if
summary
:
tf
.
scalar_summary
(
name
,
ret
)
return
ret
Write
Preview
Markdown
is supported
0%
Try again
or
attach a new file
Attach a file
Cancel
You are about to add
0
people
to the discussion. Proceed with caution.
Finish editing this message first!
Cancel
Please
register
or
sign in
to comment