diff --git a/train.py b/train.py index d8052f8..446b4b9 100644 --- a/train.py +++ b/train.py @@ -31,8 +31,7 @@ def __init__(self, net, data): self.global_step = tf.get_variable( 'global_step', [], initializer=tf.constant_initializer(0), trainable=False) -#self.learning_rate = tf.train.exponential_decay( -# self.initial_learning_rate, self.global_step, self.decay_steps, +#self.learning_rate = tf.train.exponential_decay(self.initial_learning_rate, self.global_step, self.decay_steps, # self.decay_rate, self.staircase, name='learning_rate') self.optimizer = tf.train.AdamOptimizer(