@@ -70,7 +70,7 @@ def train_batch(self, batch, step):
7070 torch .nn .utils .clip_grad_norm_ (self .model .parameters (), self .args .grad_norm )
7171
7272 self .optimizer .step ()
73- if self .args .schedule_lr :
73+ if self .args .schedule_lr and step <= self . args . schedule_step :
7474 self .scheduler .step ()
7575 return loss
7676
@@ -181,10 +181,11 @@ def parse_args(self, args=None):
181181 parser .add_argument ('-early_stop_tolerance' , type = int , default = 100 , help = '' )
182182 parser .add_argument ('-train_parallel' , action = 'store_true' , help = '' )
183183 parser .add_argument ('-schedule_lr' , action = 'store_true' , help = '' )
184- parser .add_argument ('-schedule_step' , type = int , default = 100000 , help = '' )
185- parser .add_argument ('-schedule_gamma' , type = float , default = 0.5 , help = '' )
184+ parser .add_argument ('-schedule_step' , type = int , default = 10000 , help = '' )
185+ parser .add_argument ('-schedule_gamma' , type = float , default = 0.1 , help = '' )
186186 parser .add_argument ('-processed' , action = 'store_true' , help = '' )
187187 parser .add_argument ('-prefetch' , action = 'store_true' , help = '' )
188+ parser .add_argument ('-lazy_loading' , action = 'store_true' , help = '' )
188189
189190 # model specific parameter
190191 parser .add_argument ("-embed_size" , type = int , default = 200 , help = '' )
0 commit comments