You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
New experimental model type LanguageModel to train generative language models (see the example GPT-2 configuration). The usage is the same as a sequence to sequence model except that "labels" data should not be set.
cosine_annealing learning rate decay
weight_decay parameter to apply decoupled weight decay regularization (as described in Loshchilov et al. 2017)
sampling_temperature parameter to control the randomness of the generation
Fixes and improvements
Improve correctness of MeanEncoder for variable lengths inputs (requires TensorFlow 1.13+)
Internal refactoring and changes to prepare for 2.0 transition