Skip to content

Commit

Permalink
Remove weight_decay parameter in GPT-2
Browse files Browse the repository at this point in the history
This parameter appears to be non working or incorrectly tuned.
  • Loading branch information
guillaumekln committed Apr 10, 2019
1 parent c422d94 commit ac26785
Showing 1 changed file with 0 additions and 1 deletion.
1 change: 0 additions & 1 deletion config/models/gpt_2.py
Original file line number Diff line number Diff line change
Expand Up @@ -36,7 +36,6 @@ def auto_config(self, num_devices=1):
"params": {
"average_loss_in_time": True,
"optimizer": "AdamOptimizer",
"weight_decay": 0.01,
"learning_rate": 2.5e-4,
"decay_type": "cosine_annealing",
"decay_params": {
Expand Down

0 comments on commit ac26785

Please sign in to comment.