Samuel Fipps
commited on
Commit
•
553d60b
1
Parent(s):
b7cf562
Update README.md
Browse files
README.md
CHANGED
@@ -72,5 +72,30 @@ model-index:
|
|
72 |
value: 45.928
|
73 |
verified: true
|
74 |
---
|
|
|
75 |
|
76 |
-
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
72 |
value: 45.928
|
73 |
verified: true
|
74 |
---
|
75 |
+
Trained on Samsum train split.
|
76 |
|
77 |
+
Parameters for training:
|
78 |
+
no_decay = ["bias", "LayerNorm.weight", "layer_norm.weight"]
|
79 |
+
optimizer_grouped_parameters = [
|
80 |
+
{
|
81 |
+
"params": [p for n, p in model.named_parameters() if not any(nd in n for nd in no_decay)],
|
82 |
+
"weight_decay": 0.0,
|
83 |
+
},
|
84 |
+
{
|
85 |
+
"params": [p for n, p in model.named_parameters() if any(nd in n for nd in no_decay)],
|
86 |
+
"weight_decay": 0.0,
|
87 |
+
},
|
88 |
+
]
|
89 |
+
|
90 |
+
lr = 0.00005
|
91 |
+
optimizer = torch.optim.RAdam(optimizer_grouped_parameters, lr=lr)
|
92 |
+
|
93 |
+
lr_scheduler = get_scheduler(
|
94 |
+
name="linear",
|
95 |
+
optimizer=optimizer,
|
96 |
+
num_warmup_steps=0,
|
97 |
+
num_training_steps=50005)
|
98 |
+
|
99 |
+
This was only for 10K steps
|
100 |
+
|
101 |
+
More details coming soon
|