sberbank-ai commited on
Commit
d925dc3
1 Parent(s): dd01069

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +1 -1
README.md CHANGED
@@ -19,7 +19,7 @@ Bbpe tokenizer. 50257 + special tokens 107. Prefix tokens: '\<LM\>', '\<SC1>',..
19
  First half of the time model trained on the small part of all dataset (1%,3GB) and without prefixes in each task.
20
 
21
  For RSG, we trained as described in the T5 paper. First, we trained multitask for all tasks. Then we took the best checkpoint for the task and trained it further.
22
- RSG submit here https://russiansuperglue.com/login/submit_info/1936
23
 
24
  Total training time was around 35 days on 160 V100 GPUs.
25
 
 
19
  First half of the time model trained on the small part of all dataset (1%,3GB) and without prefixes in each task.
20
 
21
  For RSG, we trained as described in the T5 paper. First, we trained multitask for all tasks. Then we took the best checkpoint for the task and trained it further.
22
+ RSG submit here https://russiansuperglue.com/login/submit_info/2060
23
 
24
  Total training time was around 35 days on 160 V100 GPUs.
25