File size: 821 Bytes
71c7586 fad1ee6 |
1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 16 17 18 19 20 21 22 23 24 25 26 27 28 29 30 31 32 33 34 35 36 37 38 39 40 41 42 |
XLM-RoBERTa base (`xlm-roberta-base`) finetuned on squad v1.1.
**Training-specifications:**
- training_epochs: 3.0
- max_seq_length: 384
- batch_size: 16
- dataset_name: squad
- doc_stride 128
**Train-results:**
```
{
"epoch": 3.0,
"init_mem_cpu_alloc_delta": 991453184,
"init_mem_cpu_peaked_delta": 0,
"init_mem_gpu_alloc_delta": 1109893120,
"init_mem_gpu_peaked_delta": 0,
"train_mem_cpu_alloc_delta": 14753792,
"train_mem_cpu_peaked_delta": 0,
"train_mem_gpu_alloc_delta": 3330195456,
"train_mem_gpu_peaked_delta": 8287144960,
"train_runtime": 11376.3034,
"train_samples": 89597,
"train_samples_per_second": 1.477
}
```
**Eval-results:**
```
{
"epoch": 3.0,
"eval_samples": 10918,
"exact_match": 82.06244087038789,
"f1": 89.09539709124654
}
```
|