XLM-RoBERTa base (xlm-roberta-base
) finetuned on squad v1.1.
Training-specifications:
- training_epochs: 3.0
- max_seq_length: 384
- batch_size: 16
- dataset_name: squad
- doc_stride 128
Train-results:
{
"epoch": 3.0,
"init_mem_cpu_alloc_delta": 991453184,
"init_mem_cpu_peaked_delta": 0,
"init_mem_gpu_alloc_delta": 1109893120,
"init_mem_gpu_peaked_delta": 0,
"train_mem_cpu_alloc_delta": 14753792,
"train_mem_cpu_peaked_delta": 0,
"train_mem_gpu_alloc_delta": 3330195456,
"train_mem_gpu_peaked_delta": 8287144960,
"train_runtime": 11376.3034,
"train_samples": 89597,
"train_samples_per_second": 1.477
}
Eval-results:
{
"epoch": 3.0,
"eval_samples": 10918,
"exact_match": 82.06244087038789,
"f1": 89.09539709124654
}