nickrwu commited on
Commit
33d3f8c
1 Parent(s): f1c01d2

End of training

Browse files
Files changed (4) hide show
  1. README.md +7 -9
  2. tokenizer.json +2 -2
  3. tokenizer_config.json +7 -0
  4. training_args.bin +1 -1
README.md CHANGED
@@ -1,6 +1,4 @@
1
  ---
2
- license: mit
3
- base_model: LIAMF-USP/roberta-large-finetuned-race
4
  tags:
5
  - generated_from_trainer
6
  metrics:
@@ -18,11 +16,11 @@ should probably proofread and complete it, then remove this comment. -->
18
 
19
  # roberta-mqa-rat
20
 
21
- This model is a fine-tuned version of [LIAMF-USP/roberta-large-finetuned-race](https://huggingface.co/LIAMF-USP/roberta-large-finetuned-race) on an unknown dataset.
22
  It achieves the following results on the evaluation set:
23
- - Loss: 1.6094
24
  - Accuracy: 0.2144
25
  - F1: 0.1861
 
26
  - Precision: 0.2104
27
  - Recall: 0.2059
28
 
@@ -54,11 +52,11 @@ The following hyperparameters were used during training:
54
 
55
  ### Training results
56
 
57
- | Training Loss | Epoch | Step | Validation Loss | Accuracy | F1 | Precision | Recall |
58
- |:-------------:|:-----:|:-----:|:---------------:|:--------:|:------:|:---------:|:------:|
59
- | 1.6189 | 1.0 | 3712 | 1.6094 | 0.2097 | 0.0731 | 0.0854 | 0.1997 |
60
- | 1.6176 | 2.0 | 7424 | 1.6094 | 0.2084 | 0.0971 | 0.1999 | 0.1982 |
61
- | 1.6154 | 3.0 | 11136 | 1.6094 | 0.2144 | 0.1861 | 0.2104 | 0.2059 |
62
 
63
 
64
  ### Framework versions
 
1
  ---
 
 
2
  tags:
3
  - generated_from_trainer
4
  metrics:
 
16
 
17
  # roberta-mqa-rat
18
 
19
+ This model was trained from scratch on an unknown dataset.
20
  It achieves the following results on the evaluation set:
 
21
  - Accuracy: 0.2144
22
  - F1: 0.1861
23
+ - Loss: 1.6094
24
  - Precision: 0.2104
25
  - Recall: 0.2059
26
 
 
52
 
53
  ### Training results
54
 
55
+ | Training Loss | Epoch | Step | Accuracy | F1 | Validation Loss | Precision | Recall |
56
+ |:-------------:|:-----:|:-----:|:--------:|:------:|:---------------:|:---------:|:------:|
57
+ | 1.6189 | 1.0 | 3712 | 0.2097 | 0.0731 | 1.6094 | 0.0854 | 0.1997 |
58
+ | 1.6176 | 2.0 | 7424 | 0.2084 | 0.0971 | 1.6094 | 0.1999 | 0.1982 |
59
+ | 1.6154 | 3.0 | 11136 | 0.2144 | 0.1861 | 1.6094 | 0.2104 | 0.2059 |
60
 
61
 
62
  ### Framework versions
tokenizer.json CHANGED
@@ -2,13 +2,13 @@
2
  "version": "1.0",
3
  "truncation": {
4
  "direction": "Right",
5
- "max_length": 256,
6
  "strategy": "LongestFirst",
7
  "stride": 0
8
  },
9
  "padding": {
10
  "strategy": {
11
- "Fixed": 256
12
  },
13
  "direction": "Right",
14
  "pad_to_multiple_of": null,
 
2
  "version": "1.0",
3
  "truncation": {
4
  "direction": "Right",
5
+ "max_length": 320,
6
  "strategy": "LongestFirst",
7
  "stride": 0
8
  },
9
  "padding": {
10
  "strategy": {
11
+ "Fixed": 320
12
  },
13
  "direction": "Right",
14
  "pad_to_multiple_of": null,
tokenizer_config.json CHANGED
@@ -48,10 +48,17 @@
48
  "eos_token": "</s>",
49
  "errors": "replace",
50
  "mask_token": "<mask>",
 
51
  "model_max_length": 512,
 
52
  "pad_token": "<pad>",
 
 
53
  "sep_token": "</s>",
 
54
  "tokenizer_class": "RobertaTokenizer",
55
  "trim_offsets": true,
 
 
56
  "unk_token": "<unk>"
57
  }
 
48
  "eos_token": "</s>",
49
  "errors": "replace",
50
  "mask_token": "<mask>",
51
+ "max_length": 128,
52
  "model_max_length": 512,
53
+ "pad_to_multiple_of": null,
54
  "pad_token": "<pad>",
55
+ "pad_token_type_id": 0,
56
+ "padding_side": "right",
57
  "sep_token": "</s>",
58
+ "stride": 0,
59
  "tokenizer_class": "RobertaTokenizer",
60
  "trim_offsets": true,
61
+ "truncation_side": "right",
62
+ "truncation_strategy": "longest_first",
63
  "unk_token": "<unk>"
64
  }
training_args.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:44f821910103f34f39e67b6d3374a73688d8e48bbecffae46dfa5c6a155737ad
3
  size 4920
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:76ee26ff7d8fb303073b380f7923ffecc736bbf9d6868b0d62a174456dfa2145
3
  size 4920