hyt1912 commited on
Commit
231ab82
verified
1 Parent(s): fe2c9d8

Model save

Browse files
Files changed (1) hide show
  1. README.md +9 -9
README.md CHANGED
@@ -1,7 +1,7 @@
1
  ---
2
  library_name: transformers
3
  license: apache-2.0
4
- base_model: google/electra-small-discriminator
5
  tags:
6
  - generated_from_trainer
7
  model-index:
@@ -14,9 +14,9 @@ should probably proofread and complete it, then remove this comment. -->
14
 
15
  # electra-small-discriminator-finetuned-squad
16
 
17
- This model is a fine-tuned version of [google/electra-small-discriminator](https://huggingface.co/google/electra-small-discriminator) on an unknown dataset.
18
  It achieves the following results on the evaluation set:
19
- - Loss: 1.1425
20
 
21
  ## Model description
22
 
@@ -39,7 +39,7 @@ The following hyperparameters were used during training:
39
  - train_batch_size: 16
40
  - eval_batch_size: 16
41
  - seed: 42
42
- - optimizer: Use adamw_torch with betas=(0.9,0.999) and epsilon=1e-08 and optimizer_args=No additional optimizer arguments
43
  - lr_scheduler_type: linear
44
  - num_epochs: 3
45
 
@@ -47,14 +47,14 @@ The following hyperparameters were used during training:
47
 
48
  | Training Loss | Epoch | Step | Validation Loss |
49
  |:-------------:|:-----:|:-----:|:---------------:|
50
- | 1.3266 | 1.0 | 5533 | 1.2489 |
51
- | 1.1417 | 2.0 | 11066 | 1.1470 |
52
- | 1.0544 | 3.0 | 16599 | 1.1425 |
53
 
54
 
55
  ### Framework versions
56
 
57
- - Transformers 4.46.2
58
- - Pytorch 2.5.1+cu121
59
  - Datasets 3.1.0
60
  - Tokenizers 0.20.3
 
1
  ---
2
  library_name: transformers
3
  license: apache-2.0
4
+ base_model: hyt1912/electra-small-discriminator-finetuned-squad
5
  tags:
6
  - generated_from_trainer
7
  model-index:
 
14
 
15
  # electra-small-discriminator-finetuned-squad
16
 
17
+ This model is a fine-tuned version of [hyt1912/electra-small-discriminator-finetuned-squad](https://huggingface.co/hyt1912/electra-small-discriminator-finetuned-squad) on an unknown dataset.
18
  It achieves the following results on the evaluation set:
19
+ - Loss: 1.0458
20
 
21
  ## Model description
22
 
 
39
  - train_batch_size: 16
40
  - eval_batch_size: 16
41
  - seed: 42
42
+ - optimizer: Use OptimizerNames.ADAMW_TORCH with betas=(0.9,0.999) and epsilon=1e-08 and optimizer_args=No additional optimizer arguments
43
  - lr_scheduler_type: linear
44
  - num_epochs: 3
45
 
 
47
 
48
  | Training Loss | Epoch | Step | Validation Loss |
49
  |:-------------:|:-----:|:-----:|:---------------:|
50
+ | 1.0184 | 1.0 | 3729 | 1.0553 |
51
+ | 0.9004 | 2.0 | 7458 | 1.0349 |
52
+ | 0.7912 | 3.0 | 11187 | 1.0458 |
53
 
54
 
55
  ### Framework versions
56
 
57
+ - Transformers 4.46.3
58
+ - Pytorch 2.5.0
59
  - Datasets 3.1.0
60
  - Tokenizers 0.20.3