alphahg commited on
Commit
2cb0248
1 Parent(s): 4b6b633

update model card README.md

Browse files
Files changed (1) hide show
  1. README.md +11 -5
README.md CHANGED
@@ -1,6 +1,8 @@
1
  ---
2
  tags:
3
  - generated_from_trainer
 
 
4
  model-index:
5
  - name: awsome_bigbird_24387490
6
  results: []
@@ -11,7 +13,9 @@ should probably proofread and complete it, then remove this comment. -->
11
 
12
  # awsome_bigbird_24387490
13
 
14
- This model is a fine-tuned version of [monologg/kobigbird-bert-base](https://huggingface.co/monologg/kobigbird-bert-base) on the None dataset.
 
 
15
 
16
  ## Model description
17
 
@@ -34,18 +38,20 @@ The following hyperparameters were used during training:
34
  - train_batch_size: 32
35
  - eval_batch_size: 32
36
  - seed: 30
37
- - gradient_accumulation_steps: 8
38
- - total_train_batch_size: 256
39
  - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
40
  - lr_scheduler_type: linear
41
- - num_epochs: 1
42
  - mixed_precision_training: Native AMP
43
 
44
  ### Training results
45
 
46
  | Training Loss | Epoch | Step | Validation Loss |
47
  |:-------------:|:-----:|:----:|:---------------:|
48
- | No log | 1.0 | 137 | 1.5316 |
 
 
49
 
50
 
51
  ### Framework versions
 
1
  ---
2
  tags:
3
  - generated_from_trainer
4
+ datasets:
5
+ - custom_squad_v2
6
  model-index:
7
  - name: awsome_bigbird_24387490
8
  results: []
 
13
 
14
  # awsome_bigbird_24387490
15
 
16
+ This model is a fine-tuned version of [monologg/kobigbird-bert-base](https://huggingface.co/monologg/kobigbird-bert-base) on the custom_squad_v2 dataset.
17
+ It achieves the following results on the evaluation set:
18
+ - Loss: 1.1222
19
 
20
  ## Model description
21
 
 
38
  - train_batch_size: 32
39
  - eval_batch_size: 32
40
  - seed: 30
41
+ - gradient_accumulation_steps: 16
42
+ - total_train_batch_size: 512
43
  - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
44
  - lr_scheduler_type: linear
45
+ - num_epochs: 3
46
  - mixed_precision_training: Native AMP
47
 
48
  ### Training results
49
 
50
  | Training Loss | Epoch | Step | Validation Loss |
51
  |:-------------:|:-----:|:----:|:---------------:|
52
+ | No log | 0.99 | 21 | 1.2008 |
53
+ | No log | 1.99 | 42 | 1.1138 |
54
+ | No log | 2.99 | 63 | 1.1222 |
55
 
56
 
57
  ### Framework versions