Amal17 commited on
Commit
ea484ba
1 Parent(s): 01228c6

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +13 -3
README.md CHANGED
@@ -6,6 +6,10 @@ tags:
6
  model-index:
7
  - name: wikipedia-20230601.ace
8
  results: []
 
 
 
 
9
  ---
10
 
11
  <!-- This model card has been generated automatically according to the information the Trainer had access to. You
@@ -13,13 +17,13 @@ should probably proofread and complete it, then remove this comment. -->
13
 
14
  # wikipedia-20230601.ace
15
 
16
- This model is a fine-tuned version of [distilgpt2](https://huggingface.co/distilgpt2) on the None dataset.
17
  It achieves the following results on the evaluation set:
18
  - Loss: 1.0119
19
 
20
  ## Model description
21
 
22
- More information needed
23
 
24
  ## Intended uses & limitations
25
 
@@ -50,10 +54,16 @@ The following hyperparameters were used during training:
50
  | 1.7159 | 2.0 | 1346 | 1.0161 |
51
  | 1.6392 | 3.0 | 2019 | 1.0119 |
52
 
 
 
 
 
 
 
53
 
54
  ### Framework versions
55
 
56
  - Transformers 4.32.0.dev0
57
  - Pytorch 2.0.1+cu117
58
  - Datasets 2.13.1
59
- - Tokenizers 0.13.3
 
6
  model-index:
7
  - name: wikipedia-20230601.ace
8
  results: []
9
+ datasets:
10
+ - graelo/wikipedia
11
+ metrics:
12
+ - perplexity
13
  ---
14
 
15
  <!-- This model card has been generated automatically according to the information the Trainer had access to. You
 
17
 
18
  # wikipedia-20230601.ace
19
 
20
+ This model is a fine-tuned version of [distilgpt2](https://huggingface.co/distilgpt2) on the graelo/wikipedia-20230601.ace dataset.
21
  It achieves the following results on the evaluation set:
22
  - Loss: 1.0119
23
 
24
  ## Model description
25
 
26
+ This model finetune distilgpt2 to Acehnese just for experiment purpose
27
 
28
  ## Intended uses & limitations
29
 
 
54
  | 1.7159 | 2.0 | 1346 | 1.0161 |
55
  | 1.6392 | 3.0 | 2019 | 1.0119 |
56
 
57
+ ### Perplexity
58
+
59
+ Datatest: `load_dataset("graelo/wikipedia", "20230601.ace", split="train[-10%:]")`
60
+
61
+ - original distilgpt2 : 40.5980
62
+ - this model (finetuned) : 3.9992
63
 
64
  ### Framework versions
65
 
66
  - Transformers 4.32.0.dev0
67
  - Pytorch 2.0.1+cu117
68
  - Datasets 2.13.1
69
+ - Tokenizers 0.13.3