rasyosef commited on
Commit
0a09afb
·
verified ·
1 Parent(s): 9e45bdd

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +23 -9
README.md CHANGED
@@ -1,18 +1,32 @@
1
  ---
2
  widget:
3
- - text: አዲስ አበባ
4
- example_title: Example 1
5
- - text: በ ኢንግሊዝ ፕሪምየር ሊግ
6
- example_title: Example 2
7
- - text: ፕሬዚዳንት ዶናልድ ትራምፕ
8
- example_title: Example 3
 
 
 
 
 
 
9
  ---
10
 
11
- # gpt2-small-amharic-128-v3
12
 
13
- This is a smaller version of the [gpt2](https://huggingface.co/openai-community/gpt2) decoder transformer model pretrained from scratch for **2 days** on **290 million tokens** of **Amharic** text. The **context size** of this model is **128** tokens. It has the same tokenizer as gpt2, trained from scratch using the same dataset with a vocabulary size of **8192**.
14
 
15
- This is a base model and hasn't undergone any supervised finetuing yet.
 
 
 
 
 
 
 
 
16
 
17
  ### Demo
18
 
 
1
  ---
2
  widget:
3
+ - text: አዲስ አበባ
4
+ example_title: Example 1
5
+ - text: በ ኢንግሊዝ ፕሪምየር ሊግ
6
+ example_title: Example 2
7
+ - text: ፕሬዚዳንት ዶናልድ ትራምፕ
8
+ example_title: Example 3
9
+ language:
10
+ - am
11
+ metrics:
12
+ - perplexity
13
+ library_name: transformers
14
+ pipeline_tag: text-generation
15
  ---
16
 
17
+ # gpt2-small-amharic-8k-128-v3
18
 
19
+ This is a smaller version of the [gpt2](https://huggingface.co/openai-community/gpt2) decoder transformer model pretrained from scratch for **1.5 days** on **290 million tokens** of **Amharic** text.
20
 
21
+ - It has **33.7 Million parameters**
22
+ - The **context size** of this model is **128** tokens.
23
+ - It has the same **tokenizer** as gpt2, trained from scratch using the same dataset with a vocabulary size of **8192**.
24
+ - This is a base model and hasn't undergone any supervised finetuing yet.
25
+
26
+ It achieves the following results on the evaluation set:
27
+
28
+ - `Loss: 3.59`
29
+ - `Perplexity: 36.23`
30
 
31
  ### Demo
32