redrussianarmy commited on
Commit
a3c0d6c
1 Parent(s): ccf68f4

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +5 -2
README.md CHANGED
@@ -4,7 +4,7 @@ In this repository I release GPT-2 model, that was trained on various texts for
4
 
5
  The model is meant to be an entry point for fine-tuning on other texts.
6
 
7
- # Training corpora
8
 
9
  I used a Turkish corpora that is taken from oscar-corpus.
10
 
@@ -14,7 +14,7 @@ With the Tokenizers library, I created a 52K byte-level BPE vocab based on the t
14
 
15
  After creating the vocab, I could train the GPT-2 for Turkish on two 2080TI over the complete training corpus (five epochs).
16
 
17
- # Using the model
18
 
19
  The model itself can be used in this way:
20
 
@@ -33,3 +33,6 @@ pipe = pipeline('text-generation', model="redrussianarmy/gpt2-turkish-cased",
33
  text = pipe("Akşamüstü yolda ilerlerken, ")[0]["generated_text"]
34
  print(text)
35
  ```
 
 
 
 
4
 
5
  The model is meant to be an entry point for fine-tuning on other texts.
6
 
7
+ ## Training corpora
8
 
9
  I used a Turkish corpora that is taken from oscar-corpus.
10
 
 
14
 
15
  After creating the vocab, I could train the GPT-2 for Turkish on two 2080TI over the complete training corpus (five epochs).
16
 
17
+ ## Using the model
18
 
19
  The model itself can be used in this way:
20
 
 
33
  text = pipe("Akşamüstü yolda ilerlerken, ")[0]["generated_text"]
34
  print(text)
35
  ```
36
+
37
+ ## Contact (Bugs, Feedback, Contribution and more)
38
+ For questions about the GPT2-Turkish model, just open an issue [here](https://github.com/redrussianarmy/gpt2-turkish/issues) 🤗