Update README.md
Browse files
README.md
CHANGED
@@ -7,20 +7,20 @@ license: apache-2.0
|
|
7 |
tags:
|
8 |
- "text generation"
|
9 |
---
|
10 |
-
|
11 |
GPT-2 model from Lithuania using Wikipedia corpus dataset based on GPT-2 small model.
|
12 |
|
13 |
This is only the first version of the model, over time model will be improved using a bigger dataset and better data preparation.
|
14 |
|
15 |
-
|
16 |
This model was pre-trained with 180MB of Lithuanian Wikipedia. The texts are tokenized using a byte-level version of Byte Pair Encoding (BPE).
|
17 |
|
18 |
-
|
19 |
The model was trained on wiki-corpus for 40 hours using NVIDIA Tesla P100 GPU.
|
20 |
|
21 |
-
|
22 |
|
23 |
-
|
24 |
|
25 |
``` from transformers import AutoTokenizer, TFAutoModelWithLMHead
|
26 |
import tensorflow as tf
|
@@ -33,7 +33,7 @@ tokenizer.model_max_length=1024
|
|
33 |
|
34 |
model.eval()
|
35 |
```
|
36 |
-
|
37 |
|
38 |
``` text = "tekstas"
|
39 |
inputs = tokenizer.encode(text, return_tensors="tf")
|
|
|
7 |
tags:
|
8 |
- "text generation"
|
9 |
---
|
10 |
+
### Model description
|
11 |
GPT-2 model from Lithuania using Wikipedia corpus dataset based on GPT-2 small model.
|
12 |
|
13 |
This is only the first version of the model, over time model will be improved using a bigger dataset and better data preparation.
|
14 |
|
15 |
+
### Training data
|
16 |
This model was pre-trained with 180MB of Lithuanian Wikipedia. The texts are tokenized using a byte-level version of Byte Pair Encoding (BPE).
|
17 |
|
18 |
+
### Training
|
19 |
The model was trained on wiki-corpus for 40 hours using NVIDIA Tesla P100 GPU.
|
20 |
|
21 |
+
### How to use
|
22 |
|
23 |
+
## Load model
|
24 |
|
25 |
``` from transformers import AutoTokenizer, TFAutoModelWithLMHead
|
26 |
import tensorflow as tf
|
|
|
33 |
|
34 |
model.eval()
|
35 |
```
|
36 |
+
## Generate text
|
37 |
|
38 |
``` text = "tekstas"
|
39 |
inputs = tokenizer.encode(text, return_tensors="tf")
|