DeividasM commited on
Commit
f6fc929
1 Parent(s): 91378fc

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +6 -6
README.md CHANGED
@@ -7,20 +7,20 @@ license: apache-2.0
7
  tags:
8
  - "text generation"
9
  ---
10
- ## Model description
11
  GPT-2 model from Lithuania using Wikipedia corpus dataset based on GPT-2 small model.
12
 
13
  This is only the first version of the model, over time model will be improved using a bigger dataset and better data preparation.
14
 
15
- ## Training data
16
  This model was pre-trained with 180MB of Lithuanian Wikipedia. The texts are tokenized using a byte-level version of Byte Pair Encoding (BPE).
17
 
18
- ## Training
19
  The model was trained on wiki-corpus for 40 hours using NVIDIA Tesla P100 GPU.
20
 
21
- ## How to use
22
 
23
- ### Load model
24
 
25
  ``` from transformers import AutoTokenizer, TFAutoModelWithLMHead
26
  import tensorflow as tf
@@ -33,7 +33,7 @@ tokenizer.model_max_length=1024
33
 
34
  model.eval()
35
  ```
36
- ### Generate text
37
 
38
  ``` text = "tekstas"
39
  inputs = tokenizer.encode(text, return_tensors="tf")
 
7
  tags:
8
  - "text generation"
9
  ---
10
+ ### Model description
11
  GPT-2 model from Lithuania using Wikipedia corpus dataset based on GPT-2 small model.
12
 
13
  This is only the first version of the model, over time model will be improved using a bigger dataset and better data preparation.
14
 
15
+ ### Training data
16
  This model was pre-trained with 180MB of Lithuanian Wikipedia. The texts are tokenized using a byte-level version of Byte Pair Encoding (BPE).
17
 
18
+ ### Training
19
  The model was trained on wiki-corpus for 40 hours using NVIDIA Tesla P100 GPU.
20
 
21
+ ### How to use
22
 
23
+ ## Load model
24
 
25
  ``` from transformers import AutoTokenizer, TFAutoModelWithLMHead
26
  import tensorflow as tf
 
33
 
34
  model.eval()
35
  ```
36
+ ## Generate text
37
 
38
  ``` text = "tekstas"
39
  inputs = tokenizer.encode(text, return_tensors="tf")