Text Generation
Transformers
PyTorch
TensorBoard
Arabic
aragpt2
custom_code
wissamantoun commited on
Commit
1ad45c4
1 Parent(s): 4c65fd0

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +4 -20
README.md CHANGED
@@ -14,6 +14,8 @@ widget:
14
 
15
  # Arabic GPT2
16
 
 
 
17
  You can find more information in our paper [AraGPT2](https://arxiv.org/abs/2012.15520)
18
 
19
  The code in this repository was used to train all GPT2 variants. The code support training and fine-tuning GPT2 on GPUs and TPUs via the TPUEstimator API.
@@ -39,7 +41,7 @@ from arabert.aragpt2.grover.modeling_gpt2 import GPT2LMHeadModel
39
 
40
  from arabert.preprocess import ArabertPreprocessor
41
 
42
- MODEL_NAME='aragpt2-mega'
43
  arabert_prep = ArabertPreprocessor(model_name=MODEL_NAME)
44
 
45
  text=""
@@ -75,25 +77,7 @@ python create_pretraining_data.py
75
 
76
  Finetuning:
77
  ```bash
78
- python3 run_pretraining.py \
79
- --input_file="gs://<GS_BUCKET>/pretraining_data/*" \
80
- --output_dir="gs://<GS_BUCKET>/pretraining_model/" \
81
- --config_file="config/small_hparams.json" \
82
- --batch_size=128 \
83
- --eval_batch_size=8 \
84
- --num_train_steps= \
85
- --num_warmup_steps= \
86
- --learning_rate= \
87
- --save_checkpoints_steps= \
88
- --max_seq_length=1024 \
89
- --max_eval_steps= \
90
- --optimizer="lamb" \
91
- --iterations_per_loop=5000 \
92
- --keep_checkpoint_max=10 \
93
- --use_tpu=True \
94
- --tpu_name=<TPU NAME> \
95
- --do_train=True \
96
- --do_eval=False
97
  ```
98
  # Model Sizes
99
 
 
14
 
15
  # Arabic GPT2
16
 
17
+ <img src="https://raw.githubusercontent.com/aub-mind/arabert/master/AraGPT2.png" width="100" align="left"/>
18
+
19
  You can find more information in our paper [AraGPT2](https://arxiv.org/abs/2012.15520)
20
 
21
  The code in this repository was used to train all GPT2 variants. The code support training and fine-tuning GPT2 on GPUs and TPUs via the TPUEstimator API.
 
41
 
42
  from arabert.preprocess import ArabertPreprocessor
43
 
44
+ MODEL_NAME='aubmindlab/aragpt2-mega'
45
  arabert_prep = ArabertPreprocessor(model_name=MODEL_NAME)
46
 
47
  text=""
 
77
 
78
  Finetuning:
79
  ```bash
80
+ python3 run_pretraining.py \\r\n --input_file="gs://<GS_BUCKET>/pretraining_data/*" \\r\n --output_dir="gs://<GS_BUCKET>/pretraining_model/" \\r\n --config_file="config/small_hparams.json" \\r\n --batch_size=128 \\r\n --eval_batch_size=8 \\r\n --num_train_steps= \\r\n --num_warmup_steps= \\r\n --learning_rate= \\r\n --save_checkpoints_steps= \\r\n --max_seq_length=1024 \\r\n --max_eval_steps= \\r\n --optimizer="lamb" \\r\n --iterations_per_loop=5000 \\r\n --keep_checkpoint_max=10 \\r\n --use_tpu=True \\r\n --tpu_name=<TPU NAME> \\r\n --do_train=True \\r\n --do_eval=False
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
81
  ```
82
  # Model Sizes
83