File size: 1,122 Bytes
938b528
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
---
language: id
widget:
- text: "Wahai rembulan yang tertutup awan hujan"
---
# Indonesian GPT-2 finetuned on Indonesian poems
This is the [Indonesian gpt2-small model](https://huggingface.co/flax-community/gpt2-small-indonesian) fine-tuned to Indonesian poems. The dataset can be found in [here](https://huggingface.co/datasets/id_puisi) All training was done on Google Colab Jupyter Notebook (soon).

The information of the sub-dataset and the distribution of the training and evaluation dataset are as follows:
| split | count | percentage |
| ---------- | ---------- | -------------- |
| train    | 7,358     | 80%         |
| validation    | 1,890      | 20%         |


### Evaluation results 
The model evaluation results after 10 epochs are as follows:
| dataset | train loss | eval loss | eval perplexity |
| ---------- | ---------- | -------------- | ---------- |
| [id puisi](https://huggingface.co/datasets/id_puisi)   | 3.43      | 3.54         | 34.47   |

The logs and parameters can be found in [wandb page here](https://wandb.ai/ayamerushia/gpt-2_poem/runs/2qdccw34/overview?workspace=user-ayamerushia)