Abinaya Mahendiran commited on
Commit
752d5c1
1 Parent(s): 0388313

Updated README

Browse files
Files changed (1) hide show
  1. README.md +1 -1
README.md CHANGED
@@ -25,7 +25,7 @@ Pretrained model on Tamil language using a causal language modeling (CLM) object
25
  The GTP-2 model is trained on [oscar dataset - ta](https://huggingface.co/datasets/oscar) and [IndicNLP dataset - ta](https://indicnlp.ai4bharat.org/corpora/)
26
 
27
  ## Intended uses & limitations:
28
- You can use the raw model for next sentence prediction, but it's mostly intended to be fine-tuned on a downstream task. See the [model hub](https://huggingface.co/models?filter=gpt) to look for fine-tuned versions on a task that interests you.
29
 
30
  ## How to pretrain the model:
31
  To perform training, do the following steps,
25
  The GTP-2 model is trained on [oscar dataset - ta](https://huggingface.co/datasets/oscar) and [IndicNLP dataset - ta](https://indicnlp.ai4bharat.org/corpora/)
26
 
27
  ## Intended uses & limitations:
28
+ You can use the raw model for next sentence prediction, but it's mostly intended to be fine-tuned on a downstream task. See the [model hub](https://huggingface.co/models?filter=gpt2) to look for fine-tuned versions on a task that interests you.
29
 
30
  ## How to pretrain the model:
31
  To perform training, do the following steps,