Text Generation
Transformers
PyTorch
Thai
English
mpt
custom_code
text-generation-inference
mrp commited on
Commit
c54d47b
1 Parent(s): 120c6fb

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +4 -5
README.md CHANGED
@@ -8,13 +8,12 @@ language:
8
 
9
  WangChanLion is a Multilingual, instruction-finetuned on Southeast Asian Languages SEA-LION 7B using open-source, commercially permissible datasets sample from LAION OIG chip2 and infill_dbpedia, DataBricks Dolly v2, OpenAI TL;DR, Hello-SimpleAI HC3, dolphin, iapp_wiki_qa_squad, thaisum, xlsum, scb_mt_enth_2020, han dataset, xp3x and Open-Platypus, a total of ~500k samples. Non-commercial datasets were filtered out. Released under apache 2.0 license. The models are trained to perform a subset of instruction-following tasks we found most relevant: reading comprehension, brainstorming, and creative writing. In this model, we focus on Thai and English datasets. We perform Vicuna-style evaluation using human evaluation. In a similar manner to Dolly v2, we only use open-source, commercially permissive pretrained models and datasets. Our models are neither restricted by non-commercial clauses like LLaMA-based models nor non-compete clauses like models that use self-instruct datasets from ChatGPT.
10
 
11
- Developers: PyThaiNLP and VISTEC-depa AI Research Institute of Thailand
12
-
13
- Model type: SEA-LION 7B (MPT architecture)
14
 
15
  ## Model Sources
16
- Repository: https://github.com/vistec-AI/WangchanLion
17
- Demo: [demo_WangchanLion.ipynb - Colaboratory](https://colab.research.google.com/drive/1y_7oOU3ZJI0h4chUrXFL3K4kelW_OI2G?usp=sharing#scrollTo=4yN3Bo6iAH2L)
18
 
19
  # Direct Use
20
  Intended to be used as an instruction-following model for reading comprehension, brainstorming, and creative writing.
 
8
 
9
  WangChanLion is a Multilingual, instruction-finetuned on Southeast Asian Languages SEA-LION 7B using open-source, commercially permissible datasets sample from LAION OIG chip2 and infill_dbpedia, DataBricks Dolly v2, OpenAI TL;DR, Hello-SimpleAI HC3, dolphin, iapp_wiki_qa_squad, thaisum, xlsum, scb_mt_enth_2020, han dataset, xp3x and Open-Platypus, a total of ~500k samples. Non-commercial datasets were filtered out. Released under apache 2.0 license. The models are trained to perform a subset of instruction-following tasks we found most relevant: reading comprehension, brainstorming, and creative writing. In this model, we focus on Thai and English datasets. We perform Vicuna-style evaluation using human evaluation. In a similar manner to Dolly v2, we only use open-source, commercially permissive pretrained models and datasets. Our models are neither restricted by non-commercial clauses like LLaMA-based models nor non-compete clauses like models that use self-instruct datasets from ChatGPT.
10
 
11
+ - Developers: PyThaiNLP and VISTEC-depa AI Research Institute of Thailand
12
+ - Model type: SEA-LION 7B (MPT architecture)
 
13
 
14
  ## Model Sources
15
+ - Repository: https://github.com/vistec-AI/WangchanLion
16
+ - Demo: [demo_WangchanLion.ipynb - Colaboratory](https://colab.research.google.com/drive/1y_7oOU3ZJI0h4chUrXFL3K4kelW_OI2G?usp=sharing#scrollTo=4yN3Bo6iAH2L)
17
 
18
  # Direct Use
19
  Intended to be used as an instruction-following model for reading comprehension, brainstorming, and creative writing.