--- license: apache-2.0 datasets: - kobkrit/rd-taxqa - iapp_wiki_qa_squad - Thaweewat/alpaca-cleaned-52k-th - Thaweewat/instruction-wild-52k-th - Thaweewat/databricks-dolly-15k-th - Thaweewat/hc3-24k-th - Thaweewat/gpteacher-20k-th - Thaweewat/onet-m6-social - Thaweewat/alpaca-finance-43k-th language: - th - en library_name: transformers pipeline_tag: text-generation tags: - openthaigpt --- # 🇹🇭 OpenThaiGPT 0.1.0-beta OpenThaiGPT Version 0.1.0-beta is a 7B-parameter LLaMA model finetuned to follow Thai translated instructions below and makes use of the Huggingface LLaMA implementation. ## Support - Official website: https://openthaigpt.aieat.or.th - Facebook page: https://web.facebook.com/groups/openthaigpt - A Discord server for discussion and support [here](https://discord.gg/rUTp6dfVUF) - E-mail: kobkrit@iapp.co.th ## License **Source Code**: License Apache Software License 2.0.
**Weight**: For research use only (due to the Facebook LLama's Weight LICENSE).
Note that: A commercial use license for OpenThaiGPT 0.1.0 weight will be released later soon! ## Code and Weight **Finetune Code**: https://github.com/OpenThaiGPT/openthaigpt-finetune-010beta
**Inference Code**: https://github.com/OpenThaiGPT/openthaigpt
**Weight**: https://huggingface.co/kobkrit/openthaigpt-0.1.0-beta ## Sponsors Pantip.com, ThaiSC
### Powered by OpenThaiGPT Volunteers, Artificial Intelligence Entrepreneur Association of Thailand (AIEAT), and Artificial Intelligence Association of Thailand (AIAT) ### Authors Kobkrit Viriyayudhakorn (kobkrit@iapp.co.th), Sumeth Yuenyong (sumeth.yue@mahidol.edu) and Thaweewat Ruksujarit (thaweewr@scg.com). Disclaimer: Provided responses are not guaranteed.