This is a variant of the pre-trained model BERT model. The model was pre-trained on texts in the Thai language and fine-tuned for word segmentation based on bert-base-multilingual-cased. This version of the model processes input texts with character-level with word-level incorporated with a lattice structure.
The scripts for the pre-training are available at tchayintr/latte-ptm-ws.
The LATTE scripts are available at tchayintr/latte-ws.
The model architecture is described in this paper.
The model is trained on multiple Thai word segmented datasets, including best2010, lst20, tlc (tnhc), vistec-tp-th-2021 (vistec2021) and wisesight_sentiment (ws160). The datasets can be accessed as follows:
The pre-trained model is distributed under the terms of the Creative Commons Attribution-ShareAlike 4.0.
This model was trained with GPU servers provided by Okumura-Funakoshi NLP Group.
- Downloads last month