
from nltk import word_tokenize
from dataset import SENTENCES


def demo(sentence:list=["nltk分词demo"]):
    ret = []
    for s in sentence:
        rs = word_tokenize(s)
        ret.append(rs)
    return ret

if __name__ == "__main__":
    ret = demo(SENTENCES)
    for idx,seq in enumerate(ret) :
        print("sentence",SENTENCES[idx])
        print("segmentation,","/".join(seq) )
        print("")