# -*- coding: utf-8 -*-
# pip install nltk

# 分词
from nltk.tokenize import sent_tokenize, word_tokenize
# 停用词
from nltk.corpus import stopwords
# WordNet词形还原器
from nltk.stem import WordNetLemmatizer
# Porter词干提取器
from nltk.stem.porter import PorterStemmer
# Lancaster词干提取器
from nltk.stem.lancaster import LancasterStemmer
# Snowball词干提取器
from nltk.stem import SnowballStemmer

text = """
"""
sents = sent_tokenize(text)

words = word_tokenize(text)