Spaces:
Running
Running
Update tubes.py
Browse files
tubes.py
CHANGED
@@ -6,6 +6,7 @@ from sklearn.feature_extraction.text import CountVectorizer
|
|
6 |
from nltk.tokenize import RegexpTokenizer
|
7 |
import nltk
|
8 |
nltk.download('wordnet')
|
|
|
9 |
from nltk.tokenize import word_tokenize
|
10 |
from nltk.stem import WordNetLemmatizer
|
11 |
nltk.download('punkt')
|
|
|
6 |
from nltk.tokenize import RegexpTokenizer
|
7 |
import nltk
|
8 |
nltk.download('wordnet')
|
9 |
+
nltk.download('stopwords')
|
10 |
from nltk.tokenize import word_tokenize
|
11 |
from nltk.stem import WordNetLemmatizer
|
12 |
nltk.download('punkt')
|