Update README.md
ef4033f
verified
-
2.31 kB
initial commit
-
1.02 kB
Update README.md
-
29.7 kB
Upload punkt.png
punkt.zip
Detected Pickle imports (300)
- "collections.defaultdict",
- "__builtin__.set",
- "__builtin__.object",
- "nltk.tokenize.punkt.PunktToken",
- "nltk.tokenize.punkt.PunktLanguageVars",
- "copy_reg._reconstructor",
- "nltk.tokenize.punkt.PunktSentenceTokenizer",
- "__builtin__.int",
- "nltk.tokenize.punkt.PunktParameters",
- "collections.defaultdict",
- "__builtin__.set",
- "__builtin__.object",
- "nltk.tokenize.punkt.PunktToken",
- "nltk.tokenize.punkt.PunktLanguageVars",
- "copy_reg._reconstructor",
- "nltk.tokenize.punkt.PunktSentenceTokenizer",
- "__builtin__.int",
- "nltk.tokenize.punkt.PunktParameters",
- "collections.defaultdict",
- "__builtin__.set",
- "__builtin__.object",
- "nltk.tokenize.punkt.PunktToken",
- "nltk.tokenize.punkt.PunktLanguageVars",
- "copy_reg._reconstructor",
- "nltk.tokenize.punkt.PunktSentenceTokenizer",
- "__builtin__.int",
- "nltk.tokenize.punkt.PunktParameters",
- "collections.defaultdict",
- "__builtin__.set",
- "__builtin__.object",
- "nltk.tokenize.punkt.PunktToken",
- "nltk.tokenize.punkt.PunktLanguageVars",
- "copy_reg._reconstructor",
- "nltk.tokenize.punkt.PunktSentenceTokenizer",
- "__builtin__.int",
- "nltk.tokenize.punkt.PunktParameters",
- "collections.defaultdict",
- "nltk.tokenize.punkt.PunktToken",
- "builtins.int",
- "nltk.tokenize.punkt.PunktLanguageVars",
- "nltk.tokenize.punkt.PunktSentenceTokenizer",
- "builtins.set",
- "nltk.tokenize.punkt.PunktParameters",
- "collections.defaultdict",
- "nltk.tokenize.punkt.PunktToken",
- "builtins.int",
- "nltk.tokenize.punkt.PunktLanguageVars",
- "nltk.tokenize.punkt.PunktSentenceTokenizer",
- "builtins.set",
- "nltk.tokenize.punkt.PunktParameters",
- "collections.defaultdict",
- "nltk.tokenize.punkt.PunktToken",
- "builtins.int",
- "nltk.tokenize.punkt.PunktLanguageVars",
- "nltk.tokenize.punkt.PunktSentenceTokenizer",
- "builtins.set",
- "nltk.tokenize.punkt.PunktParameters",
- "collections.defaultdict",
- "nltk.tokenize.punkt.PunktToken",
- "builtins.int",
- "nltk.tokenize.punkt.PunktLanguageVars",
- "nltk.tokenize.punkt.PunktSentenceTokenizer",
- "builtins.set",
- "nltk.tokenize.punkt.PunktParameters",
- "nltk.tokenize.punkt.PunktToken",
- "builtins.int",
- "nltk.tokenize.punkt.PunktParameters",
- "nltk.tokenize.punkt.PunktLanguageVars",
- "nltk.tokenize.punkt.PunktSentenceTokenizer",
- "builtins.set",
- "collections.defaultdict",
- "collections.defaultdict",
- "nltk.tokenize.punkt.PunktToken",
- "builtins.int",
- "nltk.tokenize.punkt.PunktLanguageVars",
- "nltk.tokenize.punkt.PunktSentenceTokenizer",
- "builtins.set",
- "nltk.tokenize.punkt.PunktParameters",
- "collections.defaultdict",
- "nltk.tokenize.punkt.PunktToken",
- "builtins.int",
- "nltk.tokenize.punkt.PunktLanguageVars",
- "nltk.tokenize.punkt.PunktSentenceTokenizer",
- "builtins.set",
- "nltk.tokenize.punkt.PunktParameters",
- "collections.defaultdict",
- "nltk.tokenize.punkt.PunktToken",
- "builtins.int",
- "nltk.tokenize.punkt.PunktLanguageVars",
- "nltk.tokenize.punkt.PunktSentenceTokenizer",
- "builtins.set",
- "nltk.tokenize.punkt.PunktParameters",
- "collections.defaultdict",
- "nltk.tokenize.punkt.PunktToken",
- "builtins.int",
- "nltk.tokenize.punkt.PunktLanguageVars",
- "nltk.tokenize.punkt.PunktSentenceTokenizer",
- "builtins.set",
- "nltk.tokenize.punkt.PunktParameters",
- "collections.defaultdict",
- "nltk.tokenize.punkt.PunktToken",
- "builtins.int",
- "nltk.tokenize.punkt.PunktLanguageVars",
- "nltk.tokenize.punkt.PunktSentenceTokenizer",
- "builtins.set",
- "nltk.tokenize.punkt.PunktParameters",
- "collections.defaultdict",
- "nltk.tokenize.punkt.PunktToken",
- "builtins.int",
- "nltk.tokenize.punkt.PunktLanguageVars",
- "nltk.tokenize.punkt.PunktSentenceTokenizer",
- "builtins.set",
- "nltk.tokenize.punkt.PunktParameters",
- "collections.defaultdict",
- "nltk.tokenize.punkt.PunktToken",
- "builtins.int",
- "nltk.tokenize.punkt.PunktLanguageVars",
- "nltk.tokenize.punkt.PunktSentenceTokenizer",
- "builtins.set",
- "nltk.tokenize.punkt.PunktParameters",
- "collections.defaultdict",
- "nltk.tokenize.punkt.PunktToken",
- "builtins.int",
- "nltk.tokenize.punkt.PunktLanguageVars",
- "nltk.tokenize.punkt.PunktSentenceTokenizer",
- "builtins.set",
- "nltk.tokenize.punkt.PunktParameters",
- "collections.defaultdict",
- "nltk.tokenize.punkt.PunktToken",
- "builtins.int",
- "nltk.tokenize.punkt.PunktLanguageVars",
- "nltk.tokenize.punkt.PunktSentenceTokenizer",
- "builtins.set",
- "nltk.tokenize.punkt.PunktParameters",
- "collections.defaultdict",
- "nltk.tokenize.punkt.PunktToken",
- "builtins.int",
- "nltk.tokenize.punkt.PunktLanguageVars",
- "nltk.tokenize.punkt.PunktSentenceTokenizer",
- "builtins.set",
- "nltk.tokenize.punkt.PunktParameters",
- "collections.defaultdict",
- "nltk.tokenize.punkt.PunktToken",
- "builtins.int",
- "nltk.tokenize.punkt.PunktLanguageVars",
- "nltk.tokenize.punkt.PunktSentenceTokenizer",
- "builtins.set",
- "nltk.tokenize.punkt.PunktParameters",
- "collections.defaultdict",
- "nltk.tokenize.punkt.PunktToken",
- "builtins.int",
- "nltk.tokenize.punkt.PunktLanguageVars",
- "nltk.tokenize.punkt.PunktSentenceTokenizer",
- "builtins.set",
- "nltk.tokenize.punkt.PunktParameters",
- "collections.defaultdict",
- "nltk.tokenize.punkt.PunktToken",
- "builtins.int",
- "nltk.tokenize.punkt.PunktLanguageVars",
- "nltk.tokenize.punkt.PunktSentenceTokenizer",
- "builtins.set",
- "nltk.tokenize.punkt.PunktParameters",
- "__builtin__.set",
- "nltk.tokenize.punkt.PunktToken",
- "nltk.tokenize.punkt.PunktParameters",
- "__builtin__.long",
- "nltk.tokenize.punkt.PunktLanguageVars",
- "nltk.tokenize.punkt.PunktSentenceTokenizer",
- "collections.defaultdict",
- "collections.defaultdict",
- "__builtin__.set",
- "__builtin__.object",
- "nltk.tokenize.punkt.PunktToken",
- "nltk.tokenize.punkt.PunktLanguageVars",
- "copy_reg._reconstructor",
- "nltk.tokenize.punkt.PunktSentenceTokenizer",
- "__builtin__.int",
- "nltk.tokenize.punkt.PunktParameters",
- "collections.defaultdict",
- "__builtin__.set",
- "__builtin__.object",
- "nltk.tokenize.punkt.PunktToken",
- "nltk.tokenize.punkt.PunktLanguageVars",
- "copy_reg._reconstructor",
- "nltk.tokenize.punkt.PunktSentenceTokenizer",
- "__builtin__.int",
- "nltk.tokenize.punkt.PunktParameters",
- "collections.defaultdict",
- "__builtin__.set",
- "__builtin__.object",
- "nltk.tokenize.punkt.PunktToken",
- "nltk.tokenize.punkt.PunktLanguageVars",
- "copy_reg._reconstructor",
- "nltk.tokenize.punkt.PunktSentenceTokenizer",
- "__builtin__.int",
- "nltk.tokenize.punkt.PunktParameters",
- "collections.defaultdict",
- "__builtin__.set",
- "__builtin__.object",
- "nltk.tokenize.punkt.PunktToken",
- "nltk.tokenize.punkt.PunktLanguageVars",
- "copy_reg._reconstructor",
- "nltk.tokenize.punkt.PunktSentenceTokenizer",
- "__builtin__.int",
- "nltk.tokenize.punkt.PunktParameters",
- "collections.defaultdict",
- "__builtin__.set",
- "__builtin__.object",
- "nltk.tokenize.punkt.PunktToken",
- "nltk.tokenize.punkt.PunktLanguageVars",
- "copy_reg._reconstructor",
- "nltk.tokenize.punkt.PunktSentenceTokenizer",
- "__builtin__.int",
- "nltk.tokenize.punkt.PunktParameters",
- "collections.defaultdict",
- "__builtin__.set",
- "__builtin__.object",
- "nltk.tokenize.punkt.PunktToken",
- "nltk.tokenize.punkt.PunktLanguageVars",
- "copy_reg._reconstructor",
- "nltk.tokenize.punkt.PunktSentenceTokenizer",
- "__builtin__.int",
- "nltk.tokenize.punkt.PunktParameters",
- "collections.defaultdict",
- "__builtin__.set",
- "__builtin__.object",
- "nltk.tokenize.punkt.PunktToken",
- "nltk.tokenize.punkt.PunktLanguageVars",
- "copy_reg._reconstructor",
- "nltk.tokenize.punkt.PunktSentenceTokenizer",
- "__builtin__.int",
- "nltk.tokenize.punkt.PunktParameters",
- "collections.defaultdict",
- "__builtin__.set",
- "__builtin__.object",
- "nltk.tokenize.punkt.PunktToken",
- "nltk.tokenize.punkt.PunktLanguageVars",
- "copy_reg._reconstructor",
- "nltk.tokenize.punkt.PunktSentenceTokenizer",
- "__builtin__.int",
- "nltk.tokenize.punkt.PunktParameters",
- "collections.defaultdict",
- "__builtin__.set",
- "__builtin__.object",
- "nltk.tokenize.punkt.PunktToken",
- "nltk.tokenize.punkt.PunktLanguageVars",
- "copy_reg._reconstructor",
- "nltk.tokenize.punkt.PunktSentenceTokenizer",
- "__builtin__.int",
- "nltk.tokenize.punkt.PunktParameters",
- "collections.defaultdict",
- "__builtin__.set",
- "__builtin__.object",
- "nltk.tokenize.punkt.PunktToken",
- "nltk.tokenize.punkt.PunktLanguageVars",
- "copy_reg._reconstructor",
- "nltk.tokenize.punkt.PunktSentenceTokenizer",
- "__builtin__.int",
- "nltk.tokenize.punkt.PunktParameters",
- "collections.defaultdict",
- "__builtin__.set",
- "__builtin__.object",
- "nltk.tokenize.punkt.PunktToken",
- "nltk.tokenize.punkt.PunktLanguageVars",
- "copy_reg._reconstructor",
- "nltk.tokenize.punkt.PunktSentenceTokenizer",
- "__builtin__.int",
- "nltk.tokenize.punkt.PunktParameters",
- "collections.defaultdict",
- "__builtin__.set",
- "__builtin__.object",
- "nltk.tokenize.punkt.PunktToken",
- "nltk.tokenize.punkt.PunktLanguageVars",
- "copy_reg._reconstructor",
- "nltk.tokenize.punkt.PunktSentenceTokenizer",
- "__builtin__.int",
- "nltk.tokenize.punkt.PunktParameters",
- "collections.defaultdict",
- "__builtin__.set",
- "__builtin__.object",
- "nltk.tokenize.punkt.PunktToken",
- "nltk.tokenize.punkt.PunktLanguageVars",
- "copy_reg._reconstructor",
- "nltk.tokenize.punkt.PunktSentenceTokenizer",
- "__builtin__.int",
- "nltk.tokenize.punkt.PunktParameters",
- "collections.defaultdict",
- "__builtin__.set",
- "nltk.tokenize.punkt.PunktToken",
- "nltk.tokenize.punkt.PunktLanguageVars",
- "nltk.tokenize.punkt.PunktSentenceTokenizer",
- "__builtin__.int",
- "nltk.tokenize.punkt.PunktParameters",
- "collections.defaultdict",
- "__builtin__.set",
- "nltk.tokenize.punkt.PunktToken",
- "nltk.tokenize.punkt.PunktLanguageVars",
- "nltk.tokenize.punkt.PunktSentenceTokenizer",
- "__builtin__.int",
- "nltk.tokenize.punkt.PunktParameters"
How to fix it?
13.9 MB
Upload punkt.zip