sunnychenxiwang commited on
Commit
55d9644
1 Parent(s): bd1679c
This view is limited to 50 files because it contains too many changes.   See raw diff
Files changed (50) hide show
  1. app.py +14 -43
  2. models/dbnetpp.pth +3 -0
  3. models/groundingdino_swint_ogc.pth +3 -0
  4. models/maerec_b.pth +3 -0
  5. pipeline/__pycache__/judge.cpython-39.pyc +0 -0
  6. pipeline/__pycache__/query_generate.cpython-39.pyc +0 -0
  7. pipeline/__pycache__/run_pipeline.cpython-39.pyc +0 -0
  8. pipeline/__pycache__/tool_execute.cpython-39.pyc +0 -0
  9. pipeline/cache_files/COCO_val2014_000000297425.jpg +0 -0
  10. pipeline/cache_files/COCO_val2014_000000297425/2Axrpnw7ricrqoNoRqLPPm.jpg +0 -0
  11. pipeline/cache_files/COCO_val2014_000000297425/N7r9ReNBe73cY2VL7gW9Az.jpg +0 -0
  12. pipeline/cache_files/COCO_val2014_000000297425/Wr7HRVVj6HtMGbsDbaXdU3.jpg +0 -0
  13. pipeline/cache_files/COCO_val2014_000000297425/max6iKdGyovWLMJzKQ4RHr.jpg +0 -0
  14. pipeline/judge.py +15 -11
  15. pipeline/nltk/VERSION +0 -1
  16. pipeline/nltk/__init__.py +0 -209
  17. pipeline/nltk/__pycache__/__init__.cpython-39.pyc +0 -0
  18. pipeline/nltk/__pycache__/book.cpython-39.pyc +0 -0
  19. pipeline/nltk/__pycache__/cli.cpython-39.pyc +0 -0
  20. pipeline/nltk/__pycache__/collections.cpython-39.pyc +0 -0
  21. pipeline/nltk/__pycache__/collocations.cpython-39.pyc +0 -0
  22. pipeline/nltk/__pycache__/compat.cpython-39.pyc +0 -0
  23. pipeline/nltk/__pycache__/data.cpython-39.pyc +0 -0
  24. pipeline/nltk/__pycache__/decorators.cpython-39.pyc +0 -0
  25. pipeline/nltk/__pycache__/downloader.cpython-39.pyc +0 -0
  26. pipeline/nltk/__pycache__/featstruct.cpython-39.pyc +0 -0
  27. pipeline/nltk/__pycache__/grammar.cpython-39.pyc +0 -0
  28. pipeline/nltk/__pycache__/help.cpython-39.pyc +0 -0
  29. pipeline/nltk/__pycache__/internals.cpython-39.pyc +0 -0
  30. pipeline/nltk/__pycache__/jsontags.cpython-39.pyc +0 -0
  31. pipeline/nltk/__pycache__/langnames.cpython-39.pyc +0 -0
  32. pipeline/nltk/__pycache__/lazyimport.cpython-39.pyc +0 -0
  33. pipeline/nltk/__pycache__/probability.cpython-39.pyc +0 -0
  34. pipeline/nltk/__pycache__/text.cpython-39.pyc +0 -0
  35. pipeline/nltk/__pycache__/tgrep.cpython-39.pyc +0 -0
  36. pipeline/nltk/__pycache__/toolbox.cpython-39.pyc +0 -0
  37. pipeline/nltk/__pycache__/treeprettyprinter.cpython-39.pyc +0 -0
  38. pipeline/nltk/__pycache__/treetransforms.cpython-39.pyc +0 -0
  39. pipeline/nltk/__pycache__/util.cpython-39.pyc +0 -0
  40. pipeline/nltk/__pycache__/wsd.cpython-39.pyc +0 -0
  41. pipeline/nltk/app/__init__.py +0 -47
  42. pipeline/nltk/app/__pycache__/__init__.cpython-39.pyc +0 -0
  43. pipeline/nltk/app/__pycache__/chartparser_app.cpython-39.pyc +0 -0
  44. pipeline/nltk/app/__pycache__/chunkparser_app.cpython-39.pyc +0 -0
  45. pipeline/nltk/app/__pycache__/collocations_app.cpython-39.pyc +0 -0
  46. pipeline/nltk/app/__pycache__/concordance_app.cpython-39.pyc +0 -0
  47. pipeline/nltk/app/__pycache__/nemo_app.cpython-39.pyc +0 -0
  48. pipeline/nltk/app/__pycache__/rdparser_app.cpython-39.pyc +0 -0
  49. pipeline/nltk/app/__pycache__/srparser_app.cpython-39.pyc +0 -0
  50. pipeline/nltk/app/__pycache__/wordfreq_app.cpython-39.pyc +0 -0
app.py CHANGED
@@ -1,58 +1,29 @@
1
 
2
  import gradio as gr
3
  from pipeline.run_pipeline import *
 
 
 
 
 
4
  # from run import *
5
  # '''
6
  # 把一些文件移动到此文件路径下
7
  # '''
8
  # text = "A person is cutting a birthday cake with two red candles that spell out \"21\". The surface of the cake is round, and there is a balloon in the room. The person is using a silver knife to cut the cake."
9
  # image_path = "/newdisk3/wcx/val2014/COCO_val2014_000000297425.jpg"
 
 
 
10
 
11
- def get_response(text, filepath):
12
- pipeline = Pipeline()
13
- res = pipeline.run(text=text, image_path=image_path)
14
- return res
15
 
16
  demo = gr.Interface(
17
  fn=get_response,
18
- inputs=["text", gr.Image(type="filepath")],
19
- outputs=["text"],
20
  )
21
 
22
- demo.launch()
23
-
24
- # def generate_mutimodal(title, context, img):
25
- # return f"Title:{title}\nContext:{context}\n...{img}"
26
-
27
- # server = gr.Interface(
28
- # fn=generate_mutimodal,
29
- # inputs=[
30
- # gr.Textbox(lines=1, placeholder="请输入标题"),
31
- # gr.Textbox(lines=2, placeholder="请输入正文"),
32
- # gr.Image(shape=(200, 200), label="请上传图片(可选)")
33
- # ],
34
- # outputs="text"
35
- # )
36
-
37
- # server.launch()
38
-
39
- # import numpy as np
40
- # import gradio as gr
41
- # def sepia(input_img):
42
- # #处理图像
43
- # sepia_filter = np.array([
44
- # [0.393, 0.769, 0.189],
45
- # [0.349, 0.686, 0.168],
46
- # [0.272, 0.534, 0.131]
47
- # ])
48
- # sepia_img = input_img.dot(sepia_filter.T)
49
- # sepia_img /= sepia_img.max()
50
- # return sepia_img
51
- # #shape设置输入图像大小
52
- # demo = gr.Interface(sepia, gr.Image(), "image")
53
- # demo.launch()
54
-
55
- # Download human-readable labels for ImageNet.
56
-
57
- # gr.Interface(fn=sepia,inputs=gr.Image(type="pil"),outputs="image").launch()
58
-
 
1
 
2
  import gradio as gr
3
  from pipeline.run_pipeline import *
4
+ '''
5
+ 时间优化
6
+ 并发优化
7
+ '''
8
+
9
  # from run import *
10
  # '''
11
  # 把一些文件移动到此文件路径下
12
  # '''
13
  # text = "A person is cutting a birthday cake with two red candles that spell out \"21\". The surface of the cake is round, and there is a balloon in the room. The person is using a silver knife to cut the cake."
14
  # image_path = "/newdisk3/wcx/val2014/COCO_val2014_000000297425.jpg"
15
+ pipeline = Pipeline(type="image-to-text", api_key="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")
16
+ # res,claim_list = pipeline.run(text=text, image_path=image_path,type="image-to-text")
17
+ # print(res)
18
 
19
+ def get_response(text, filepath, type):
20
+ res, claim_list = pipeline.run(text=text, image_path=filepath, type=type)
21
+ return claim_list, res
 
22
 
23
  demo = gr.Interface(
24
  fn=get_response,
25
+ inputs=[gr.Textbox(placeholder="Input I2T model's response or T2I model's prompt", label="text input"), gr.Image(type="filepath", label="image input"), gr.Radio(['image-to-text','text-to-image'], label='task type', value='image-to-text')],
26
+ outputs=[gr.Textbox(label="claim list"), gr.Textbox(label="detect results")],
27
  )
28
 
29
+ demo.queue().launch(share=True)
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
models/dbnetpp.pth ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:4ecb39ac54bd954fa44fd25b419b7a51539ded0f79408eb2419ca78f9f0299c6
3
+ size 113299489
models/groundingdino_swint_ogc.pth ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:3b3ca2563c77c69f651d7bd133e97139c186df06231157a64c507099c52bc799
3
+ size 693997677
models/maerec_b.pth ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:69a98230f5421517bce9fdc197261ec6845e353ceb7fa970194c04403e0ba89b
3
+ size 1764777173
pipeline/__pycache__/judge.cpython-39.pyc CHANGED
Binary files a/pipeline/__pycache__/judge.cpython-39.pyc and b/pipeline/__pycache__/judge.cpython-39.pyc differ
 
pipeline/__pycache__/query_generate.cpython-39.pyc CHANGED
Binary files a/pipeline/__pycache__/query_generate.cpython-39.pyc and b/pipeline/__pycache__/query_generate.cpython-39.pyc differ
 
pipeline/__pycache__/run_pipeline.cpython-39.pyc CHANGED
Binary files a/pipeline/__pycache__/run_pipeline.cpython-39.pyc and b/pipeline/__pycache__/run_pipeline.cpython-39.pyc differ
 
pipeline/__pycache__/tool_execute.cpython-39.pyc CHANGED
Binary files a/pipeline/__pycache__/tool_execute.cpython-39.pyc and b/pipeline/__pycache__/tool_execute.cpython-39.pyc differ
 
pipeline/cache_files/COCO_val2014_000000297425.jpg ADDED
pipeline/cache_files/COCO_val2014_000000297425/2Axrpnw7ricrqoNoRqLPPm.jpg ADDED
pipeline/cache_files/COCO_val2014_000000297425/N7r9ReNBe73cY2VL7gW9Az.jpg ADDED
pipeline/cache_files/COCO_val2014_000000297425/Wr7HRVVj6HtMGbsDbaXdU3.jpg ADDED
pipeline/cache_files/COCO_val2014_000000297425/max6iKdGyovWLMJzKQ4RHr.jpg ADDED
pipeline/judge.py CHANGED
@@ -2,9 +2,9 @@ import json
2
  import yaml
3
  import base64
4
  class Judger:
5
- def __init__(self, prompt_path, chat, type):
6
  with open(prompt_path,"r",encoding='utf-8') as file:
7
- self.prompt = yaml.load(file, yaml.FullLoader)[type]
8
  self.chat = chat
9
 
10
 
@@ -12,7 +12,7 @@ class Judger:
12
  with open(image_path, "rb") as image_file:
13
  return base64.b64encode(image_file.read()).decode('utf-8')
14
 
15
- def get_response(self, object_res, attribue_res, text_res, fact_res, claim_list, image_path):
16
  input = '''
17
  Here is the object detection expert model's result:
18
  {object}
@@ -40,24 +40,28 @@ class Judger:
40
  else:
41
  text_det = "none information"
42
 
43
-
44
- img1 = self.encode_image("/home/wcx/wcx/GroundingDINO/LVLM/cot/img_examples/animal.jpg")
45
- img2 = self.encode_image("/home/wcx/wcx/GroundingDINO/LVLM/cot/img_examples/ball.jpg")
 
 
 
46
  base64_source_image = self.encode_image(image_path)
47
  content = [
48
- {"type": "text", "text": self.prompt["user"]},
49
  {"type": "image_url","image_url": f"data:image/jpeg;base64,{img1}"},
50
- {"type": "text", "text": self.prompt["example1"]},
51
  {"type": "image_url","image_url": f"data:image/jpeg;base64,{img2}"},
52
- {"type": "text", "text": self.prompt["example2"]},
53
  {"type": "image_url","image_url": f"data:image/jpeg;base64,{base64_source_image}"},
54
  {"type": "text", "text": input.format(object=object_det,text=text_det,fact=fact_res,claims=claim_list)}
55
  ]
56
-
 
57
  message = [
58
  {
59
  'role': 'system',
60
- 'content': self.prompt["system"]
61
  },
62
  {
63
  "role": "user",
 
2
  import yaml
3
  import base64
4
  class Judger:
5
+ def __init__(self, prompt_path, chat):
6
  with open(prompt_path,"r",encoding='utf-8') as file:
7
+ self.prompt = yaml.load(file, yaml.FullLoader)
8
  self.chat = chat
9
 
10
 
 
12
  with open(image_path, "rb") as image_file:
13
  return base64.b64encode(image_file.read()).decode('utf-8')
14
 
15
+ def get_response(self, type, object_res, attribue_res, text_res, fact_res, claim_list, image_path):
16
  input = '''
17
  Here is the object detection expert model's result:
18
  {object}
 
40
  else:
41
  text_det = "none information"
42
 
43
+ if type == "image-to-text":
44
+ img1 = self.encode_image("/home/wcx/wcx/GroundingDINO/LVLM/cot/img_examples/sandbeach.jpg")
45
+ img2 = self.encode_image("/home/wcx/wcx/GroundingDINO/LVLM/cot/img_examples/football.jpg")
46
+ else:
47
+ img1 = self.encode_image("/home/wcx/wcx/GroundingDINO/LVLM/cot/img_examples/animal.jpg")
48
+ img2 = self.encode_image("/home/wcx/wcx/GroundingDINO/LVLM/cot/img_examples/ball.jpg")
49
  base64_source_image = self.encode_image(image_path)
50
  content = [
51
+ {"type": "text", "text": self.prompt[type]["user"]},
52
  {"type": "image_url","image_url": f"data:image/jpeg;base64,{img1}"},
53
+ {"type": "text", "text": self.prompt[type]["example1"]},
54
  {"type": "image_url","image_url": f"data:image/jpeg;base64,{img2}"},
55
+ {"type": "text", "text": self.prompt[type]["example2"]},
56
  {"type": "image_url","image_url": f"data:image/jpeg;base64,{base64_source_image}"},
57
  {"type": "text", "text": input.format(object=object_det,text=text_det,fact=fact_res,claims=claim_list)}
58
  ]
59
+
60
+
61
  message = [
62
  {
63
  'role': 'system',
64
+ 'content': self.prompt[type]["system"]
65
  },
66
  {
67
  "role": "user",
pipeline/nltk/VERSION DELETED
@@ -1 +0,0 @@
1
- 3.8.1
 
 
pipeline/nltk/__init__.py DELETED
@@ -1,209 +0,0 @@
1
- # Natural Language Toolkit (NLTK)
2
- #
3
- # Copyright (C) 2001-2023 NLTK Project
4
- # Authors: Steven Bird <stevenbird1@gmail.com>
5
- # Edward Loper <edloper@gmail.com>
6
- # URL: <https://www.nltk.org/>
7
- # For license information, see LICENSE.TXT
8
-
9
- """
10
- The Natural Language Toolkit (NLTK) is an open source Python library
11
- for Natural Language Processing. A free online book is available.
12
- (If you use the library for academic research, please cite the book.)
13
-
14
- Steven Bird, Ewan Klein, and Edward Loper (2009).
15
- Natural Language Processing with Python. O'Reilly Media Inc.
16
- https://www.nltk.org/book/
17
-
18
- isort:skip_file
19
- """
20
-
21
- import os
22
-
23
- # //////////////////////////////////////////////////////
24
- # Metadata
25
- # //////////////////////////////////////////////////////
26
-
27
- # Version. For each new release, the version number should be updated
28
- # in the file VERSION.
29
- try:
30
- # If a VERSION file exists, use it!
31
- version_file = os.path.join(os.path.dirname(__file__), "VERSION")
32
- with open(version_file) as infile:
33
- __version__ = infile.read().strip()
34
- except NameError:
35
- __version__ = "unknown (running code interactively?)"
36
- except OSError as ex:
37
- __version__ = "unknown (%s)" % ex
38
-
39
- if __doc__ is not None: # fix for the ``python -OO``
40
- __doc__ += "\n@version: " + __version__
41
-
42
-
43
- # Copyright notice
44
- __copyright__ = """\
45
- Copyright (C) 2001-2023 NLTK Project.
46
-
47
- Distributed and Licensed under the Apache License, Version 2.0,
48
- which is included by reference.
49
- """
50
-
51
- __license__ = "Apache License, Version 2.0"
52
- # Description of the toolkit, keywords, and the project's primary URL.
53
- __longdescr__ = """\
54
- The Natural Language Toolkit (NLTK) is a Python package for
55
- natural language processing. NLTK requires Python 3.7, 3.8, 3.9, 3.10 or 3.11."""
56
- __keywords__ = [
57
- "NLP",
58
- "CL",
59
- "natural language processing",
60
- "computational linguistics",
61
- "parsing",
62
- "tagging",
63
- "tokenizing",
64
- "syntax",
65
- "linguistics",
66
- "language",
67
- "natural language",
68
- "text analytics",
69
- ]
70
- __url__ = "https://www.nltk.org/"
71
-
72
- # Maintainer, contributors, etc.
73
- __maintainer__ = "NLTK Team"
74
- __maintainer_email__ = "nltk.team@gmail.com"
75
- __author__ = __maintainer__
76
- __author_email__ = __maintainer_email__
77
-
78
- # "Trove" classifiers for Python Package Index.
79
- __classifiers__ = [
80
- "Development Status :: 5 - Production/Stable",
81
- "Intended Audience :: Developers",
82
- "Intended Audience :: Education",
83
- "Intended Audience :: Information Technology",
84
- "Intended Audience :: Science/Research",
85
- "License :: OSI Approved :: Apache Software License",
86
- "Operating System :: OS Independent",
87
- "Programming Language :: Python :: 3.7",
88
- "Programming Language :: Python :: 3.8",
89
- "Programming Language :: Python :: 3.9",
90
- "Programming Language :: Python :: 3.10",
91
- "Programming Language :: Python :: 3.11",
92
- "Topic :: Scientific/Engineering",
93
- "Topic :: Scientific/Engineering :: Artificial Intelligence",
94
- "Topic :: Scientific/Engineering :: Human Machine Interfaces",
95
- "Topic :: Scientific/Engineering :: Information Analysis",
96
- "Topic :: Text Processing",
97
- "Topic :: Text Processing :: Filters",
98
- "Topic :: Text Processing :: General",
99
- "Topic :: Text Processing :: Indexing",
100
- "Topic :: Text Processing :: Linguistic",
101
- ]
102
-
103
- from nltk.internals import config_java
104
-
105
- # support numpy from pypy
106
- try:
107
- import numpypy
108
- except ImportError:
109
- pass
110
-
111
- # Override missing methods on environments where it cannot be used like GAE.
112
- import subprocess
113
-
114
- if not hasattr(subprocess, "PIPE"):
115
-
116
- def _fake_PIPE(*args, **kwargs):
117
- raise NotImplementedError("subprocess.PIPE is not supported.")
118
-
119
- subprocess.PIPE = _fake_PIPE
120
- if not hasattr(subprocess, "Popen"):
121
-
122
- def _fake_Popen(*args, **kwargs):
123
- raise NotImplementedError("subprocess.Popen is not supported.")
124
-
125
- subprocess.Popen = _fake_Popen
126
-
127
- ###########################################################
128
- # TOP-LEVEL MODULES
129
- ###########################################################
130
-
131
- # Import top-level functionality into top-level namespace
132
-
133
- from nltk.collocations import *
134
- from nltk.decorators import decorator, memoize
135
- from nltk.featstruct import *
136
- from nltk.grammar import *
137
- from nltk.probability import *
138
- from nltk.text import *
139
- from nltk.util import *
140
- from nltk.jsontags import *
141
-
142
- ###########################################################
143
- # PACKAGES
144
- ###########################################################
145
-
146
- from nltk.chunk import *
147
- from nltk.classify import *
148
- from nltk.inference import *
149
- from nltk.metrics import *
150
- from nltk.parse import *
151
- from nltk.tag import *
152
- from nltk.tokenize import *
153
- from nltk.translate import *
154
- from nltk.tree import *
155
- from nltk.sem import *
156
- from nltk.stem import *
157
-
158
- # Packages which can be lazily imported
159
- # (a) we don't import *
160
- # (b) they're slow to import or have run-time dependencies
161
- # that can safely fail at run time
162
-
163
- from nltk import lazyimport
164
-
165
- app = lazyimport.LazyModule("app", locals(), globals())
166
- chat = lazyimport.LazyModule("chat", locals(), globals())
167
- corpus = lazyimport.LazyModule("corpus", locals(), globals())
168
- draw = lazyimport.LazyModule("draw", locals(), globals())
169
- toolbox = lazyimport.LazyModule("toolbox", locals(), globals())
170
-
171
- # Optional loading
172
-
173
- try:
174
- import numpy
175
- except ImportError:
176
- pass
177
- else:
178
- from nltk import cluster
179
-
180
- from nltk.downloader import download, download_shell
181
-
182
- try:
183
- import tkinter
184
- except ImportError:
185
- pass
186
- else:
187
- try:
188
- from nltk.downloader import download_gui
189
- except RuntimeError as e:
190
- import warnings
191
-
192
- warnings.warn(
193
- "Corpus downloader GUI not loaded "
194
- "(RuntimeError during import: %s)" % str(e)
195
- )
196
-
197
- # explicitly import all top-level modules (ensuring
198
- # they override the same names inadvertently imported
199
- # from a subpackage)
200
-
201
- from nltk import ccg, chunk, classify, collocations
202
- from nltk import data, featstruct, grammar, help, inference, metrics
203
- from nltk import misc, parse, probability, sem, stem, wsd
204
- from nltk import tag, tbl, text, tokenize, translate, tree, util
205
-
206
-
207
- # FIXME: override any accidentally imported demo, see https://github.com/nltk/nltk/issues/2116
208
- def demo():
209
- print("To run the demo code for a module, type nltk.module.demo()")
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
pipeline/nltk/__pycache__/__init__.cpython-39.pyc DELETED
Binary file (4.84 kB)
 
pipeline/nltk/__pycache__/book.cpython-39.pyc DELETED
Binary file (2.99 kB)
 
pipeline/nltk/__pycache__/cli.cpython-39.pyc DELETED
Binary file (1.63 kB)
 
pipeline/nltk/__pycache__/collections.cpython-39.pyc DELETED
Binary file (23.4 kB)
 
pipeline/nltk/__pycache__/collocations.cpython-39.pyc DELETED
Binary file (14.9 kB)
 
pipeline/nltk/__pycache__/compat.cpython-39.pyc DELETED
Binary file (1.13 kB)
 
pipeline/nltk/__pycache__/data.cpython-39.pyc DELETED
Binary file (38.6 kB)
 
pipeline/nltk/__pycache__/decorators.cpython-39.pyc DELETED
Binary file (6.43 kB)
 
pipeline/nltk/__pycache__/downloader.cpython-39.pyc DELETED
Binary file (61.9 kB)
 
pipeline/nltk/__pycache__/featstruct.cpython-39.pyc DELETED
Binary file (74.1 kB)
 
pipeline/nltk/__pycache__/grammar.cpython-39.pyc DELETED
Binary file (53.7 kB)
 
pipeline/nltk/__pycache__/help.cpython-39.pyc DELETED
Binary file (1.63 kB)
 
pipeline/nltk/__pycache__/internals.cpython-39.pyc DELETED
Binary file (29 kB)
 
pipeline/nltk/__pycache__/jsontags.cpython-39.pyc DELETED
Binary file (2.31 kB)
 
pipeline/nltk/__pycache__/langnames.cpython-39.pyc DELETED
Binary file (15.3 kB)
 
pipeline/nltk/__pycache__/lazyimport.cpython-39.pyc DELETED
Binary file (3.73 kB)
 
pipeline/nltk/__pycache__/probability.cpython-39.pyc DELETED
Binary file (87.2 kB)
 
pipeline/nltk/__pycache__/text.cpython-39.pyc DELETED
Binary file (28.4 kB)
 
pipeline/nltk/__pycache__/tgrep.cpython-39.pyc DELETED
Binary file (35.3 kB)
 
pipeline/nltk/__pycache__/toolbox.cpython-39.pyc DELETED
Binary file (15.8 kB)
 
pipeline/nltk/__pycache__/treeprettyprinter.cpython-39.pyc DELETED
Binary file (952 Bytes)
 
pipeline/nltk/__pycache__/treetransforms.cpython-39.pyc DELETED
Binary file (4.99 kB)
 
pipeline/nltk/__pycache__/util.cpython-39.pyc DELETED
Binary file (32.5 kB)
 
pipeline/nltk/__pycache__/wsd.cpython-39.pyc DELETED
Binary file (1.82 kB)
 
pipeline/nltk/app/__init__.py DELETED
@@ -1,47 +0,0 @@
1
- # Natural Language Toolkit: Applications package
2
- #
3
- # Copyright (C) 2001-2023 NLTK Project
4
- # Author: Edward Loper <edloper@gmail.com>
5
- # Steven Bird <stevenbird1@gmail.com>
6
- # URL: <https://www.nltk.org/>
7
- # For license information, see LICENSE.TXT
8
-
9
- """
10
- Interactive NLTK Applications:
11
-
12
- chartparser: Chart Parser
13
- chunkparser: Regular-Expression Chunk Parser
14
- collocations: Find collocations in text
15
- concordance: Part-of-speech concordancer
16
- nemo: Finding (and Replacing) Nemo regular expression tool
17
- rdparser: Recursive Descent Parser
18
- srparser: Shift-Reduce Parser
19
- wordnet: WordNet Browser
20
- """
21
-
22
-
23
- # Import Tkinter-based modules if Tkinter is installed
24
- try:
25
- import tkinter
26
- except ImportError:
27
- import warnings
28
-
29
- warnings.warn("nltk.app package not loaded (please install Tkinter library).")
30
- else:
31
- from nltk.app.chartparser_app import app as chartparser
32
- from nltk.app.chunkparser_app import app as chunkparser
33
- from nltk.app.collocations_app import app as collocations
34
- from nltk.app.concordance_app import app as concordance
35
- from nltk.app.nemo_app import app as nemo
36
- from nltk.app.rdparser_app import app as rdparser
37
- from nltk.app.srparser_app import app as srparser
38
- from nltk.app.wordnet_app import app as wordnet
39
-
40
- try:
41
- from matplotlib import pylab
42
- except ImportError:
43
- import warnings
44
-
45
- warnings.warn("nltk.app.wordfreq not loaded (requires the matplotlib library).")
46
- else:
47
- from nltk.app.wordfreq_app import app as wordfreq
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
pipeline/nltk/app/__pycache__/__init__.cpython-39.pyc DELETED
Binary file (1.31 kB)
 
pipeline/nltk/app/__pycache__/chartparser_app.cpython-39.pyc DELETED
Binary file (63.2 kB)
 
pipeline/nltk/app/__pycache__/chunkparser_app.cpython-39.pyc DELETED
Binary file (33.4 kB)
 
pipeline/nltk/app/__pycache__/collocations_app.cpython-39.pyc DELETED
Binary file (14.8 kB)
 
pipeline/nltk/app/__pycache__/concordance_app.cpython-39.pyc DELETED
Binary file (22.7 kB)
 
pipeline/nltk/app/__pycache__/nemo_app.cpython-39.pyc DELETED
Binary file (12.3 kB)
 
pipeline/nltk/app/__pycache__/rdparser_app.cpython-39.pyc DELETED
Binary file (26 kB)
 
pipeline/nltk/app/__pycache__/srparser_app.cpython-39.pyc DELETED
Binary file (22 kB)
 
pipeline/nltk/app/__pycache__/wordfreq_app.cpython-39.pyc DELETED
Binary file (1.46 kB)