fix: add inference
Browse files- .gitignore +6 -1
- app.py +0 -1
- gradio_web_server.log +185 -0
- gradio_web_server_multi.log +4 -0
- src/__pycache__/__init__.cpython-310.pyc +0 -0
- src/__pycache__/constants.cpython-310.pyc +0 -0
- src/__pycache__/conversation.cpython-310.pyc +0 -0
- src/__pycache__/utils.cpython-310.pyc +0 -0
- src/model/__pycache__/__init__.cpython-310.pyc +0 -0
- src/model/__pycache__/compression.cpython-310.pyc +0 -0
- src/model/__pycache__/llama_condense_monkey_patch.cpython-310.pyc +0 -0
- src/model/__pycache__/model_adapter.cpython-310.pyc +0 -0
- src/model/__pycache__/model_chatglm.cpython-310.pyc +0 -0
- src/model/__pycache__/model_cllm.cpython-310.pyc +0 -0
- src/model/__pycache__/model_codet5p.cpython-310.pyc +0 -0
- src/model/__pycache__/model_exllama.cpython-310.pyc +0 -0
- src/model/__pycache__/model_falcon.cpython-310.pyc +0 -0
- src/model/__pycache__/model_registry.cpython-310.pyc +0 -0
- src/model/__pycache__/model_xfastertransformer.cpython-310.pyc +0 -0
- src/model/__pycache__/model_yuan2.cpython-310.pyc +0 -0
- src/model/__pycache__/monkey_patch_non_inplace.cpython-310.pyc +0 -0
- src/modules/__pycache__/__init__.cpython-310.pyc +0 -0
- src/modules/__pycache__/awq.cpython-310.pyc +0 -0
- src/modules/__pycache__/exllama.cpython-310.pyc +0 -0
- src/modules/__pycache__/gptq.cpython-310.pyc +0 -0
- src/modules/__pycache__/xfastertransformer.cpython-310.pyc +0 -0
- src/serve/__pycache__/__init__.cpython-310.pyc +0 -0
- src/serve/__pycache__/api_provider.cpython-310.pyc +0 -0
- src/serve/__pycache__/gradio_block_arena_named.cpython-310.pyc +0 -0
- src/serve/__pycache__/gradio_block_arena_vision.cpython-310.pyc +0 -0
- src/serve/__pycache__/gradio_block_arena_vision_named.cpython-310.pyc +0 -0
- src/serve/__pycache__/gradio_web_server.cpython-310.pyc +0 -0
- src/serve/__pycache__/remote_logger.cpython-310.pyc +0 -0
- src/serve/gradio_web_server.py +6 -2
- vision-tmp-2024-07-10-conv.json +3 -0
.gitignore
CHANGED
@@ -1 +1,6 @@
|
|
1 |
-
dependency
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
dependency
|
2 |
+
*.json
|
3 |
+
*.log
|
4 |
+
*.pyc
|
5 |
+
*__pycache__*
|
6 |
+
*.json
|
app.py
CHANGED
@@ -6,7 +6,6 @@ import gradio as gr
|
|
6 |
from src.model.model_llava import inference
|
7 |
from src.serve.gradio_block_arena_vision_named import build_side_by_side_vision_ui_named
|
8 |
|
9 |
-
@spaces.GPU(duration=120)
|
10 |
def main():
|
11 |
with gr.Blocks() as demo:
|
12 |
states = build_side_by_side_vision_ui_named(
|
|
|
6 |
from src.model.model_llava import inference
|
7 |
from src.serve.gradio_block_arena_vision_named import build_side_by_side_vision_ui_named
|
8 |
|
|
|
9 |
def main():
|
10 |
with gr.Blocks() as demo:
|
11 |
states = build_side_by_side_vision_ui_named(
|
gradio_web_server.log
CHANGED
@@ -254,3 +254,188 @@
|
|
254 |
2024-07-10 12:01:17 | ERROR | stderr | File "/home/bofei-zhang/miniconda3/envs/fire_demo/lib/python3.10/threading.py", line 1116, in _wait_for_tstate_lock
|
255 |
2024-07-10 12:01:17 | ERROR | stderr | if lock.acquire(block, timeout):
|
256 |
2024-07-10 12:01:17 | ERROR | stderr | KeyboardInterrupt
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
254 |
2024-07-10 12:01:17 | ERROR | stderr | File "/home/bofei-zhang/miniconda3/envs/fire_demo/lib/python3.10/threading.py", line 1116, in _wait_for_tstate_lock
|
255 |
2024-07-10 12:01:17 | ERROR | stderr | if lock.acquire(block, timeout):
|
256 |
2024-07-10 12:01:17 | ERROR | stderr | KeyboardInterrupt
|
257 |
+
2024-07-10 07:18:04 | ERROR | stderr | /usr/local/lib/python3.10/site-packages/spaces/zero/wrappers.py:77: UserWarning: Using a ZeroGPU function outside of Gradio caching or request might block the app
|
258 |
+
2024-07-10 07:18:04 | ERROR | stderr | warnings.warn("Using a ZeroGPU function outside of Gradio caching or request might block the app")
|
259 |
+
2024-07-10 07:18:11 | INFO | stdout | Running on local URL: http://0.0.0.0:7860
|
260 |
+
2024-07-10 07:18:11 | INFO | stdout |
|
261 |
+
2024-07-10 07:18:11 | INFO | stdout | To create a public link, set `share=True` in `launch()`.
|
262 |
+
2024-07-10 07:18:28 | ERROR | stderr | ERROR: Exception in ASGI application
|
263 |
+
2024-07-10 07:18:28 | ERROR | stderr | Traceback (most recent call last):
|
264 |
+
2024-07-10 07:18:28 | ERROR | stderr | File "/usr/local/lib/python3.10/site-packages/starlette/responses.py", line 265, in __call__
|
265 |
+
2024-07-10 07:18:28 | ERROR | stderr | await wrap(partial(self.listen_for_disconnect, receive))
|
266 |
+
2024-07-10 07:18:28 | ERROR | stderr | File "/usr/local/lib/python3.10/site-packages/starlette/responses.py", line 261, in wrap
|
267 |
+
2024-07-10 07:18:28 | ERROR | stderr | await func()
|
268 |
+
2024-07-10 07:18:28 | ERROR | stderr | File "/usr/local/lib/python3.10/site-packages/starlette/responses.py", line 238, in listen_for_disconnect
|
269 |
+
2024-07-10 07:18:28 | ERROR | stderr | message = await receive()
|
270 |
+
2024-07-10 07:18:28 | ERROR | stderr | File "/usr/local/lib/python3.10/site-packages/uvicorn/protocols/http/httptools_impl.py", line 553, in receive
|
271 |
+
2024-07-10 07:18:28 | ERROR | stderr | await self.message_event.wait()
|
272 |
+
2024-07-10 07:18:28 | ERROR | stderr | File "/usr/local/lib/python3.10/asyncio/locks.py", line 214, in wait
|
273 |
+
2024-07-10 07:18:28 | ERROR | stderr | await fut
|
274 |
+
2024-07-10 07:18:28 | ERROR | stderr | asyncio.exceptions.CancelledError: Cancelled by cancel scope 7ffa672fb370
|
275 |
+
2024-07-10 07:18:28 | ERROR | stderr |
|
276 |
+
2024-07-10 07:18:28 | ERROR | stderr | During handling of the above exception, another exception occurred:
|
277 |
+
2024-07-10 07:18:28 | ERROR | stderr |
|
278 |
+
2024-07-10 07:18:28 | ERROR | stderr | + Exception Group Traceback (most recent call last):
|
279 |
+
2024-07-10 07:18:28 | ERROR | stderr | | File "/usr/local/lib/python3.10/site-packages/uvicorn/protocols/http/httptools_impl.py", line 399, in run_asgi
|
280 |
+
2024-07-10 07:18:28 | ERROR | stderr | | result = await app( # type: ignore[func-returns-value]
|
281 |
+
2024-07-10 07:18:28 | ERROR | stderr | | File "/usr/local/lib/python3.10/site-packages/uvicorn/middleware/proxy_headers.py", line 70, in __call__
|
282 |
+
2024-07-10 07:18:28 | ERROR | stderr | | return await self.app(scope, receive, send)
|
283 |
+
2024-07-10 07:18:28 | ERROR | stderr | | File "/usr/local/lib/python3.10/site-packages/fastapi/applications.py", line 1054, in __call__
|
284 |
+
2024-07-10 07:18:28 | ERROR | stderr | | await super().__call__(scope, receive, send)
|
285 |
+
2024-07-10 07:18:28 | ERROR | stderr | | File "/usr/local/lib/python3.10/site-packages/starlette/applications.py", line 123, in __call__
|
286 |
+
2024-07-10 07:18:28 | ERROR | stderr | | await self.middleware_stack(scope, receive, send)
|
287 |
+
2024-07-10 07:18:28 | ERROR | stderr | | File "/usr/local/lib/python3.10/site-packages/starlette/middleware/errors.py", line 186, in __call__
|
288 |
+
2024-07-10 07:18:28 | ERROR | stderr | | raise exc
|
289 |
+
2024-07-10 07:18:28 | ERROR | stderr | | File "/usr/local/lib/python3.10/site-packages/starlette/middleware/errors.py", line 164, in __call__
|
290 |
+
2024-07-10 07:18:28 | ERROR | stderr | | await self.app(scope, receive, _send)
|
291 |
+
2024-07-10 07:18:28 | ERROR | stderr | | File "/usr/local/lib/python3.10/site-packages/gradio/route_utils.py", line 714, in __call__
|
292 |
+
2024-07-10 07:18:28 | ERROR | stderr | | await self.app(scope, receive, send)
|
293 |
+
2024-07-10 07:18:28 | ERROR | stderr | | File "/usr/local/lib/python3.10/site-packages/starlette/middleware/exceptions.py", line 65, in __call__
|
294 |
+
2024-07-10 07:18:28 | ERROR | stderr | | await wrap_app_handling_exceptions(self.app, conn)(scope, receive, send)
|
295 |
+
2024-07-10 07:18:28 | ERROR | stderr | | File "/usr/local/lib/python3.10/site-packages/starlette/_exception_handler.py", line 64, in wrapped_app
|
296 |
+
2024-07-10 07:18:28 | ERROR | stderr | | raise exc
|
297 |
+
2024-07-10 07:18:28 | ERROR | stderr | | File "/usr/local/lib/python3.10/site-packages/starlette/_exception_handler.py", line 53, in wrapped_app
|
298 |
+
2024-07-10 07:18:28 | ERROR | stderr | | await app(scope, receive, sender)
|
299 |
+
2024-07-10 07:18:28 | ERROR | stderr | | File "/usr/local/lib/python3.10/site-packages/starlette/routing.py", line 756, in __call__
|
300 |
+
2024-07-10 07:18:28 | ERROR | stderr | | await self.middleware_stack(scope, receive, send)
|
301 |
+
2024-07-10 07:18:28 | ERROR | stderr | | File "/usr/local/lib/python3.10/site-packages/starlette/routing.py", line 776, in app
|
302 |
+
2024-07-10 07:18:28 | ERROR | stderr | | await route.handle(scope, receive, send)
|
303 |
+
2024-07-10 07:18:28 | ERROR | stderr | | File "/usr/local/lib/python3.10/site-packages/starlette/routing.py", line 297, in handle
|
304 |
+
2024-07-10 07:18:28 | ERROR | stderr | | await self.app(scope, receive, send)
|
305 |
+
2024-07-10 07:18:28 | ERROR | stderr | | File "/usr/local/lib/python3.10/site-packages/starlette/routing.py", line 77, in app
|
306 |
+
2024-07-10 07:18:28 | ERROR | stderr | | await wrap_app_handling_exceptions(app, request)(scope, receive, send)
|
307 |
+
2024-07-10 07:18:28 | ERROR | stderr | | File "/usr/local/lib/python3.10/site-packages/starlette/_exception_handler.py", line 64, in wrapped_app
|
308 |
+
2024-07-10 07:18:28 | ERROR | stderr | | raise exc
|
309 |
+
2024-07-10 07:18:28 | ERROR | stderr | | File "/usr/local/lib/python3.10/site-packages/starlette/_exception_handler.py", line 53, in wrapped_app
|
310 |
+
2024-07-10 07:18:28 | ERROR | stderr | | await app(scope, receive, sender)
|
311 |
+
2024-07-10 07:18:28 | ERROR | stderr | | File "/usr/local/lib/python3.10/site-packages/starlette/routing.py", line 75, in app
|
312 |
+
2024-07-10 07:18:28 | ERROR | stderr | | await response(scope, receive, send)
|
313 |
+
2024-07-10 07:18:28 | ERROR | stderr | | File "/usr/local/lib/python3.10/site-packages/starlette/responses.py", line 258, in __call__
|
314 |
+
2024-07-10 07:18:28 | ERROR | stderr | | async with anyio.create_task_group() as task_group:
|
315 |
+
2024-07-10 07:18:28 | ERROR | stderr | | File "/usr/local/lib/python3.10/site-packages/anyio/_backends/_asyncio.py", line 680, in __aexit__
|
316 |
+
2024-07-10 07:18:28 | ERROR | stderr | | raise BaseExceptionGroup(
|
317 |
+
2024-07-10 07:18:28 | ERROR | stderr | | exceptiongroup.ExceptionGroup: unhandled errors in a TaskGroup (1 sub-exception)
|
318 |
+
2024-07-10 07:18:28 | ERROR | stderr | +-+---------------- 1 ----------------
|
319 |
+
2024-07-10 07:18:28 | ERROR | stderr | | Traceback (most recent call last):
|
320 |
+
2024-07-10 07:18:28 | ERROR | stderr | | File "/usr/local/lib/python3.10/site-packages/starlette/responses.py", line 261, in wrap
|
321 |
+
2024-07-10 07:18:28 | ERROR | stderr | | await func()
|
322 |
+
2024-07-10 07:18:28 | ERROR | stderr | | File "/usr/local/lib/python3.10/site-packages/starlette/responses.py", line 250, in stream_response
|
323 |
+
2024-07-10 07:18:28 | ERROR | stderr | | async for chunk in self.body_iterator:
|
324 |
+
2024-07-10 07:18:28 | ERROR | stderr | | File "/usr/local/lib/python3.10/site-packages/gradio/routes.py", line 670, in iterator
|
325 |
+
2024-07-10 07:18:28 | ERROR | stderr | | done = [d.result() for d in done]
|
326 |
+
2024-07-10 07:18:28 | ERROR | stderr | | File "/usr/local/lib/python3.10/site-packages/gradio/routes.py", line 670, in <listcomp>
|
327 |
+
2024-07-10 07:18:28 | ERROR | stderr | | done = [d.result() for d in done]
|
328 |
+
2024-07-10 07:18:28 | ERROR | stderr | | File "/usr/local/lib/python3.10/site-packages/gradio/routes.py", line 655, in stop_stream
|
329 |
+
2024-07-10 07:18:28 | ERROR | stderr | | await app.stop_event.wait()
|
330 |
+
2024-07-10 07:18:28 | ERROR | stderr | | AttributeError: 'NoneType' object has no attribute 'wait'
|
331 |
+
2024-07-10 07:18:28 | ERROR | stderr | +------------------------------------
|
332 |
+
2024-07-10 07:19:12 | ERROR | stderr | ERROR: Exception in ASGI application
|
333 |
+
2024-07-10 07:19:12 | ERROR | stderr | Traceback (most recent call last):
|
334 |
+
2024-07-10 07:19:12 | ERROR | stderr | File "/usr/local/lib/python3.10/site-packages/starlette/responses.py", line 265, in __call__
|
335 |
+
2024-07-10 07:19:12 | ERROR | stderr | await wrap(partial(self.listen_for_disconnect, receive))
|
336 |
+
2024-07-10 07:19:12 | ERROR | stderr | File "/usr/local/lib/python3.10/site-packages/starlette/responses.py", line 261, in wrap
|
337 |
+
2024-07-10 07:19:12 | ERROR | stderr | await func()
|
338 |
+
2024-07-10 07:19:12 | ERROR | stderr | File "/usr/local/lib/python3.10/site-packages/starlette/responses.py", line 238, in listen_for_disconnect
|
339 |
+
2024-07-10 07:19:12 | ERROR | stderr | message = await receive()
|
340 |
+
2024-07-10 07:19:12 | ERROR | stderr | File "/usr/local/lib/python3.10/site-packages/uvicorn/protocols/http/httptools_impl.py", line 553, in receive
|
341 |
+
2024-07-10 07:19:12 | ERROR | stderr | await self.message_event.wait()
|
342 |
+
2024-07-10 07:19:12 | ERROR | stderr | File "/usr/local/lib/python3.10/asyncio/locks.py", line 214, in wait
|
343 |
+
2024-07-10 07:19:12 | ERROR | stderr | await fut
|
344 |
+
2024-07-10 07:19:12 | ERROR | stderr | asyncio.exceptions.CancelledError: Cancelled by cancel scope 7ffa696ff760
|
345 |
+
2024-07-10 07:19:12 | ERROR | stderr |
|
346 |
+
2024-07-10 07:19:12 | ERROR | stderr | During handling of the above exception, another exception occurred:
|
347 |
+
2024-07-10 07:19:12 | ERROR | stderr |
|
348 |
+
2024-07-10 07:19:12 | ERROR | stderr | + Exception Group Traceback (most recent call last):
|
349 |
+
2024-07-10 07:19:12 | ERROR | stderr | | File "/usr/local/lib/python3.10/site-packages/uvicorn/protocols/http/httptools_impl.py", line 399, in run_asgi
|
350 |
+
2024-07-10 07:19:12 | ERROR | stderr | | result = await app( # type: ignore[func-returns-value]
|
351 |
+
2024-07-10 07:19:12 | ERROR | stderr | | File "/usr/local/lib/python3.10/site-packages/uvicorn/middleware/proxy_headers.py", line 70, in __call__
|
352 |
+
2024-07-10 07:19:12 | ERROR | stderr | | return await self.app(scope, receive, send)
|
353 |
+
2024-07-10 07:19:12 | ERROR | stderr | | File "/usr/local/lib/python3.10/site-packages/fastapi/applications.py", line 1054, in __call__
|
354 |
+
2024-07-10 07:19:12 | ERROR | stderr | | await super().__call__(scope, receive, send)
|
355 |
+
2024-07-10 07:19:12 | ERROR | stderr | | File "/usr/local/lib/python3.10/site-packages/starlette/applications.py", line 123, in __call__
|
356 |
+
2024-07-10 07:19:12 | ERROR | stderr | | await self.middleware_stack(scope, receive, send)
|
357 |
+
2024-07-10 07:19:12 | ERROR | stderr | | File "/usr/local/lib/python3.10/site-packages/starlette/middleware/errors.py", line 186, in __call__
|
358 |
+
2024-07-10 07:19:12 | ERROR | stderr | | raise exc
|
359 |
+
2024-07-10 07:19:12 | ERROR | stderr | | File "/usr/local/lib/python3.10/site-packages/starlette/middleware/errors.py", line 164, in __call__
|
360 |
+
2024-07-10 07:19:12 | ERROR | stderr | | await self.app(scope, receive, _send)
|
361 |
+
2024-07-10 07:19:12 | ERROR | stderr | | File "/usr/local/lib/python3.10/site-packages/gradio/route_utils.py", line 714, in __call__
|
362 |
+
2024-07-10 07:19:12 | ERROR | stderr | | await self.app(scope, receive, send)
|
363 |
+
2024-07-10 07:19:12 | ERROR | stderr | | File "/usr/local/lib/python3.10/site-packages/starlette/middleware/exceptions.py", line 65, in __call__
|
364 |
+
2024-07-10 07:19:12 | ERROR | stderr | | await wrap_app_handling_exceptions(self.app, conn)(scope, receive, send)
|
365 |
+
2024-07-10 07:19:12 | ERROR | stderr | | File "/usr/local/lib/python3.10/site-packages/starlette/_exception_handler.py", line 64, in wrapped_app
|
366 |
+
2024-07-10 07:19:12 | ERROR | stderr | | raise exc
|
367 |
+
2024-07-10 07:19:12 | ERROR | stderr | | File "/usr/local/lib/python3.10/site-packages/starlette/_exception_handler.py", line 53, in wrapped_app
|
368 |
+
2024-07-10 07:19:12 | ERROR | stderr | | await app(scope, receive, sender)
|
369 |
+
2024-07-10 07:19:12 | ERROR | stderr | | File "/usr/local/lib/python3.10/site-packages/starlette/routing.py", line 756, in __call__
|
370 |
+
2024-07-10 07:19:12 | ERROR | stderr | | await self.middleware_stack(scope, receive, send)
|
371 |
+
2024-07-10 07:19:12 | ERROR | stderr | | File "/usr/local/lib/python3.10/site-packages/starlette/routing.py", line 776, in app
|
372 |
+
2024-07-10 07:19:12 | ERROR | stderr | | await route.handle(scope, receive, send)
|
373 |
+
2024-07-10 07:19:12 | ERROR | stderr | | File "/usr/local/lib/python3.10/site-packages/starlette/routing.py", line 297, in handle
|
374 |
+
2024-07-10 07:19:12 | ERROR | stderr | | await self.app(scope, receive, send)
|
375 |
+
2024-07-10 07:19:12 | ERROR | stderr | | File "/usr/local/lib/python3.10/site-packages/starlette/routing.py", line 77, in app
|
376 |
+
2024-07-10 07:19:12 | ERROR | stderr | | await wrap_app_handling_exceptions(app, request)(scope, receive, send)
|
377 |
+
2024-07-10 07:19:12 | ERROR | stderr | | File "/usr/local/lib/python3.10/site-packages/starlette/_exception_handler.py", line 64, in wrapped_app
|
378 |
+
2024-07-10 07:19:12 | ERROR | stderr | | raise exc
|
379 |
+
2024-07-10 07:19:12 | ERROR | stderr | | File "/usr/local/lib/python3.10/site-packages/starlette/_exception_handler.py", line 53, in wrapped_app
|
380 |
+
2024-07-10 07:19:12 | ERROR | stderr | | await app(scope, receive, sender)
|
381 |
+
2024-07-10 07:19:12 | ERROR | stderr | | File "/usr/local/lib/python3.10/site-packages/starlette/routing.py", line 75, in app
|
382 |
+
2024-07-10 07:19:12 | ERROR | stderr | | await response(scope, receive, send)
|
383 |
+
2024-07-10 07:19:12 | ERROR | stderr | | File "/usr/local/lib/python3.10/site-packages/starlette/responses.py", line 258, in __call__
|
384 |
+
2024-07-10 07:19:12 | ERROR | stderr | | async with anyio.create_task_group() as task_group:
|
385 |
+
2024-07-10 07:19:12 | ERROR | stderr | | File "/usr/local/lib/python3.10/site-packages/anyio/_backends/_asyncio.py", line 680, in __aexit__
|
386 |
+
2024-07-10 07:19:12 | ERROR | stderr | | raise BaseExceptionGroup(
|
387 |
+
2024-07-10 07:19:12 | ERROR | stderr | | exceptiongroup.ExceptionGroup: unhandled errors in a TaskGroup (1 sub-exception)
|
388 |
+
2024-07-10 07:19:12 | ERROR | stderr | +-+---------------- 1 ----------------
|
389 |
+
2024-07-10 07:19:12 | ERROR | stderr | | Traceback (most recent call last):
|
390 |
+
2024-07-10 07:19:12 | ERROR | stderr | | File "/usr/local/lib/python3.10/site-packages/starlette/responses.py", line 261, in wrap
|
391 |
+
2024-07-10 07:19:12 | ERROR | stderr | | await func()
|
392 |
+
2024-07-10 07:19:12 | ERROR | stderr | | File "/usr/local/lib/python3.10/site-packages/starlette/responses.py", line 250, in stream_response
|
393 |
+
2024-07-10 07:19:12 | ERROR | stderr | | async for chunk in self.body_iterator:
|
394 |
+
2024-07-10 07:19:12 | ERROR | stderr | | File "/usr/local/lib/python3.10/site-packages/gradio/routes.py", line 670, in iterator
|
395 |
+
2024-07-10 07:19:12 | ERROR | stderr | | done = [d.result() for d in done]
|
396 |
+
2024-07-10 07:19:12 | ERROR | stderr | | File "/usr/local/lib/python3.10/site-packages/gradio/routes.py", line 670, in <listcomp>
|
397 |
+
2024-07-10 07:19:12 | ERROR | stderr | | done = [d.result() for d in done]
|
398 |
+
2024-07-10 07:19:12 | ERROR | stderr | | File "/usr/local/lib/python3.10/site-packages/gradio/routes.py", line 655, in stop_stream
|
399 |
+
2024-07-10 07:19:12 | ERROR | stderr | | await app.stop_event.wait()
|
400 |
+
2024-07-10 07:19:12 | ERROR | stderr | | AttributeError: 'NoneType' object has no attribute 'wait'
|
401 |
+
2024-07-10 07:19:12 | ERROR | stderr | +------------------------------------
|
402 |
+
2024-07-10 07:19:24 | ERROR | stderr | Traceback (most recent call last):
|
403 |
+
2024-07-10 07:19:24 | ERROR | stderr | File "/home/user/app/app.py", line 18, in <module>
|
404 |
+
2024-07-10 07:19:24 | ERROR | stderr | main()
|
405 |
+
2024-07-10 07:19:24 | ERROR | stderr | File "/home/user/app/app.py", line 15, in main
|
406 |
+
2024-07-10 07:19:24 | ERROR | stderr | demo.launch()
|
407 |
+
2024-07-10 07:19:24 | ERROR | stderr | File "/usr/local/lib/python3.10/site-packages/gradio/blocks.py", line 2327, in launch
|
408 |
+
2024-07-10 07:19:24 | ERROR | stderr | ) = http_server.start_server(
|
409 |
+
2024-07-10 07:19:24 | ERROR | stderr | File "/usr/local/lib/python3.10/site-packages/gradio/http_server.py", line 154, in start_server
|
410 |
+
2024-07-10 07:19:24 | ERROR | stderr | raise OSError(
|
411 |
+
2024-07-10 07:19:24 | ERROR | stderr | OSError: Cannot find empty port in range: 7860-7860. You can specify a different port by setting the GRADIO_SERVER_PORT environment variable or passing the `server_port` parameter to `launch()`.
|
412 |
+
2024-07-10 07:20:37 | INFO | stdout | Running on local URL: http://0.0.0.0:7860
|
413 |
+
2024-07-10 07:20:37 | INFO | stdout |
|
414 |
+
2024-07-10 07:20:37 | INFO | stdout | To create a public link, set `share=True` in `launch()`.
|
415 |
+
2024-07-10 07:20:58 | INFO | gradio_web_server | bot_response. ip: 46.3.240.105
|
416 |
+
2024-07-10 07:20:58 | INFO | gradio_web_server | monitor error: HTTPConnectionPool(host='localhost', port=9090): Max retries exceeded with url: /is_limit_reached?model=llava-fire&user_id=46.3.240.105 (Caused by NewConnectionError('<urllib3.connection.HTTPConnection object at 0x7f75d015d8d0>: Failed to establish a new connection: [Errno 111] Connection refused'))
|
417 |
+
2024-07-10 07:20:58 | INFO | gradio_web_server | model_name: llava-fire;model_api_dict: None
|
418 |
+
2024-07-10 07:20:58 | INFO | gradio_web_server | bot_response. ip: 46.3.240.105
|
419 |
+
2024-07-10 07:20:58 | INFO | gradio_web_server | monitor error: HTTPConnectionPool(host='localhost', port=9090): Max retries exceeded with url: /is_limit_reached?model=llava-original&user_id=46.3.240.105 (Caused by NewConnectionError('<urllib3.connection.HTTPConnection object at 0x7f75d015cf70>: Failed to establish a new connection: [Errno 111] Connection refused'))
|
420 |
+
2024-07-10 07:20:58 | INFO | gradio_web_server | model_name: llava-original;model_api_dict: None
|
421 |
+
2024-07-10 07:20:58 | INFO | gradio_web_server | hello
|
422 |
+
2024-07-10 07:20:58 | INFO | gradio_web_server | hello
|
423 |
+
2024-07-10 07:23:46 | INFO | stdout | Running on local URL: http://0.0.0.0:7860
|
424 |
+
2024-07-10 07:23:46 | INFO | stdout |
|
425 |
+
2024-07-10 07:23:46 | INFO | stdout | To create a public link, set `share=True` in `launch()`.
|
426 |
+
2024-07-10 07:24:07 | INFO | gradio_web_server | bot_response. ip: 46.3.240.105
|
427 |
+
2024-07-10 07:24:07 | INFO | gradio_web_server | monitor error: HTTPConnectionPool(host='localhost', port=9090): Max retries exceeded with url: /is_limit_reached?model=llava-fire&user_id=46.3.240.105 (Caused by NewConnectionError('<urllib3.connection.HTTPConnection object at 0x7f36e964f100>: Failed to establish a new connection: [Errno 111] Connection refused'))
|
428 |
+
2024-07-10 07:24:07 | INFO | gradio_web_server | model_name: llava-fire;model_api_dict: None
|
429 |
+
2024-07-10 07:24:07 | INFO | gradio_web_server | bot_response. ip: 46.3.240.105
|
430 |
+
2024-07-10 07:24:07 | INFO | gradio_web_server | monitor error: HTTPConnectionPool(host='localhost', port=9090): Max retries exceeded with url: /is_limit_reached?model=llava-original&user_id=46.3.240.105 (Caused by NewConnectionError('<urllib3.connection.HTTPConnection object at 0x7f36e964d4e0>: Failed to establish a new connection: [Errno 111] Connection refused'))
|
431 |
+
2024-07-10 07:24:07 | INFO | gradio_web_server | model_name: llava-original;model_api_dict: None
|
432 |
+
2024-07-10 07:24:15 | INFO | stdout | torch.Size([1, 58]) torch.Size([1, 5, 3, 336, 336])
|
433 |
+
2024-07-10 07:24:16 | ERROR | stderr | /usr/local/lib/python3.10/site-packages/transformers/generation/configuration_utils.py:392: UserWarning: `do_sample` is set to `False`. However, `temperature` is set to `0` -- this flag is only used in sample-based generation modes. You should set `do_sample=True` or unset `temperature`.
|
434 |
+
2024-07-10 07:24:16 | ERROR | stderr | warnings.warn(
|
435 |
+
2024-07-10 07:24:16 | ERROR | stderr | /usr/local/lib/python3.10/site-packages/transformers/generation/configuration_utils.py:397: UserWarning: `do_sample` is set to `False`. However, `top_p` is set to `0.9` -- this flag is only used in sample-based generation modes. You should set `do_sample=True` or unset `top_p`.
|
436 |
+
2024-07-10 07:24:16 | ERROR | stderr | warnings.warn(
|
437 |
+
2024-07-10 07:24:22 | INFO | stdout | ["The figure in the image is a can of Södra Almighy, which is a beer. Södra is a Swedish brewery known for its lager beers. The can's design is modern and minimalist, with a color scheme that includes black and white, and the text is in English. The label indicates that it is a dry-hopped beer, which means it has been flavored with hops that have not been steeped in the brewing process, giving it a unique taste profile. The can's design suggests a contemporary and possibly craft beer, which is often associated with a more complex flavor profile than traditional lagers."]
|
438 |
+
2024-07-10 07:24:22 | INFO | gradio_web_server | hello
|
439 |
+
2024-07-10 07:24:48 | INFO | stdout | Running on local URL: http://0.0.0.0:7860
|
440 |
+
2024-07-10 07:24:48 | INFO | stdout |
|
441 |
+
2024-07-10 07:24:48 | INFO | stdout | To create a public link, set `share=True` in `launch()`.
|
gradio_web_server_multi.log
CHANGED
@@ -8,3 +8,7 @@
|
|
8 |
2024-07-10 11:58:24 | INFO | gradio_web_server_multi | bot_response_multi (named). ip: 127.0.0.1
|
9 |
2024-07-10 11:58:46 | INFO | gradio_web_server_multi | add_text (named). ip: 127.0.0.1. len: 5
|
10 |
2024-07-10 11:58:46 | INFO | gradio_web_server_multi | bot_response_multi (named). ip: 127.0.0.1
|
|
|
|
|
|
|
|
|
|
8 |
2024-07-10 11:58:24 | INFO | gradio_web_server_multi | bot_response_multi (named). ip: 127.0.0.1
|
9 |
2024-07-10 11:58:46 | INFO | gradio_web_server_multi | add_text (named). ip: 127.0.0.1. len: 5
|
10 |
2024-07-10 11:58:46 | INFO | gradio_web_server_multi | bot_response_multi (named). ip: 127.0.0.1
|
11 |
+
2024-07-10 07:20:57 | INFO | gradio_web_server_multi | add_text (named). ip: 46.3.240.105. len: 5
|
12 |
+
2024-07-10 07:20:58 | INFO | gradio_web_server_multi | bot_response_multi (named). ip: 46.3.240.105
|
13 |
+
2024-07-10 07:24:07 | INFO | gradio_web_server_multi | add_text (named). ip: 46.3.240.105. len: 5
|
14 |
+
2024-07-10 07:24:07 | INFO | gradio_web_server_multi | bot_response_multi (named). ip: 46.3.240.105
|
src/__pycache__/__init__.cpython-310.pyc
CHANGED
Binary files a/src/__pycache__/__init__.cpython-310.pyc and b/src/__pycache__/__init__.cpython-310.pyc differ
|
|
src/__pycache__/constants.cpython-310.pyc
CHANGED
Binary files a/src/__pycache__/constants.cpython-310.pyc and b/src/__pycache__/constants.cpython-310.pyc differ
|
|
src/__pycache__/conversation.cpython-310.pyc
CHANGED
Binary files a/src/__pycache__/conversation.cpython-310.pyc and b/src/__pycache__/conversation.cpython-310.pyc differ
|
|
src/__pycache__/utils.cpython-310.pyc
CHANGED
Binary files a/src/__pycache__/utils.cpython-310.pyc and b/src/__pycache__/utils.cpython-310.pyc differ
|
|
src/model/__pycache__/__init__.cpython-310.pyc
CHANGED
Binary files a/src/model/__pycache__/__init__.cpython-310.pyc and b/src/model/__pycache__/__init__.cpython-310.pyc differ
|
|
src/model/__pycache__/compression.cpython-310.pyc
CHANGED
Binary files a/src/model/__pycache__/compression.cpython-310.pyc and b/src/model/__pycache__/compression.cpython-310.pyc differ
|
|
src/model/__pycache__/llama_condense_monkey_patch.cpython-310.pyc
CHANGED
Binary files a/src/model/__pycache__/llama_condense_monkey_patch.cpython-310.pyc and b/src/model/__pycache__/llama_condense_monkey_patch.cpython-310.pyc differ
|
|
src/model/__pycache__/model_adapter.cpython-310.pyc
CHANGED
Binary files a/src/model/__pycache__/model_adapter.cpython-310.pyc and b/src/model/__pycache__/model_adapter.cpython-310.pyc differ
|
|
src/model/__pycache__/model_chatglm.cpython-310.pyc
CHANGED
Binary files a/src/model/__pycache__/model_chatglm.cpython-310.pyc and b/src/model/__pycache__/model_chatglm.cpython-310.pyc differ
|
|
src/model/__pycache__/model_cllm.cpython-310.pyc
CHANGED
Binary files a/src/model/__pycache__/model_cllm.cpython-310.pyc and b/src/model/__pycache__/model_cllm.cpython-310.pyc differ
|
|
src/model/__pycache__/model_codet5p.cpython-310.pyc
CHANGED
Binary files a/src/model/__pycache__/model_codet5p.cpython-310.pyc and b/src/model/__pycache__/model_codet5p.cpython-310.pyc differ
|
|
src/model/__pycache__/model_exllama.cpython-310.pyc
CHANGED
Binary files a/src/model/__pycache__/model_exllama.cpython-310.pyc and b/src/model/__pycache__/model_exllama.cpython-310.pyc differ
|
|
src/model/__pycache__/model_falcon.cpython-310.pyc
CHANGED
Binary files a/src/model/__pycache__/model_falcon.cpython-310.pyc and b/src/model/__pycache__/model_falcon.cpython-310.pyc differ
|
|
src/model/__pycache__/model_registry.cpython-310.pyc
CHANGED
Binary files a/src/model/__pycache__/model_registry.cpython-310.pyc and b/src/model/__pycache__/model_registry.cpython-310.pyc differ
|
|
src/model/__pycache__/model_xfastertransformer.cpython-310.pyc
CHANGED
Binary files a/src/model/__pycache__/model_xfastertransformer.cpython-310.pyc and b/src/model/__pycache__/model_xfastertransformer.cpython-310.pyc differ
|
|
src/model/__pycache__/model_yuan2.cpython-310.pyc
CHANGED
Binary files a/src/model/__pycache__/model_yuan2.cpython-310.pyc and b/src/model/__pycache__/model_yuan2.cpython-310.pyc differ
|
|
src/model/__pycache__/monkey_patch_non_inplace.cpython-310.pyc
CHANGED
Binary files a/src/model/__pycache__/monkey_patch_non_inplace.cpython-310.pyc and b/src/model/__pycache__/monkey_patch_non_inplace.cpython-310.pyc differ
|
|
src/modules/__pycache__/__init__.cpython-310.pyc
CHANGED
Binary files a/src/modules/__pycache__/__init__.cpython-310.pyc and b/src/modules/__pycache__/__init__.cpython-310.pyc differ
|
|
src/modules/__pycache__/awq.cpython-310.pyc
CHANGED
Binary files a/src/modules/__pycache__/awq.cpython-310.pyc and b/src/modules/__pycache__/awq.cpython-310.pyc differ
|
|
src/modules/__pycache__/exllama.cpython-310.pyc
CHANGED
Binary files a/src/modules/__pycache__/exllama.cpython-310.pyc and b/src/modules/__pycache__/exllama.cpython-310.pyc differ
|
|
src/modules/__pycache__/gptq.cpython-310.pyc
CHANGED
Binary files a/src/modules/__pycache__/gptq.cpython-310.pyc and b/src/modules/__pycache__/gptq.cpython-310.pyc differ
|
|
src/modules/__pycache__/xfastertransformer.cpython-310.pyc
CHANGED
Binary files a/src/modules/__pycache__/xfastertransformer.cpython-310.pyc and b/src/modules/__pycache__/xfastertransformer.cpython-310.pyc differ
|
|
src/serve/__pycache__/__init__.cpython-310.pyc
CHANGED
Binary files a/src/serve/__pycache__/__init__.cpython-310.pyc and b/src/serve/__pycache__/__init__.cpython-310.pyc differ
|
|
src/serve/__pycache__/api_provider.cpython-310.pyc
CHANGED
Binary files a/src/serve/__pycache__/api_provider.cpython-310.pyc and b/src/serve/__pycache__/api_provider.cpython-310.pyc differ
|
|
src/serve/__pycache__/gradio_block_arena_named.cpython-310.pyc
CHANGED
Binary files a/src/serve/__pycache__/gradio_block_arena_named.cpython-310.pyc and b/src/serve/__pycache__/gradio_block_arena_named.cpython-310.pyc differ
|
|
src/serve/__pycache__/gradio_block_arena_vision.cpython-310.pyc
CHANGED
Binary files a/src/serve/__pycache__/gradio_block_arena_vision.cpython-310.pyc and b/src/serve/__pycache__/gradio_block_arena_vision.cpython-310.pyc differ
|
|
src/serve/__pycache__/gradio_block_arena_vision_named.cpython-310.pyc
CHANGED
Binary files a/src/serve/__pycache__/gradio_block_arena_vision_named.cpython-310.pyc and b/src/serve/__pycache__/gradio_block_arena_vision_named.cpython-310.pyc differ
|
|
src/serve/__pycache__/gradio_web_server.cpython-310.pyc
CHANGED
Binary files a/src/serve/__pycache__/gradio_web_server.cpython-310.pyc and b/src/serve/__pycache__/gradio_web_server.cpython-310.pyc differ
|
|
src/serve/__pycache__/remote_logger.cpython-310.pyc
CHANGED
Binary files a/src/serve/__pycache__/remote_logger.cpython-310.pyc and b/src/serve/__pycache__/remote_logger.cpython-310.pyc differ
|
|
src/serve/gradio_web_server.py
CHANGED
@@ -451,10 +451,14 @@ def bot_response(
|
|
451 |
images = conv.get_images()
|
452 |
logger.info(f"model_name: {model_name};model_api_dict: {model_api_dict}")
|
453 |
if model_api_dict is None:
|
454 |
-
|
|
|
|
|
|
|
|
|
455 |
stream_iter = [{
|
456 |
"error_code": 0,
|
457 |
-
"text":
|
458 |
}]
|
459 |
# Query worker address
|
460 |
# ret = requests.post(
|
|
|
451 |
images = conv.get_images()
|
452 |
logger.info(f"model_name: {model_name};model_api_dict: {model_api_dict}")
|
453 |
if model_api_dict is None:
|
454 |
+
if model_name == "llava-original":
|
455 |
+
from src.model.model_llava import inference
|
456 |
+
output_text = inference()[0]
|
457 |
+
else:
|
458 |
+
output_text = "hello"
|
459 |
stream_iter = [{
|
460 |
"error_code": 0,
|
461 |
+
"text": output_text
|
462 |
}]
|
463 |
# Query worker address
|
464 |
# ret = requests.post(
|
vision-tmp-2024-07-10-conv.json
CHANGED
@@ -1,2 +1,5 @@
|
|
1 |
{"tstamp": 1720583926.6994, "type": "chat", "model": "llava-fire", "gen_params": {"temperature": 0.7, "top_p": 1.0, "max_new_tokens": 1024}, "start": 1720583926.678, "finish": 1720583926.6994, "state": {"template_name": "vicuna_v1.1", "system_message": "A chat between a curious user and an artificial intelligence assistant. The assistant gives helpful, detailed, and polite answers to the user's questions.", "roles": ["USER", "ASSISTANT"], "messages": [["USER", "Hello"], ["ASSISTANT", "hello"]], "offset": 0, "conv_id": "a48164214dce453a9ad276401b7d095c", "model_name": "llava-fire", "has_csam_image": false}, "ip": "127.0.0.1"}
|
2 |
{"tstamp": 1720583926.6998, "type": "chat", "model": "llava-original", "gen_params": {"temperature": 0.7, "top_p": 1.0, "max_new_tokens": 1024}, "start": 1720583926.6815, "finish": 1720583926.6998, "state": {"template_name": "vicuna_v1.1", "system_message": "A chat between a curious user and an artificial intelligence assistant. The assistant gives helpful, detailed, and polite answers to the user's questions.", "roles": ["USER", "ASSISTANT"], "messages": [["USER", "Hello"], ["ASSISTANT", "hello"]], "offset": 0, "conv_id": "8bed262a5728409284a7a56a0fe66a75", "model_name": "llava-original", "has_csam_image": false}, "ip": "127.0.0.1"}
|
|
|
|
|
|
|
|
1 |
{"tstamp": 1720583926.6994, "type": "chat", "model": "llava-fire", "gen_params": {"temperature": 0.7, "top_p": 1.0, "max_new_tokens": 1024}, "start": 1720583926.678, "finish": 1720583926.6994, "state": {"template_name": "vicuna_v1.1", "system_message": "A chat between a curious user and an artificial intelligence assistant. The assistant gives helpful, detailed, and polite answers to the user's questions.", "roles": ["USER", "ASSISTANT"], "messages": [["USER", "Hello"], ["ASSISTANT", "hello"]], "offset": 0, "conv_id": "a48164214dce453a9ad276401b7d095c", "model_name": "llava-fire", "has_csam_image": false}, "ip": "127.0.0.1"}
|
2 |
{"tstamp": 1720583926.6998, "type": "chat", "model": "llava-original", "gen_params": {"temperature": 0.7, "top_p": 1.0, "max_new_tokens": 1024}, "start": 1720583926.6815, "finish": 1720583926.6998, "state": {"template_name": "vicuna_v1.1", "system_message": "A chat between a curious user and an artificial intelligence assistant. The assistant gives helpful, detailed, and polite answers to the user's questions.", "roles": ["USER", "ASSISTANT"], "messages": [["USER", "Hello"], ["ASSISTANT", "hello"]], "offset": 0, "conv_id": "8bed262a5728409284a7a56a0fe66a75", "model_name": "llava-original", "has_csam_image": false}, "ip": "127.0.0.1"}
|
3 |
+
{"tstamp": 1720588858.8939, "type": "chat", "model": "llava-fire", "gen_params": {"temperature": 0.7, "top_p": 1.0, "max_new_tokens": 1024}, "start": 1720588858.8843, "finish": 1720588858.8939, "state": {"template_name": "vicuna_v1.1", "system_message": "A chat between a curious user and an artificial intelligence assistant. The assistant gives helpful, detailed, and polite answers to the user's questions.", "roles": ["USER", "ASSISTANT"], "messages": [["USER", "Hello"], ["ASSISTANT", "hello"]], "offset": 0, "conv_id": "d994e3d859c94bddbf0dfcaed6c63079", "model_name": "llava-fire", "has_csam_image": false}, "ip": "46.3.240.105"}
|
4 |
+
{"tstamp": 1720588858.8951, "type": "chat", "model": "llava-original", "gen_params": {"temperature": 0.7, "top_p": 1.0, "max_new_tokens": 1024}, "start": 1720588858.8863, "finish": 1720588858.8951, "state": {"template_name": "vicuna_v1.1", "system_message": "A chat between a curious user and an artificial intelligence assistant. The assistant gives helpful, detailed, and polite answers to the user's questions.", "roles": ["USER", "ASSISTANT"], "messages": [["USER", "Hello"], ["ASSISTANT", "hello"]], "offset": 0, "conv_id": "e41b47f05f8b44ff9d520c8e94c6e8de", "model_name": "llava-original", "has_csam_image": false}, "ip": "46.3.240.105"}
|
5 |
+
{"tstamp": 1720589062.1758, "type": "chat", "model": "llava-fire", "gen_params": {"temperature": 0.7, "top_p": 1.0, "max_new_tokens": 1024}, "start": 1720589047.9171, "finish": 1720589062.1758, "state": {"template_name": "vicuna_v1.1", "system_message": "A chat between a curious user and an artificial intelligence assistant. The assistant gives helpful, detailed, and polite answers to the user's questions.", "roles": ["USER", "ASSISTANT"], "messages": [["USER", "Hello"], ["ASSISTANT", "hello"]], "offset": 0, "conv_id": "963f15cd5e224eb8ae02c67ed37b93c4", "model_name": "llava-fire", "has_csam_image": false}, "ip": "46.3.240.105"}
|