li-qing commited on
Commit
682d6db
1 Parent(s): 73237fe

fix: add inference

Browse files
Files changed (35) hide show
  1. .gitignore +6 -1
  2. app.py +0 -1
  3. gradio_web_server.log +185 -0
  4. gradio_web_server_multi.log +4 -0
  5. src/__pycache__/__init__.cpython-310.pyc +0 -0
  6. src/__pycache__/constants.cpython-310.pyc +0 -0
  7. src/__pycache__/conversation.cpython-310.pyc +0 -0
  8. src/__pycache__/utils.cpython-310.pyc +0 -0
  9. src/model/__pycache__/__init__.cpython-310.pyc +0 -0
  10. src/model/__pycache__/compression.cpython-310.pyc +0 -0
  11. src/model/__pycache__/llama_condense_monkey_patch.cpython-310.pyc +0 -0
  12. src/model/__pycache__/model_adapter.cpython-310.pyc +0 -0
  13. src/model/__pycache__/model_chatglm.cpython-310.pyc +0 -0
  14. src/model/__pycache__/model_cllm.cpython-310.pyc +0 -0
  15. src/model/__pycache__/model_codet5p.cpython-310.pyc +0 -0
  16. src/model/__pycache__/model_exllama.cpython-310.pyc +0 -0
  17. src/model/__pycache__/model_falcon.cpython-310.pyc +0 -0
  18. src/model/__pycache__/model_registry.cpython-310.pyc +0 -0
  19. src/model/__pycache__/model_xfastertransformer.cpython-310.pyc +0 -0
  20. src/model/__pycache__/model_yuan2.cpython-310.pyc +0 -0
  21. src/model/__pycache__/monkey_patch_non_inplace.cpython-310.pyc +0 -0
  22. src/modules/__pycache__/__init__.cpython-310.pyc +0 -0
  23. src/modules/__pycache__/awq.cpython-310.pyc +0 -0
  24. src/modules/__pycache__/exllama.cpython-310.pyc +0 -0
  25. src/modules/__pycache__/gptq.cpython-310.pyc +0 -0
  26. src/modules/__pycache__/xfastertransformer.cpython-310.pyc +0 -0
  27. src/serve/__pycache__/__init__.cpython-310.pyc +0 -0
  28. src/serve/__pycache__/api_provider.cpython-310.pyc +0 -0
  29. src/serve/__pycache__/gradio_block_arena_named.cpython-310.pyc +0 -0
  30. src/serve/__pycache__/gradio_block_arena_vision.cpython-310.pyc +0 -0
  31. src/serve/__pycache__/gradio_block_arena_vision_named.cpython-310.pyc +0 -0
  32. src/serve/__pycache__/gradio_web_server.cpython-310.pyc +0 -0
  33. src/serve/__pycache__/remote_logger.cpython-310.pyc +0 -0
  34. src/serve/gradio_web_server.py +6 -2
  35. vision-tmp-2024-07-10-conv.json +3 -0
.gitignore CHANGED
@@ -1 +1,6 @@
1
- dependency
 
 
 
 
 
 
1
+ dependency
2
+ *.json
3
+ *.log
4
+ *.pyc
5
+ *__pycache__*
6
+ *.json
app.py CHANGED
@@ -6,7 +6,6 @@ import gradio as gr
6
  from src.model.model_llava import inference
7
  from src.serve.gradio_block_arena_vision_named import build_side_by_side_vision_ui_named
8
 
9
- @spaces.GPU(duration=120)
10
  def main():
11
  with gr.Blocks() as demo:
12
  states = build_side_by_side_vision_ui_named(
 
6
  from src.model.model_llava import inference
7
  from src.serve.gradio_block_arena_vision_named import build_side_by_side_vision_ui_named
8
 
 
9
  def main():
10
  with gr.Blocks() as demo:
11
  states = build_side_by_side_vision_ui_named(
gradio_web_server.log CHANGED
@@ -254,3 +254,188 @@
254
  2024-07-10 12:01:17 | ERROR | stderr | File "/home/bofei-zhang/miniconda3/envs/fire_demo/lib/python3.10/threading.py", line 1116, in _wait_for_tstate_lock
255
  2024-07-10 12:01:17 | ERROR | stderr | if lock.acquire(block, timeout):
256
  2024-07-10 12:01:17 | ERROR | stderr | KeyboardInterrupt
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
254
  2024-07-10 12:01:17 | ERROR | stderr | File "/home/bofei-zhang/miniconda3/envs/fire_demo/lib/python3.10/threading.py", line 1116, in _wait_for_tstate_lock
255
  2024-07-10 12:01:17 | ERROR | stderr | if lock.acquire(block, timeout):
256
  2024-07-10 12:01:17 | ERROR | stderr | KeyboardInterrupt
257
+ 2024-07-10 07:18:04 | ERROR | stderr | /usr/local/lib/python3.10/site-packages/spaces/zero/wrappers.py:77: UserWarning: Using a ZeroGPU function outside of Gradio caching or request might block the app
258
+ 2024-07-10 07:18:04 | ERROR | stderr | warnings.warn("Using a ZeroGPU function outside of Gradio caching or request might block the app")
259
+ 2024-07-10 07:18:11 | INFO | stdout | Running on local URL: http://0.0.0.0:7860
260
+ 2024-07-10 07:18:11 | INFO | stdout |
261
+ 2024-07-10 07:18:11 | INFO | stdout | To create a public link, set `share=True` in `launch()`.
262
+ 2024-07-10 07:18:28 | ERROR | stderr | ERROR: Exception in ASGI application
263
+ 2024-07-10 07:18:28 | ERROR | stderr | Traceback (most recent call last):
264
+ 2024-07-10 07:18:28 | ERROR | stderr | File "/usr/local/lib/python3.10/site-packages/starlette/responses.py", line 265, in __call__
265
+ 2024-07-10 07:18:28 | ERROR | stderr | await wrap(partial(self.listen_for_disconnect, receive))
266
+ 2024-07-10 07:18:28 | ERROR | stderr | File "/usr/local/lib/python3.10/site-packages/starlette/responses.py", line 261, in wrap
267
+ 2024-07-10 07:18:28 | ERROR | stderr | await func()
268
+ 2024-07-10 07:18:28 | ERROR | stderr | File "/usr/local/lib/python3.10/site-packages/starlette/responses.py", line 238, in listen_for_disconnect
269
+ 2024-07-10 07:18:28 | ERROR | stderr | message = await receive()
270
+ 2024-07-10 07:18:28 | ERROR | stderr | File "/usr/local/lib/python3.10/site-packages/uvicorn/protocols/http/httptools_impl.py", line 553, in receive
271
+ 2024-07-10 07:18:28 | ERROR | stderr | await self.message_event.wait()
272
+ 2024-07-10 07:18:28 | ERROR | stderr | File "/usr/local/lib/python3.10/asyncio/locks.py", line 214, in wait
273
+ 2024-07-10 07:18:28 | ERROR | stderr | await fut
274
+ 2024-07-10 07:18:28 | ERROR | stderr | asyncio.exceptions.CancelledError: Cancelled by cancel scope 7ffa672fb370
275
+ 2024-07-10 07:18:28 | ERROR | stderr |
276
+ 2024-07-10 07:18:28 | ERROR | stderr | During handling of the above exception, another exception occurred:
277
+ 2024-07-10 07:18:28 | ERROR | stderr |
278
+ 2024-07-10 07:18:28 | ERROR | stderr | + Exception Group Traceback (most recent call last):
279
+ 2024-07-10 07:18:28 | ERROR | stderr | | File "/usr/local/lib/python3.10/site-packages/uvicorn/protocols/http/httptools_impl.py", line 399, in run_asgi
280
+ 2024-07-10 07:18:28 | ERROR | stderr | | result = await app( # type: ignore[func-returns-value]
281
+ 2024-07-10 07:18:28 | ERROR | stderr | | File "/usr/local/lib/python3.10/site-packages/uvicorn/middleware/proxy_headers.py", line 70, in __call__
282
+ 2024-07-10 07:18:28 | ERROR | stderr | | return await self.app(scope, receive, send)
283
+ 2024-07-10 07:18:28 | ERROR | stderr | | File "/usr/local/lib/python3.10/site-packages/fastapi/applications.py", line 1054, in __call__
284
+ 2024-07-10 07:18:28 | ERROR | stderr | | await super().__call__(scope, receive, send)
285
+ 2024-07-10 07:18:28 | ERROR | stderr | | File "/usr/local/lib/python3.10/site-packages/starlette/applications.py", line 123, in __call__
286
+ 2024-07-10 07:18:28 | ERROR | stderr | | await self.middleware_stack(scope, receive, send)
287
+ 2024-07-10 07:18:28 | ERROR | stderr | | File "/usr/local/lib/python3.10/site-packages/starlette/middleware/errors.py", line 186, in __call__
288
+ 2024-07-10 07:18:28 | ERROR | stderr | | raise exc
289
+ 2024-07-10 07:18:28 | ERROR | stderr | | File "/usr/local/lib/python3.10/site-packages/starlette/middleware/errors.py", line 164, in __call__
290
+ 2024-07-10 07:18:28 | ERROR | stderr | | await self.app(scope, receive, _send)
291
+ 2024-07-10 07:18:28 | ERROR | stderr | | File "/usr/local/lib/python3.10/site-packages/gradio/route_utils.py", line 714, in __call__
292
+ 2024-07-10 07:18:28 | ERROR | stderr | | await self.app(scope, receive, send)
293
+ 2024-07-10 07:18:28 | ERROR | stderr | | File "/usr/local/lib/python3.10/site-packages/starlette/middleware/exceptions.py", line 65, in __call__
294
+ 2024-07-10 07:18:28 | ERROR | stderr | | await wrap_app_handling_exceptions(self.app, conn)(scope, receive, send)
295
+ 2024-07-10 07:18:28 | ERROR | stderr | | File "/usr/local/lib/python3.10/site-packages/starlette/_exception_handler.py", line 64, in wrapped_app
296
+ 2024-07-10 07:18:28 | ERROR | stderr | | raise exc
297
+ 2024-07-10 07:18:28 | ERROR | stderr | | File "/usr/local/lib/python3.10/site-packages/starlette/_exception_handler.py", line 53, in wrapped_app
298
+ 2024-07-10 07:18:28 | ERROR | stderr | | await app(scope, receive, sender)
299
+ 2024-07-10 07:18:28 | ERROR | stderr | | File "/usr/local/lib/python3.10/site-packages/starlette/routing.py", line 756, in __call__
300
+ 2024-07-10 07:18:28 | ERROR | stderr | | await self.middleware_stack(scope, receive, send)
301
+ 2024-07-10 07:18:28 | ERROR | stderr | | File "/usr/local/lib/python3.10/site-packages/starlette/routing.py", line 776, in app
302
+ 2024-07-10 07:18:28 | ERROR | stderr | | await route.handle(scope, receive, send)
303
+ 2024-07-10 07:18:28 | ERROR | stderr | | File "/usr/local/lib/python3.10/site-packages/starlette/routing.py", line 297, in handle
304
+ 2024-07-10 07:18:28 | ERROR | stderr | | await self.app(scope, receive, send)
305
+ 2024-07-10 07:18:28 | ERROR | stderr | | File "/usr/local/lib/python3.10/site-packages/starlette/routing.py", line 77, in app
306
+ 2024-07-10 07:18:28 | ERROR | stderr | | await wrap_app_handling_exceptions(app, request)(scope, receive, send)
307
+ 2024-07-10 07:18:28 | ERROR | stderr | | File "/usr/local/lib/python3.10/site-packages/starlette/_exception_handler.py", line 64, in wrapped_app
308
+ 2024-07-10 07:18:28 | ERROR | stderr | | raise exc
309
+ 2024-07-10 07:18:28 | ERROR | stderr | | File "/usr/local/lib/python3.10/site-packages/starlette/_exception_handler.py", line 53, in wrapped_app
310
+ 2024-07-10 07:18:28 | ERROR | stderr | | await app(scope, receive, sender)
311
+ 2024-07-10 07:18:28 | ERROR | stderr | | File "/usr/local/lib/python3.10/site-packages/starlette/routing.py", line 75, in app
312
+ 2024-07-10 07:18:28 | ERROR | stderr | | await response(scope, receive, send)
313
+ 2024-07-10 07:18:28 | ERROR | stderr | | File "/usr/local/lib/python3.10/site-packages/starlette/responses.py", line 258, in __call__
314
+ 2024-07-10 07:18:28 | ERROR | stderr | | async with anyio.create_task_group() as task_group:
315
+ 2024-07-10 07:18:28 | ERROR | stderr | | File "/usr/local/lib/python3.10/site-packages/anyio/_backends/_asyncio.py", line 680, in __aexit__
316
+ 2024-07-10 07:18:28 | ERROR | stderr | | raise BaseExceptionGroup(
317
+ 2024-07-10 07:18:28 | ERROR | stderr | | exceptiongroup.ExceptionGroup: unhandled errors in a TaskGroup (1 sub-exception)
318
+ 2024-07-10 07:18:28 | ERROR | stderr | +-+---------------- 1 ----------------
319
+ 2024-07-10 07:18:28 | ERROR | stderr | | Traceback (most recent call last):
320
+ 2024-07-10 07:18:28 | ERROR | stderr | | File "/usr/local/lib/python3.10/site-packages/starlette/responses.py", line 261, in wrap
321
+ 2024-07-10 07:18:28 | ERROR | stderr | | await func()
322
+ 2024-07-10 07:18:28 | ERROR | stderr | | File "/usr/local/lib/python3.10/site-packages/starlette/responses.py", line 250, in stream_response
323
+ 2024-07-10 07:18:28 | ERROR | stderr | | async for chunk in self.body_iterator:
324
+ 2024-07-10 07:18:28 | ERROR | stderr | | File "/usr/local/lib/python3.10/site-packages/gradio/routes.py", line 670, in iterator
325
+ 2024-07-10 07:18:28 | ERROR | stderr | | done = [d.result() for d in done]
326
+ 2024-07-10 07:18:28 | ERROR | stderr | | File "/usr/local/lib/python3.10/site-packages/gradio/routes.py", line 670, in <listcomp>
327
+ 2024-07-10 07:18:28 | ERROR | stderr | | done = [d.result() for d in done]
328
+ 2024-07-10 07:18:28 | ERROR | stderr | | File "/usr/local/lib/python3.10/site-packages/gradio/routes.py", line 655, in stop_stream
329
+ 2024-07-10 07:18:28 | ERROR | stderr | | await app.stop_event.wait()
330
+ 2024-07-10 07:18:28 | ERROR | stderr | | AttributeError: 'NoneType' object has no attribute 'wait'
331
+ 2024-07-10 07:18:28 | ERROR | stderr | +------------------------------------
332
+ 2024-07-10 07:19:12 | ERROR | stderr | ERROR: Exception in ASGI application
333
+ 2024-07-10 07:19:12 | ERROR | stderr | Traceback (most recent call last):
334
+ 2024-07-10 07:19:12 | ERROR | stderr | File "/usr/local/lib/python3.10/site-packages/starlette/responses.py", line 265, in __call__
335
+ 2024-07-10 07:19:12 | ERROR | stderr | await wrap(partial(self.listen_for_disconnect, receive))
336
+ 2024-07-10 07:19:12 | ERROR | stderr | File "/usr/local/lib/python3.10/site-packages/starlette/responses.py", line 261, in wrap
337
+ 2024-07-10 07:19:12 | ERROR | stderr | await func()
338
+ 2024-07-10 07:19:12 | ERROR | stderr | File "/usr/local/lib/python3.10/site-packages/starlette/responses.py", line 238, in listen_for_disconnect
339
+ 2024-07-10 07:19:12 | ERROR | stderr | message = await receive()
340
+ 2024-07-10 07:19:12 | ERROR | stderr | File "/usr/local/lib/python3.10/site-packages/uvicorn/protocols/http/httptools_impl.py", line 553, in receive
341
+ 2024-07-10 07:19:12 | ERROR | stderr | await self.message_event.wait()
342
+ 2024-07-10 07:19:12 | ERROR | stderr | File "/usr/local/lib/python3.10/asyncio/locks.py", line 214, in wait
343
+ 2024-07-10 07:19:12 | ERROR | stderr | await fut
344
+ 2024-07-10 07:19:12 | ERROR | stderr | asyncio.exceptions.CancelledError: Cancelled by cancel scope 7ffa696ff760
345
+ 2024-07-10 07:19:12 | ERROR | stderr |
346
+ 2024-07-10 07:19:12 | ERROR | stderr | During handling of the above exception, another exception occurred:
347
+ 2024-07-10 07:19:12 | ERROR | stderr |
348
+ 2024-07-10 07:19:12 | ERROR | stderr | + Exception Group Traceback (most recent call last):
349
+ 2024-07-10 07:19:12 | ERROR | stderr | | File "/usr/local/lib/python3.10/site-packages/uvicorn/protocols/http/httptools_impl.py", line 399, in run_asgi
350
+ 2024-07-10 07:19:12 | ERROR | stderr | | result = await app( # type: ignore[func-returns-value]
351
+ 2024-07-10 07:19:12 | ERROR | stderr | | File "/usr/local/lib/python3.10/site-packages/uvicorn/middleware/proxy_headers.py", line 70, in __call__
352
+ 2024-07-10 07:19:12 | ERROR | stderr | | return await self.app(scope, receive, send)
353
+ 2024-07-10 07:19:12 | ERROR | stderr | | File "/usr/local/lib/python3.10/site-packages/fastapi/applications.py", line 1054, in __call__
354
+ 2024-07-10 07:19:12 | ERROR | stderr | | await super().__call__(scope, receive, send)
355
+ 2024-07-10 07:19:12 | ERROR | stderr | | File "/usr/local/lib/python3.10/site-packages/starlette/applications.py", line 123, in __call__
356
+ 2024-07-10 07:19:12 | ERROR | stderr | | await self.middleware_stack(scope, receive, send)
357
+ 2024-07-10 07:19:12 | ERROR | stderr | | File "/usr/local/lib/python3.10/site-packages/starlette/middleware/errors.py", line 186, in __call__
358
+ 2024-07-10 07:19:12 | ERROR | stderr | | raise exc
359
+ 2024-07-10 07:19:12 | ERROR | stderr | | File "/usr/local/lib/python3.10/site-packages/starlette/middleware/errors.py", line 164, in __call__
360
+ 2024-07-10 07:19:12 | ERROR | stderr | | await self.app(scope, receive, _send)
361
+ 2024-07-10 07:19:12 | ERROR | stderr | | File "/usr/local/lib/python3.10/site-packages/gradio/route_utils.py", line 714, in __call__
362
+ 2024-07-10 07:19:12 | ERROR | stderr | | await self.app(scope, receive, send)
363
+ 2024-07-10 07:19:12 | ERROR | stderr | | File "/usr/local/lib/python3.10/site-packages/starlette/middleware/exceptions.py", line 65, in __call__
364
+ 2024-07-10 07:19:12 | ERROR | stderr | | await wrap_app_handling_exceptions(self.app, conn)(scope, receive, send)
365
+ 2024-07-10 07:19:12 | ERROR | stderr | | File "/usr/local/lib/python3.10/site-packages/starlette/_exception_handler.py", line 64, in wrapped_app
366
+ 2024-07-10 07:19:12 | ERROR | stderr | | raise exc
367
+ 2024-07-10 07:19:12 | ERROR | stderr | | File "/usr/local/lib/python3.10/site-packages/starlette/_exception_handler.py", line 53, in wrapped_app
368
+ 2024-07-10 07:19:12 | ERROR | stderr | | await app(scope, receive, sender)
369
+ 2024-07-10 07:19:12 | ERROR | stderr | | File "/usr/local/lib/python3.10/site-packages/starlette/routing.py", line 756, in __call__
370
+ 2024-07-10 07:19:12 | ERROR | stderr | | await self.middleware_stack(scope, receive, send)
371
+ 2024-07-10 07:19:12 | ERROR | stderr | | File "/usr/local/lib/python3.10/site-packages/starlette/routing.py", line 776, in app
372
+ 2024-07-10 07:19:12 | ERROR | stderr | | await route.handle(scope, receive, send)
373
+ 2024-07-10 07:19:12 | ERROR | stderr | | File "/usr/local/lib/python3.10/site-packages/starlette/routing.py", line 297, in handle
374
+ 2024-07-10 07:19:12 | ERROR | stderr | | await self.app(scope, receive, send)
375
+ 2024-07-10 07:19:12 | ERROR | stderr | | File "/usr/local/lib/python3.10/site-packages/starlette/routing.py", line 77, in app
376
+ 2024-07-10 07:19:12 | ERROR | stderr | | await wrap_app_handling_exceptions(app, request)(scope, receive, send)
377
+ 2024-07-10 07:19:12 | ERROR | stderr | | File "/usr/local/lib/python3.10/site-packages/starlette/_exception_handler.py", line 64, in wrapped_app
378
+ 2024-07-10 07:19:12 | ERROR | stderr | | raise exc
379
+ 2024-07-10 07:19:12 | ERROR | stderr | | File "/usr/local/lib/python3.10/site-packages/starlette/_exception_handler.py", line 53, in wrapped_app
380
+ 2024-07-10 07:19:12 | ERROR | stderr | | await app(scope, receive, sender)
381
+ 2024-07-10 07:19:12 | ERROR | stderr | | File "/usr/local/lib/python3.10/site-packages/starlette/routing.py", line 75, in app
382
+ 2024-07-10 07:19:12 | ERROR | stderr | | await response(scope, receive, send)
383
+ 2024-07-10 07:19:12 | ERROR | stderr | | File "/usr/local/lib/python3.10/site-packages/starlette/responses.py", line 258, in __call__
384
+ 2024-07-10 07:19:12 | ERROR | stderr | | async with anyio.create_task_group() as task_group:
385
+ 2024-07-10 07:19:12 | ERROR | stderr | | File "/usr/local/lib/python3.10/site-packages/anyio/_backends/_asyncio.py", line 680, in __aexit__
386
+ 2024-07-10 07:19:12 | ERROR | stderr | | raise BaseExceptionGroup(
387
+ 2024-07-10 07:19:12 | ERROR | stderr | | exceptiongroup.ExceptionGroup: unhandled errors in a TaskGroup (1 sub-exception)
388
+ 2024-07-10 07:19:12 | ERROR | stderr | +-+---------------- 1 ----------------
389
+ 2024-07-10 07:19:12 | ERROR | stderr | | Traceback (most recent call last):
390
+ 2024-07-10 07:19:12 | ERROR | stderr | | File "/usr/local/lib/python3.10/site-packages/starlette/responses.py", line 261, in wrap
391
+ 2024-07-10 07:19:12 | ERROR | stderr | | await func()
392
+ 2024-07-10 07:19:12 | ERROR | stderr | | File "/usr/local/lib/python3.10/site-packages/starlette/responses.py", line 250, in stream_response
393
+ 2024-07-10 07:19:12 | ERROR | stderr | | async for chunk in self.body_iterator:
394
+ 2024-07-10 07:19:12 | ERROR | stderr | | File "/usr/local/lib/python3.10/site-packages/gradio/routes.py", line 670, in iterator
395
+ 2024-07-10 07:19:12 | ERROR | stderr | | done = [d.result() for d in done]
396
+ 2024-07-10 07:19:12 | ERROR | stderr | | File "/usr/local/lib/python3.10/site-packages/gradio/routes.py", line 670, in <listcomp>
397
+ 2024-07-10 07:19:12 | ERROR | stderr | | done = [d.result() for d in done]
398
+ 2024-07-10 07:19:12 | ERROR | stderr | | File "/usr/local/lib/python3.10/site-packages/gradio/routes.py", line 655, in stop_stream
399
+ 2024-07-10 07:19:12 | ERROR | stderr | | await app.stop_event.wait()
400
+ 2024-07-10 07:19:12 | ERROR | stderr | | AttributeError: 'NoneType' object has no attribute 'wait'
401
+ 2024-07-10 07:19:12 | ERROR | stderr | +------------------------------------
402
+ 2024-07-10 07:19:24 | ERROR | stderr | Traceback (most recent call last):
403
+ 2024-07-10 07:19:24 | ERROR | stderr | File "/home/user/app/app.py", line 18, in <module>
404
+ 2024-07-10 07:19:24 | ERROR | stderr | main()
405
+ 2024-07-10 07:19:24 | ERROR | stderr | File "/home/user/app/app.py", line 15, in main
406
+ 2024-07-10 07:19:24 | ERROR | stderr | demo.launch()
407
+ 2024-07-10 07:19:24 | ERROR | stderr | File "/usr/local/lib/python3.10/site-packages/gradio/blocks.py", line 2327, in launch
408
+ 2024-07-10 07:19:24 | ERROR | stderr | ) = http_server.start_server(
409
+ 2024-07-10 07:19:24 | ERROR | stderr | File "/usr/local/lib/python3.10/site-packages/gradio/http_server.py", line 154, in start_server
410
+ 2024-07-10 07:19:24 | ERROR | stderr | raise OSError(
411
+ 2024-07-10 07:19:24 | ERROR | stderr | OSError: Cannot find empty port in range: 7860-7860. You can specify a different port by setting the GRADIO_SERVER_PORT environment variable or passing the `server_port` parameter to `launch()`.
412
+ 2024-07-10 07:20:37 | INFO | stdout | Running on local URL: http://0.0.0.0:7860
413
+ 2024-07-10 07:20:37 | INFO | stdout |
414
+ 2024-07-10 07:20:37 | INFO | stdout | To create a public link, set `share=True` in `launch()`.
415
+ 2024-07-10 07:20:58 | INFO | gradio_web_server | bot_response. ip: 46.3.240.105
416
+ 2024-07-10 07:20:58 | INFO | gradio_web_server | monitor error: HTTPConnectionPool(host='localhost', port=9090): Max retries exceeded with url: /is_limit_reached?model=llava-fire&user_id=46.3.240.105 (Caused by NewConnectionError('<urllib3.connection.HTTPConnection object at 0x7f75d015d8d0>: Failed to establish a new connection: [Errno 111] Connection refused'))
417
+ 2024-07-10 07:20:58 | INFO | gradio_web_server | model_name: llava-fire;model_api_dict: None
418
+ 2024-07-10 07:20:58 | INFO | gradio_web_server | bot_response. ip: 46.3.240.105
419
+ 2024-07-10 07:20:58 | INFO | gradio_web_server | monitor error: HTTPConnectionPool(host='localhost', port=9090): Max retries exceeded with url: /is_limit_reached?model=llava-original&user_id=46.3.240.105 (Caused by NewConnectionError('<urllib3.connection.HTTPConnection object at 0x7f75d015cf70>: Failed to establish a new connection: [Errno 111] Connection refused'))
420
+ 2024-07-10 07:20:58 | INFO | gradio_web_server | model_name: llava-original;model_api_dict: None
421
+ 2024-07-10 07:20:58 | INFO | gradio_web_server | hello
422
+ 2024-07-10 07:20:58 | INFO | gradio_web_server | hello
423
+ 2024-07-10 07:23:46 | INFO | stdout | Running on local URL: http://0.0.0.0:7860
424
+ 2024-07-10 07:23:46 | INFO | stdout |
425
+ 2024-07-10 07:23:46 | INFO | stdout | To create a public link, set `share=True` in `launch()`.
426
+ 2024-07-10 07:24:07 | INFO | gradio_web_server | bot_response. ip: 46.3.240.105
427
+ 2024-07-10 07:24:07 | INFO | gradio_web_server | monitor error: HTTPConnectionPool(host='localhost', port=9090): Max retries exceeded with url: /is_limit_reached?model=llava-fire&user_id=46.3.240.105 (Caused by NewConnectionError('<urllib3.connection.HTTPConnection object at 0x7f36e964f100>: Failed to establish a new connection: [Errno 111] Connection refused'))
428
+ 2024-07-10 07:24:07 | INFO | gradio_web_server | model_name: llava-fire;model_api_dict: None
429
+ 2024-07-10 07:24:07 | INFO | gradio_web_server | bot_response. ip: 46.3.240.105
430
+ 2024-07-10 07:24:07 | INFO | gradio_web_server | monitor error: HTTPConnectionPool(host='localhost', port=9090): Max retries exceeded with url: /is_limit_reached?model=llava-original&user_id=46.3.240.105 (Caused by NewConnectionError('<urllib3.connection.HTTPConnection object at 0x7f36e964d4e0>: Failed to establish a new connection: [Errno 111] Connection refused'))
431
+ 2024-07-10 07:24:07 | INFO | gradio_web_server | model_name: llava-original;model_api_dict: None
432
+ 2024-07-10 07:24:15 | INFO | stdout | torch.Size([1, 58]) torch.Size([1, 5, 3, 336, 336])
433
+ 2024-07-10 07:24:16 | ERROR | stderr | /usr/local/lib/python3.10/site-packages/transformers/generation/configuration_utils.py:392: UserWarning: `do_sample` is set to `False`. However, `temperature` is set to `0` -- this flag is only used in sample-based generation modes. You should set `do_sample=True` or unset `temperature`.
434
+ 2024-07-10 07:24:16 | ERROR | stderr | warnings.warn(
435
+ 2024-07-10 07:24:16 | ERROR | stderr | /usr/local/lib/python3.10/site-packages/transformers/generation/configuration_utils.py:397: UserWarning: `do_sample` is set to `False`. However, `top_p` is set to `0.9` -- this flag is only used in sample-based generation modes. You should set `do_sample=True` or unset `top_p`.
436
+ 2024-07-10 07:24:16 | ERROR | stderr | warnings.warn(
437
+ 2024-07-10 07:24:22 | INFO | stdout | ["The figure in the image is a can of Södra Almighy, which is a beer. Södra is a Swedish brewery known for its lager beers. The can's design is modern and minimalist, with a color scheme that includes black and white, and the text is in English. The label indicates that it is a dry-hopped beer, which means it has been flavored with hops that have not been steeped in the brewing process, giving it a unique taste profile. The can's design suggests a contemporary and possibly craft beer, which is often associated with a more complex flavor profile than traditional lagers."]
438
+ 2024-07-10 07:24:22 | INFO | gradio_web_server | hello
439
+ 2024-07-10 07:24:48 | INFO | stdout | Running on local URL: http://0.0.0.0:7860
440
+ 2024-07-10 07:24:48 | INFO | stdout |
441
+ 2024-07-10 07:24:48 | INFO | stdout | To create a public link, set `share=True` in `launch()`.
gradio_web_server_multi.log CHANGED
@@ -8,3 +8,7 @@
8
  2024-07-10 11:58:24 | INFO | gradio_web_server_multi | bot_response_multi (named). ip: 127.0.0.1
9
  2024-07-10 11:58:46 | INFO | gradio_web_server_multi | add_text (named). ip: 127.0.0.1. len: 5
10
  2024-07-10 11:58:46 | INFO | gradio_web_server_multi | bot_response_multi (named). ip: 127.0.0.1
 
 
 
 
 
8
  2024-07-10 11:58:24 | INFO | gradio_web_server_multi | bot_response_multi (named). ip: 127.0.0.1
9
  2024-07-10 11:58:46 | INFO | gradio_web_server_multi | add_text (named). ip: 127.0.0.1. len: 5
10
  2024-07-10 11:58:46 | INFO | gradio_web_server_multi | bot_response_multi (named). ip: 127.0.0.1
11
+ 2024-07-10 07:20:57 | INFO | gradio_web_server_multi | add_text (named). ip: 46.3.240.105. len: 5
12
+ 2024-07-10 07:20:58 | INFO | gradio_web_server_multi | bot_response_multi (named). ip: 46.3.240.105
13
+ 2024-07-10 07:24:07 | INFO | gradio_web_server_multi | add_text (named). ip: 46.3.240.105. len: 5
14
+ 2024-07-10 07:24:07 | INFO | gradio_web_server_multi | bot_response_multi (named). ip: 46.3.240.105
src/__pycache__/__init__.cpython-310.pyc CHANGED
Binary files a/src/__pycache__/__init__.cpython-310.pyc and b/src/__pycache__/__init__.cpython-310.pyc differ
 
src/__pycache__/constants.cpython-310.pyc CHANGED
Binary files a/src/__pycache__/constants.cpython-310.pyc and b/src/__pycache__/constants.cpython-310.pyc differ
 
src/__pycache__/conversation.cpython-310.pyc CHANGED
Binary files a/src/__pycache__/conversation.cpython-310.pyc and b/src/__pycache__/conversation.cpython-310.pyc differ
 
src/__pycache__/utils.cpython-310.pyc CHANGED
Binary files a/src/__pycache__/utils.cpython-310.pyc and b/src/__pycache__/utils.cpython-310.pyc differ
 
src/model/__pycache__/__init__.cpython-310.pyc CHANGED
Binary files a/src/model/__pycache__/__init__.cpython-310.pyc and b/src/model/__pycache__/__init__.cpython-310.pyc differ
 
src/model/__pycache__/compression.cpython-310.pyc CHANGED
Binary files a/src/model/__pycache__/compression.cpython-310.pyc and b/src/model/__pycache__/compression.cpython-310.pyc differ
 
src/model/__pycache__/llama_condense_monkey_patch.cpython-310.pyc CHANGED
Binary files a/src/model/__pycache__/llama_condense_monkey_patch.cpython-310.pyc and b/src/model/__pycache__/llama_condense_monkey_patch.cpython-310.pyc differ
 
src/model/__pycache__/model_adapter.cpython-310.pyc CHANGED
Binary files a/src/model/__pycache__/model_adapter.cpython-310.pyc and b/src/model/__pycache__/model_adapter.cpython-310.pyc differ
 
src/model/__pycache__/model_chatglm.cpython-310.pyc CHANGED
Binary files a/src/model/__pycache__/model_chatglm.cpython-310.pyc and b/src/model/__pycache__/model_chatglm.cpython-310.pyc differ
 
src/model/__pycache__/model_cllm.cpython-310.pyc CHANGED
Binary files a/src/model/__pycache__/model_cllm.cpython-310.pyc and b/src/model/__pycache__/model_cllm.cpython-310.pyc differ
 
src/model/__pycache__/model_codet5p.cpython-310.pyc CHANGED
Binary files a/src/model/__pycache__/model_codet5p.cpython-310.pyc and b/src/model/__pycache__/model_codet5p.cpython-310.pyc differ
 
src/model/__pycache__/model_exllama.cpython-310.pyc CHANGED
Binary files a/src/model/__pycache__/model_exllama.cpython-310.pyc and b/src/model/__pycache__/model_exllama.cpython-310.pyc differ
 
src/model/__pycache__/model_falcon.cpython-310.pyc CHANGED
Binary files a/src/model/__pycache__/model_falcon.cpython-310.pyc and b/src/model/__pycache__/model_falcon.cpython-310.pyc differ
 
src/model/__pycache__/model_registry.cpython-310.pyc CHANGED
Binary files a/src/model/__pycache__/model_registry.cpython-310.pyc and b/src/model/__pycache__/model_registry.cpython-310.pyc differ
 
src/model/__pycache__/model_xfastertransformer.cpython-310.pyc CHANGED
Binary files a/src/model/__pycache__/model_xfastertransformer.cpython-310.pyc and b/src/model/__pycache__/model_xfastertransformer.cpython-310.pyc differ
 
src/model/__pycache__/model_yuan2.cpython-310.pyc CHANGED
Binary files a/src/model/__pycache__/model_yuan2.cpython-310.pyc and b/src/model/__pycache__/model_yuan2.cpython-310.pyc differ
 
src/model/__pycache__/monkey_patch_non_inplace.cpython-310.pyc CHANGED
Binary files a/src/model/__pycache__/monkey_patch_non_inplace.cpython-310.pyc and b/src/model/__pycache__/monkey_patch_non_inplace.cpython-310.pyc differ
 
src/modules/__pycache__/__init__.cpython-310.pyc CHANGED
Binary files a/src/modules/__pycache__/__init__.cpython-310.pyc and b/src/modules/__pycache__/__init__.cpython-310.pyc differ
 
src/modules/__pycache__/awq.cpython-310.pyc CHANGED
Binary files a/src/modules/__pycache__/awq.cpython-310.pyc and b/src/modules/__pycache__/awq.cpython-310.pyc differ
 
src/modules/__pycache__/exllama.cpython-310.pyc CHANGED
Binary files a/src/modules/__pycache__/exllama.cpython-310.pyc and b/src/modules/__pycache__/exllama.cpython-310.pyc differ
 
src/modules/__pycache__/gptq.cpython-310.pyc CHANGED
Binary files a/src/modules/__pycache__/gptq.cpython-310.pyc and b/src/modules/__pycache__/gptq.cpython-310.pyc differ
 
src/modules/__pycache__/xfastertransformer.cpython-310.pyc CHANGED
Binary files a/src/modules/__pycache__/xfastertransformer.cpython-310.pyc and b/src/modules/__pycache__/xfastertransformer.cpython-310.pyc differ
 
src/serve/__pycache__/__init__.cpython-310.pyc CHANGED
Binary files a/src/serve/__pycache__/__init__.cpython-310.pyc and b/src/serve/__pycache__/__init__.cpython-310.pyc differ
 
src/serve/__pycache__/api_provider.cpython-310.pyc CHANGED
Binary files a/src/serve/__pycache__/api_provider.cpython-310.pyc and b/src/serve/__pycache__/api_provider.cpython-310.pyc differ
 
src/serve/__pycache__/gradio_block_arena_named.cpython-310.pyc CHANGED
Binary files a/src/serve/__pycache__/gradio_block_arena_named.cpython-310.pyc and b/src/serve/__pycache__/gradio_block_arena_named.cpython-310.pyc differ
 
src/serve/__pycache__/gradio_block_arena_vision.cpython-310.pyc CHANGED
Binary files a/src/serve/__pycache__/gradio_block_arena_vision.cpython-310.pyc and b/src/serve/__pycache__/gradio_block_arena_vision.cpython-310.pyc differ
 
src/serve/__pycache__/gradio_block_arena_vision_named.cpython-310.pyc CHANGED
Binary files a/src/serve/__pycache__/gradio_block_arena_vision_named.cpython-310.pyc and b/src/serve/__pycache__/gradio_block_arena_vision_named.cpython-310.pyc differ
 
src/serve/__pycache__/gradio_web_server.cpython-310.pyc CHANGED
Binary files a/src/serve/__pycache__/gradio_web_server.cpython-310.pyc and b/src/serve/__pycache__/gradio_web_server.cpython-310.pyc differ
 
src/serve/__pycache__/remote_logger.cpython-310.pyc CHANGED
Binary files a/src/serve/__pycache__/remote_logger.cpython-310.pyc and b/src/serve/__pycache__/remote_logger.cpython-310.pyc differ
 
src/serve/gradio_web_server.py CHANGED
@@ -451,10 +451,14 @@ def bot_response(
451
  images = conv.get_images()
452
  logger.info(f"model_name: {model_name};model_api_dict: {model_api_dict}")
453
  if model_api_dict is None:
454
-
 
 
 
 
455
  stream_iter = [{
456
  "error_code": 0,
457
- "text": "hello"
458
  }]
459
  # Query worker address
460
  # ret = requests.post(
 
451
  images = conv.get_images()
452
  logger.info(f"model_name: {model_name};model_api_dict: {model_api_dict}")
453
  if model_api_dict is None:
454
+ if model_name == "llava-original":
455
+ from src.model.model_llava import inference
456
+ output_text = inference()[0]
457
+ else:
458
+ output_text = "hello"
459
  stream_iter = [{
460
  "error_code": 0,
461
+ "text": output_text
462
  }]
463
  # Query worker address
464
  # ret = requests.post(
vision-tmp-2024-07-10-conv.json CHANGED
@@ -1,2 +1,5 @@
1
  {"tstamp": 1720583926.6994, "type": "chat", "model": "llava-fire", "gen_params": {"temperature": 0.7, "top_p": 1.0, "max_new_tokens": 1024}, "start": 1720583926.678, "finish": 1720583926.6994, "state": {"template_name": "vicuna_v1.1", "system_message": "A chat between a curious user and an artificial intelligence assistant. The assistant gives helpful, detailed, and polite answers to the user's questions.", "roles": ["USER", "ASSISTANT"], "messages": [["USER", "Hello"], ["ASSISTANT", "hello"]], "offset": 0, "conv_id": "a48164214dce453a9ad276401b7d095c", "model_name": "llava-fire", "has_csam_image": false}, "ip": "127.0.0.1"}
2
  {"tstamp": 1720583926.6998, "type": "chat", "model": "llava-original", "gen_params": {"temperature": 0.7, "top_p": 1.0, "max_new_tokens": 1024}, "start": 1720583926.6815, "finish": 1720583926.6998, "state": {"template_name": "vicuna_v1.1", "system_message": "A chat between a curious user and an artificial intelligence assistant. The assistant gives helpful, detailed, and polite answers to the user's questions.", "roles": ["USER", "ASSISTANT"], "messages": [["USER", "Hello"], ["ASSISTANT", "hello"]], "offset": 0, "conv_id": "8bed262a5728409284a7a56a0fe66a75", "model_name": "llava-original", "has_csam_image": false}, "ip": "127.0.0.1"}
 
 
 
 
1
  {"tstamp": 1720583926.6994, "type": "chat", "model": "llava-fire", "gen_params": {"temperature": 0.7, "top_p": 1.0, "max_new_tokens": 1024}, "start": 1720583926.678, "finish": 1720583926.6994, "state": {"template_name": "vicuna_v1.1", "system_message": "A chat between a curious user and an artificial intelligence assistant. The assistant gives helpful, detailed, and polite answers to the user's questions.", "roles": ["USER", "ASSISTANT"], "messages": [["USER", "Hello"], ["ASSISTANT", "hello"]], "offset": 0, "conv_id": "a48164214dce453a9ad276401b7d095c", "model_name": "llava-fire", "has_csam_image": false}, "ip": "127.0.0.1"}
2
  {"tstamp": 1720583926.6998, "type": "chat", "model": "llava-original", "gen_params": {"temperature": 0.7, "top_p": 1.0, "max_new_tokens": 1024}, "start": 1720583926.6815, "finish": 1720583926.6998, "state": {"template_name": "vicuna_v1.1", "system_message": "A chat between a curious user and an artificial intelligence assistant. The assistant gives helpful, detailed, and polite answers to the user's questions.", "roles": ["USER", "ASSISTANT"], "messages": [["USER", "Hello"], ["ASSISTANT", "hello"]], "offset": 0, "conv_id": "8bed262a5728409284a7a56a0fe66a75", "model_name": "llava-original", "has_csam_image": false}, "ip": "127.0.0.1"}
3
+ {"tstamp": 1720588858.8939, "type": "chat", "model": "llava-fire", "gen_params": {"temperature": 0.7, "top_p": 1.0, "max_new_tokens": 1024}, "start": 1720588858.8843, "finish": 1720588858.8939, "state": {"template_name": "vicuna_v1.1", "system_message": "A chat between a curious user and an artificial intelligence assistant. The assistant gives helpful, detailed, and polite answers to the user's questions.", "roles": ["USER", "ASSISTANT"], "messages": [["USER", "Hello"], ["ASSISTANT", "hello"]], "offset": 0, "conv_id": "d994e3d859c94bddbf0dfcaed6c63079", "model_name": "llava-fire", "has_csam_image": false}, "ip": "46.3.240.105"}
4
+ {"tstamp": 1720588858.8951, "type": "chat", "model": "llava-original", "gen_params": {"temperature": 0.7, "top_p": 1.0, "max_new_tokens": 1024}, "start": 1720588858.8863, "finish": 1720588858.8951, "state": {"template_name": "vicuna_v1.1", "system_message": "A chat between a curious user and an artificial intelligence assistant. The assistant gives helpful, detailed, and polite answers to the user's questions.", "roles": ["USER", "ASSISTANT"], "messages": [["USER", "Hello"], ["ASSISTANT", "hello"]], "offset": 0, "conv_id": "e41b47f05f8b44ff9d520c8e94c6e8de", "model_name": "llava-original", "has_csam_image": false}, "ip": "46.3.240.105"}
5
+ {"tstamp": 1720589062.1758, "type": "chat", "model": "llava-fire", "gen_params": {"temperature": 0.7, "top_p": 1.0, "max_new_tokens": 1024}, "start": 1720589047.9171, "finish": 1720589062.1758, "state": {"template_name": "vicuna_v1.1", "system_message": "A chat between a curious user and an artificial intelligence assistant. The assistant gives helpful, detailed, and polite answers to the user's questions.", "roles": ["USER", "ASSISTANT"], "messages": [["USER", "Hello"], ["ASSISTANT", "hello"]], "offset": 0, "conv_id": "963f15cd5e224eb8ae02c67ed37b93c4", "model_name": "llava-fire", "has_csam_image": false}, "ip": "46.3.240.105"}