File size: 14,847 Bytes
5472531
 
 
 
 
 
 
 
1
2
3
4
5
6
7
8
9
{"tstamp": 1709252309.8682, "type": "chat", "model": "MobiLlama-05B-Chat", "gen_params": {"temperature": 0.7, "top_p": 1.0, "max_new_tokens": 1024}, "start": 1709252307.0823, "finish": 1709252309.8682, "state": {"template_name": "zero_shot", "system_message": "A chat between a curious human and an artificial intelligence assistant. The assistant gives helpful, detailed, and polite answers to the human's questions.", "roles": ["Human", "Assistant"], "messages": [["Human", "Who is Prof Yi-Zhe Song? What is his research background? "], ["Assistant", "Prof Yi-Zhe Song is a Professor of Electrical Engineering and Computer Science at the University of Texas at Austin. He received his Ph.D. in Electrical Engineering from the University of California, Berkeley, in 2004. He is also a Senior Research Scientist at the University of Texas at Austin. Prof Song's research interests include signal processing, machine learning, and optimization. He has published over 300 papers in top-tier conferences and journals, including IEEE Transactions on Pattern Analysis and Machine Intelligence, IEEE Transactions on Signal Processing, and IEEE Transactions on Knowledge and Data Engineering. Prof Song is a recipient of numerous awards and grants, including the National Science Foundation Graduate Research Fellowship and the National Institutes of Health Young Investigator Award."]], "offset": 0, "conv_id": "d5bad730cd4b48bab48a80156b991ebd", "model_name": "MobiLlama-05B-Chat"}, "ip": "172.31.57.119"}
{"tstamp": 1709252330.9337, "type": "chat", "model": "MobiLlama-05B-Chat", "gen_params": {"temperature": 0.7, "top_p": 1.0, "max_new_tokens": 1024}, "start": 1709252330.4529, "finish": 1709252330.9337, "state": {"template_name": "zero_shot", "system_message": "A chat between a curious human and an artificial intelligence assistant. The assistant gives helpful, detailed, and polite answers to the human's questions.", "roles": ["Human", "Assistant"], "messages": [["Human", "can you give any link?"], ["Assistant", "Sure, here is the link: https://www.google.com/search?q=google+search+api"]], "offset": 0, "conv_id": "d5bad730cd4b48bab48a80156b991ebd", "model_name": "MobiLlama-05B-Chat"}, "ip": "172.31.57.119"}
{"tstamp": 1709252352.6014, "type": "chat", "model": "MobiLlama-05B-Chat", "gen_params": {"temperature": 0.7, "top_p": 1.0, "max_new_tokens": 1024}, "start": 1709252351.7288, "finish": 1709252352.6014, "state": {"template_name": "zero_shot", "system_message": "A chat between a curious human and an artificial intelligence assistant. The assistant gives helpful, detailed, and polite answers to the human's questions.", "roles": ["Human", "Assistant"], "messages": [["Human", "do you know Ayan Kumar Bhunia"], ["Assistant", "Yes, Ayan Kumar Bhunia is a famous Indian film actor, producer, and director. He is known for his role in the Bollywood film 'Dil Dosti Duniya Do' and 'Meri Aashiqui'."]], "offset": 0, "conv_id": "d5bad730cd4b48bab48a80156b991ebd", "model_name": "MobiLlama-05B-Chat"}, "ip": "172.31.57.119"}
{"tstamp": 1709261178.0713, "type": "chat", "model": "MobiLlama-05B-Chat", "gen_params": {"temperature": 0.7, "top_p": 1.0, "max_new_tokens": 1024}, "start": 1709261171.7866, "finish": 1709261178.0713, "state": {"template_name": "zero_shot", "system_message": "A chat between a curious human and an artificial intelligence assistant. The assistant gives helpful, detailed, and polite answers to the human's questions.", "roles": ["Human", "Assistant"], "messages": [["Human", "\u5c06\u4e0b\u9762\u6587\u4ef6\u7ffb\u8bd1\u6210\u4e2d\u6587 \u201cDescribe the bug\nI'm seeing 5x i/s improvements in Ooba when using the exLlama loading and was hopeful that it would help serve up some gains when using ooba's multimodal extension.\n\nWhen attempting to use exllama as the loader with any of the 4 stock Multimodal setups in either chat, instruct-chat or instruct modesl, text will work on its own but adding an image to the multimodal extension results in this error.\n\nMaybe you can point me in the right direction to try and resolve this?\n\nIs there an existing issue for this?\n[X] I have searched the existing issues\nReproduction\nDownload Multimodal models Install miniGPT if necessary Install exLlama repository (now a part of the latest one click isntall)\n\nScreenshot\nNo response\n\nLogs\nFile \"D:\\00\\text-generation-webui\\modules\\text_generation.py\", line 300, in generate_reply_custom\n    for reply in shared.model.generate_with_streaming(question, state):\n  File \"D:\\00\\text-generation-webui\\modules\\exllama.py\", line 68, in generate_with_streaming\n    self.generator.gen_begin_reuse(ids)\n  File \"D:\\00\\text-generation-webui\\repositories\\exllama\\generator.py\", line 191, in gen_begin_reuse\n    if reuse < in_tokens.shape[-1]: self.gen_feed_tokens(in_tokens[:, reuse:])\n  File \"D:\\00\\text-generation-webui\\repositories\\exllama\\generator.py\", line 209, in gen_feed_tokens\n    self.model.forward(self.sequence[:, start:-1], self.cache, preprocess_only = True, lora = self.lora)\n  File \"D:\\00\\text-generation-webui\\repositories\\exllama\\model.py\", line 841, in forward\n    hidden_states = decoder_layer.forward(hidden_states, cache, buffers[device], lora)\n  File \"D:\\00\\text-generation-webui\\repositories\\exllama\\model.py\", line 459, in forward\n    hidden_states = self.self_attn.forward(hidden_states, cache, buffer, lora)\n  File \"D:\\00\\text-generation-webui\\repositories\\exllama\\model.py\", line 381, in forward\n    new_keys = cache.key_states[self.index].narrow(2, past_len, q_len)\nRuntimeError: start (49) + length (13970) exceeds dimension size (2048).\nOutput generated in 1.51 seconds (0.00 tokens/s, 0 tokens, context 14020, seed 979644525)\nSystem Info\nWindows 10, 96 GB RAM, Nvidia GTX 2080i 11GB\nAlso confirmed same issue on\nWindows 11, Titan RTX 24 GB\u201d"], ["Assistant", "\u5728\u4e0b\u9762\u6587\u4ef6\u4e2d\u7ffb\u8bd1\u6210\u4e2d\u6587 \u201cDescribe the bug\u201d\n\n\u6587\u4ef6\uff1a\u201cDescribe the bug\u201d\n\nIssue\uff1a\n\n1. \u4f7f\u7528 ooba \u7684 multimodal \u6a21\u578b\u65f6\uff0c\u5982\u679c\u5728\u4f7f\u7528 ooba \u7684 multimodal \u6a21\u578b\u65f6\uff0c\u4f7f\u7528 ooba \u7684 exLlama \u6a21\u578b\uff0c\u5c06\u4f1a\u51fa\u73b0\u4ee5\u4e0b\u9519\u8bef\uff1a\n\n```\nFile \"D:\\00\\text-generation-webui\\modules\\text_generation.py\", line 300, in generate_reply_custom\n    for reply in shared.model.generate_with_streaming(question, state):\n  File \"D:\\00\\text-generation-webui\\modules\\exllama.py\", line 68, in generate_with_streaming\n    self.generator.gen_begin_reuse(ids)\n  File \"D:\\00\\text-generation-webui\\repositories\\exllama\\generator.py\", line 191, in gen_begin_reuse\n    if reuse < in_tokens.shape[-1]: self.gen_feed_tokens(in_tokens[:, reuse:])\n  File \"D:\\00\\text-generation-webui\\repositories\\exllama\\generator.py\", line 209, in gen_feed_tokens\n    if reuse < in_tokens.shape[-1]: self.gen_feed_tokens(in_tokens[:, reuse:])\n  File \"D:\\00\\text-generation-webui\\repositories\\exllama\\generator.py\", line 209, in gen"]], "offset": 0, "conv_id": "bff45589528e4269a2e8ae0d31c801bd", "model_name": "MobiLlama-05B-Chat"}, "ip": "172.31.18.237"}
{"tstamp": 1709261335.2558, "type": "chat", "model": "MobiLlama-05B-Chat", "gen_params": {"temperature": 0.7, "top_p": 1.0, "max_new_tokens": 1024}, "start": 1709261332.0782, "finish": 1709261335.2558, "state": {"template_name": "zero_shot", "system_message": "A chat between a curious human and an artificial intelligence assistant. The assistant gives helpful, detailed, and polite answers to the human's questions.", "roles": ["Human", "Assistant"], "messages": [["Human", "\u7ffb\u8bd1\u6210\u4e2d\u6587\u201cDescribe the bug\nI'm seeing 5x i/s improvements in Ooba when using the exLlama loading and was hopeful that it would help serve up some gains when using ooba's multimodal extension.\n\nWhen attempting to use exllama as the loader with any of the 4 stock Multimodal setups in either chat, instruct-chat or instruct modesl, text will work on its own but adding an image to the multimodal extension results in this error.\n\nMaybe you can point me in the right direction to try and resolve this?\n\nIs there an existing issue for this?\n[X] I have searched the existing issues\nReproduction\nDownload Multimodal models Install miniGPT if necessary Install exLlama repository (now a part of the latest one click isntall)\n\nScreenshot\nNo response\u201d"], ["Assistant", "\u7ffb\u8bd1\u6210\u4e2d\u6587\u201cI'm seeing 5x i/s improvements in Ooba when using the exLlama loading and was hopeful that it would help serve up some gains when using ooba's multimodal extension.\n\nWhen attempting to use exllama as the loader with any of the 4 stock Multimodal setups in either chat, instruct-chat or instruct modesl, text will work on its own but adding an image to the multimodal extension results in this error.\n\nMaybe you can point me in the right direction to try and resolve this?\n\nIs there an existing issue for this?\n[X] I have searched the existing issues\nReproduction\nDownload Multimodal models Install miniGPT if necessary Install exLlama repository (now a part of the latest one click isntall)\n\nScreenshot\nNo response\u201d"]], "offset": 0, "conv_id": "bff45589528e4269a2e8ae0d31c801bd", "model_name": "MobiLlama-05B-Chat"}, "ip": "172.31.18.237"}
{"tstamp": 1709261369.0236, "type": "chat", "model": "MobiLlama-05B-Chat", "gen_params": {"temperature": 0.7, "top_p": 1.0, "max_new_tokens": 1024}, "start": 1709261362.7107, "finish": 1709261369.0236, "state": {"template_name": "zero_shot", "system_message": "A chat between a curious human and an artificial intelligence assistant. The assistant gives helpful, detailed, and polite answers to the human's questions.", "roles": ["Human", "Assistant"], "messages": [["Human", "\u5c06\u4e0b\u9762\u53cc\u5f15\u53f7\u4e2d\u7684\u6587\u5b57\u7ffb\u8bd1\u6210\u4e2d\u6587\u201cDescribe the bug\nI'm seeing 5x i/s improvements in Ooba when using the exLlama loading and was hopeful that it would help serve up some gains when using ooba's multimodal extension.\n\nWhen attempting to use exllama as the loader with any of the 4 stock Multimodal setups in either chat, instruct-chat or instruct modesl, text will work on its own but adding an image to the multimodal extension results in this error.\n\nMaybe you can point me in the right direction to try and resolve this?\n\nIs there an existing issue for this?\n[X] I have searched the existing issues\nReproduction\nDownload Multimodal models Install miniGPT if necessary Install exLlama repository (now a part of the latest one click isntall)\n\nScreenshot\nNo response\u201d"], ["Assistant", "\u6211\u53ef\u4ee5\u4e3a\u4f60\u63d0\u4f9b\u8be6\u7ec6\u7684\u89e3\u51b3\u65b9\u6848\u3002\n\u5982\u679c\u4f60\u662f\u5426\u6709\u4efb\u4f55\u95ee\u9898\uff0c\u8bf7\u95ee\u4f60\u6709\u600e\u4e48\u89e3\u51b3\u8fd9\u4e2a\u95ee\u9898\uff1f\n[X] I have searched the existing issues\n[ ] I have found an existing issue\n[ ] I have searched for a solution to this problem\n[ ] I have found a solution to this problem\n[ ] I have found a solution to this problem\n[ ] I have found a solution to this problem\n[ ] I have found a solution to this problem\n[ ] I have found a solution to this problem\n[ ] I have found a solution to this problem\n[ ] I have found a solution to this problem\n[ ] I have found a solution to this problem\n[ ] I have found a solution to this problem\n[ ] I have found a solution to this problem\n[ ] I have found a solution to this problem\n[ ] I have found a solution to this problem\n[ ] I have found a solution to this problem\n[ ] I have found a solution to this problem\n[ ] I have found a solution to this problem\n[ ] I have found a solution to this problem\n[ ] I have found a solution to this problem\n[ ] I have found a solution to this problem\n[ ] I have found a solution to this problem\n[ ] I have found a solution to this problem\n[ ] I have found a solution to this problem\n[ ] I have found a solution to this problem\n[ ] I have found a solution to this problem\n[ ] I have found a solution to this problem\n[ ] I have found a solution to this problem\n[ ] I have found a solution to this problem\n[ ] I have found a solution to this problem\n[ ] I have found a solution to this problem\n[ ] I have found"]], "offset": 0, "conv_id": "bff45589528e4269a2e8ae0d31c801bd", "model_name": "MobiLlama-05B-Chat"}, "ip": "172.31.18.237"}
{"tstamp": 1709261390.198, "type": "chat", "model": "MobiLlama-05B-Chat", "gen_params": {"temperature": 0.7, "top_p": 1.0, "max_new_tokens": 1024}, "start": 1709261389.3716, "finish": 1709261390.198, "state": {"template_name": "zero_shot", "system_message": "A chat between a curious human and an artificial intelligence assistant. The assistant gives helpful, detailed, and polite answers to the human's questions.", "roles": ["Human", "Assistant"], "messages": [["Human", "\u5c06\u4e0b\u9762\u53cc\u5f15\u53f7\u4e2d\u7684\u6587\u5b57\u7ffb\u8bd1\u6210\u4e2d\u6587\u201cDescribe the bug\nI'm seeing 5x i/s improvements in Ooba when using the exLlama loading and was hopeful that it would help serve up some gains when using ooba's multimodal extension.\n\nWhen attempting to use exllama as the loader with any of the 4 stock Multimodal setups in either chat, instruct-chat or instruct modesl, text will work on its own but adding an image to the multimodal extension results in this error.\n\nMaybe you can point me in the right direction to try and resolve this?\n\nIs there an existing issue for this?\n[X] I have searched the existing issues\nReproduction\nDownload Multimodal models Install miniGPT if necessary Install exLlama repository (now a part of the latest one click isntall)\n\nScreenshot\nNo response\u201d"], ["Assistant", "I'm sorry, but I cannot provide a response as I do not have access to your account or any other relevant information. Can you please provide me with the information you mentioned so that I can assist you further?"]], "offset": 0, "conv_id": "bff45589528e4269a2e8ae0d31c801bd", "model_name": "MobiLlama-05B-Chat"}, "ip": "172.31.18.237"}
{"tstamp": 1709279504.3264, "type": "chat", "model": "MobiLlama-1B-Chat", "gen_params": {"temperature": 0.7, "top_p": 1.0, "max_new_tokens": 1024}, "start": 1709279504.0812, "finish": 1709279504.3264, "state": {"template_name": "zero_shot", "system_message": "A chat between a curious human and an artificial intelligence assistant. The assistant gives helpful, detailed, and polite answers to the human's questions.", "roles": ["Human", "Assistant"], "messages": [["Human", "hello"], ["Assistant", "Hello! How can I assist you today?"]], "offset": 0, "conv_id": "c8a58b9ebd9e42e7b5b2e01cd648829d", "model_name": "MobiLlama-1B-Chat"}, "ip": "172.31.38.44"}