import sys import os sys.path.append(os.path.dirname(os.path.abspath(__file__)) + '/../../') import asyncio from argparse import Namespace from models.loader.args import parser from models.loader import LoaderCheckPoint import models.shared as shared async def dispatch(args: Namespace): args_dict = vars(args) shared.loaderCheckPoint = LoaderCheckPoint(args_dict) llm_model_ins = shared.loaderLLM() history = [ ("which city is this?", "tokyo"), ("why?", "she's japanese"), ] for answer_result in llm_model_ins.generatorAnswer(prompt="你好? ", history=history, streaming=False): resp = answer_result.llm_output["answer"] print(resp) if __name__ == '__main__': args = None args = parser.parse_args(args=['--model-dir', '/media/checkpoint/', '--model', 'fastchat-chatglm-6b', '--no-remote-model']) loop = asyncio.new_event_loop() asyncio.set_event_loop(loop) loop.run_until_complete(dispatch(args))