Spaces:
Sleeping
Sleeping
| # """ | |
| # 对各个llm模型进行单元测试 | |
| # """ | |
| def validate_path(): | |
| import os, sys | |
| dir_name = os.path.dirname(__file__) | |
| root_dir_assume = os.path.abspath(os.path.dirname(__file__) + '/..') | |
| os.chdir(root_dir_assume) | |
| sys.path.append(root_dir_assume) | |
| validate_path() # validate path so you can run from base directory | |
| from request_llm.bridge_moss import predict_no_ui_long_connection | |
| # from request_llm.bridge_jittorllms_pangualpha import predict_no_ui_long_connection | |
| # from request_llm.bridge_jittorllms_llama import predict_no_ui_long_connection | |
| llm_kwargs = { | |
| 'max_length': 512, | |
| 'top_p': 1, | |
| 'temperature': 1, | |
| } | |
| result = predict_no_ui_long_connection(inputs="你好", | |
| llm_kwargs=llm_kwargs, | |
| history=[], | |
| sys_prompt="") | |
| print('final result:', result) | |
| result = predict_no_ui_long_connection(inputs="what is a hero?", | |
| llm_kwargs=llm_kwargs, | |
| history=["hello world"], | |
| sys_prompt="") | |
| print('final result:', result) | |
| result = predict_no_ui_long_connection(inputs="如何理解传奇?", | |
| llm_kwargs=llm_kwargs, | |
| history=[], | |
| sys_prompt="") | |
| print('final result:', result) | |
| # # print(result) | |
| # from multiprocessing import Process, Pipe | |
| # class GetGLMHandle(Process): | |
| # def __init__(self): | |
| # super().__init__(daemon=True) | |
| # pass | |
| # def run(self): | |
| # # 子进程执行 | |
| # # 第一次运行,加载参数 | |
| # def validate_path(): | |
| # import os, sys | |
| # dir_name = os.path.dirname(__file__) | |
| # root_dir_assume = os.path.abspath(os.path.dirname(__file__) + '/..') | |
| # os.chdir(root_dir_assume + '/request_llm/jittorllms') | |
| # sys.path.append(root_dir_assume + '/request_llm/jittorllms') | |
| # validate_path() # validate path so you can run from base directory | |
| # jittorllms_model = None | |
| # import types | |
| # try: | |
| # if jittorllms_model is None: | |
| # from models import get_model | |
| # # availabel_models = ["chatglm", "pangualpha", "llama", "chatrwkv"] | |
| # args_dict = {'model': 'chatrwkv'} | |
| # print('self.jittorllms_model = get_model(types.SimpleNamespace(**args_dict))') | |
| # jittorllms_model = get_model(types.SimpleNamespace(**args_dict)) | |
| # print('done get model') | |
| # except: | |
| # # self.child.send('[Local Message] Call jittorllms fail 不能正常加载jittorllms的参数。') | |
| # raise RuntimeError("不能正常加载jittorllms的参数!") | |
| # x = GetGLMHandle() | |
| # x.start() | |
| # input() |