whyazalea commited on
Commit
629687a
·
1 Parent(s): 5464bfb

Upload folder using huggingface_hub

Browse files
.ipynb_checkpoints/web_demo-checkpoint.py CHANGED
@@ -3,8 +3,12 @@ import gradio as gr
3
  import mdtex2html
4
  from utils import load_model_on_gpus
5
 
6
- tokenizer = AutoTokenizer.from_pretrained(r"/mnt/workspace/chatglm-6b", trust_remote_code=True)
7
- model = AutoModel.from_pretrained(r"/mnt/workspace/chatglm-6b", trust_remote_code=True).half().cuda()
 
 
 
 
8
  # 多显卡支持,使用下面两行代替上面一行,将num_gpus改为你实际的显卡数量
9
  # from utils import load_model_on_gpus
10
  # model = load_model_on_gpus("THUDM/chatglm2-6b", num_gpus=2)
 
3
  import mdtex2html
4
  from utils import load_model_on_gpus
5
 
6
+
7
+ tokenizer = AutoTokenizer.from_pretrained(r"THUDM/chatglm2-6b", trust_remote_code=True)
8
+ model = AutoModel.from_pretrained(r"THUDM/chatglm2-6b", trust_remote_code=True).half().cuda()
9
+
10
+ # tokenizer = AutoTokenizer.from_pretrained(r"/mnt/workspace/chatglm-6b", trust_remote_code=True)
11
+ # model = AutoModel.from_pretrained(r"/mnt/workspace/chatglm-6b", trust_remote_code=True).half().cuda()
12
  # 多显卡支持,使用下面两行代替上面一行,将num_gpus改为你实际的显卡数量
13
  # from utils import load_model_on_gpus
14
  # model = load_model_on_gpus("THUDM/chatglm2-6b", num_gpus=2)
web_demo.py CHANGED
@@ -3,8 +3,12 @@ import gradio as gr
3
  import mdtex2html
4
  from utils import load_model_on_gpus
5
 
6
- tokenizer = AutoTokenizer.from_pretrained(r"/mnt/workspace/chatglm-6b", trust_remote_code=True)
7
- model = AutoModel.from_pretrained(r"/mnt/workspace/chatglm-6b", trust_remote_code=True).half().cuda()
 
 
 
 
8
  # 多显卡支持,使用下面两行代替上面一行,将num_gpus改为你实际的显卡数量
9
  # from utils import load_model_on_gpus
10
  # model = load_model_on_gpus("THUDM/chatglm2-6b", num_gpus=2)
 
3
  import mdtex2html
4
  from utils import load_model_on_gpus
5
 
6
+
7
+ tokenizer = AutoTokenizer.from_pretrained(r"THUDM/chatglm2-6b", trust_remote_code=True)
8
+ model = AutoModel.from_pretrained(r"THUDM/chatglm2-6b", trust_remote_code=True).half().cuda()
9
+
10
+ # tokenizer = AutoTokenizer.from_pretrained(r"/mnt/workspace/chatglm-6b", trust_remote_code=True)
11
+ # model = AutoModel.from_pretrained(r"/mnt/workspace/chatglm-6b", trust_remote_code=True).half().cuda()
12
  # 多显卡支持,使用下面两行代替上面一行,将num_gpus改为你实际的显卡数量
13
  # from utils import load_model_on_gpus
14
  # model = load_model_on_gpus("THUDM/chatglm2-6b", num_gpus=2)