Matt Wallace commited on
Commit
a130843
1 Parent(s): 993a943

trying gguf for cpu inf

Browse files
Files changed (1) hide show
  1. app.py +2 -2
app.py CHANGED
@@ -1,11 +1,11 @@
1
  import gradio as gr
2
  from transformers import AutoModelForCausalLM, AutoTokenizer, pipeline
3
 
4
- model_name_or_path = "TheBloke/Llama-2-13B-chat-GPTQ"
5
  model = AutoModelForCausalLM.from_pretrained(model_name_or_path,
6
  device_map="auto",
7
  trust_remote_code=False,
8
- revision="gptq-4bit-32g-actorder_True")
9
 
10
  tokenizer = AutoTokenizer.from_pretrained(model_name_or_path, use_fast=True)
11
 
 
1
  import gradio as gr
2
  from transformers import AutoModelForCausalLM, AutoTokenizer, pipeline
3
 
4
+ model_name_or_path = "TheBloke/Llama-2-13B-chat-GGUF"
5
  model = AutoModelForCausalLM.from_pretrained(model_name_or_path,
6
  device_map="auto",
7
  trust_remote_code=False,
8
+ revision="llama-2-13b-chat.Q4_K_M.gguf")
9
 
10
  tokenizer = AutoTokenizer.from_pretrained(model_name_or_path, use_fast=True)
11