charanhu commited on
Commit
31f279f
β€’
1 Parent(s): 567a560

Update app.py

Browse files
Files changed (1) hide show
  1. app.py +2 -9
app.py CHANGED
@@ -2,15 +2,8 @@ import gradio as gr
2
  import torch
3
  from transformers import AutoTokenizer, AutoModelForCausalLM, AutoConfig
4
 
5
- tokenizer = AutoTokenizer.from_pretrained("TheBloke/SOLAR-10.7B-Instruct-v1.0-uncensored-GPTQ")
6
- config = AutoConfig.from_pretrained("TheBloke/SOLAR-10.7B-Instruct-v1.0-uncensored-GPTQ", trust_remote_code=True)
7
- # # config.quantization_config["use_exllama"] = True
8
- # config.quantization_config["disable_exllama"] = True
9
- # config.quantization_config["exllama_config"] = {"version":2}
10
- # # model = AutoModelForCausalLM.from_config(config, trust_remote_code=True)
11
- config = AutoConfig.from_pretrained("TheBloke/SOLAR-10.7B-Instruct-v1.0-uncensored-GPTQ")
12
- config.quantization_config["disable_exllama"] = True
13
- model = AutoModelForCausalLM.from_config(config)
14
 
15
  def generate_response(prompt):
16
  conversation = [{'role': 'user', 'content': prompt}]
 
2
  import torch
3
  from transformers import AutoTokenizer, AutoModelForCausalLM, AutoConfig
4
 
5
+ tokenizer = AutoTokenizer.from_pretrained("kimnt93/chat-llama2-1b-1.0")
6
+ model = AutoModelForCausalLM.from_pretrained("kimnt93/chat-llama2-1b-1.0")
 
 
 
 
 
 
 
7
 
8
  def generate_response(prompt):
9
  conversation = [{'role': 'user', 'content': prompt}]