eswardivi commited on
Commit
ac6fdaf
1 Parent(s): ad891da

Update app.py

Browse files
Files changed (1) hide show
  1. app.py +3 -2
app.py CHANGED
@@ -18,9 +18,9 @@ quantization_config = BitsAndBytesConfig(
18
  )
19
 
20
  model = AutoModelForCausalLM.from_pretrained(
21
- "google/gemma-1.1-7b-it", quantization_config=quantization_config, token=token
22
  )
23
- tok = AutoTokenizer.from_pretrained("google/gemma-1.1-7b-it", token=token)
24
 
25
  if torch.cuda.is_available():
26
  device = torch.device("cuda")
@@ -104,5 +104,6 @@ demo = gr.ChatInterface(
104
  ],
105
  stop_btn="Stop Generation",
106
  title="Chat With LLMs",
 
107
  )
108
  demo.launch()
 
18
  )
19
 
20
  model = AutoModelForCausalLM.from_pretrained(
21
+ "meta-llama/Meta-Llama-3-70B-Instruct", quantization_config=quantization_config, token=token
22
  )
23
+ tok = AutoTokenizer.from_pretrained("meta-llama/Meta-Llama-3-70B-Instruct", token=token)
24
 
25
  if torch.cuda.is_available():
26
  device = torch.device("cuda")
 
104
  ],
105
  stop_btn="Stop Generation",
106
  title="Chat With LLMs",
107
+ description="Now Running ![meta-llama/Meta-Llama-3-70B-Instruct](https://huggingface.co/meta-llama/Meta-Llama-3-70B-Instruct) in 4bit"
108
  )
109
  demo.launch()