Upto12forenglish commited on
Commit
ddaaa7d
1 Parent(s): 79ecda9

Update app.py

Browse files
Files changed (1) hide show
  1. app.py +8 -10
app.py CHANGED
@@ -2,19 +2,17 @@ import os
2
  HF_TOKEN = os.getenv('HF_TOKEN')
3
  print("Token loaded")
4
 
5
- from transformers import GemmaTokenizer, AutoModelForCausalLM
6
- from transformers import AutoModelForCausalLM, AutoTokenizer, TextIteratorStreamer
7
- from threading import Thread
8
  import transformers
9
  import torch
10
 
11
- # Load the tokenizer and model
12
- tokenizer = AutoTokenizer.from_pretrained("meta-llama/Meta-Llama-3-8B-Instruct")
13
- model = AutoModelForCausalLM.from_pretrained("meta-llama/Meta-Llama-3-8B-Instruct", device_map="auto") # to("cuda:0")
14
- terminators = [
15
- tokenizer.eos_token_id,
16
- tokenizer.convert_tokens_to_ids("<|eot_id|>")
17
- ]
 
18
 
19
  print("llama download successfully")
20
 
 
2
  HF_TOKEN = os.getenv('HF_TOKEN')
3
  print("Token loaded")
4
 
 
 
 
5
  import transformers
6
  import torch
7
 
8
+ model_id = "meta-llama/Meta-Llama-3-8B-Instruct"
9
+
10
+ pipeline = transformers.pipeline(
11
+ "text-generation",
12
+ model="meta-llama/Meta-Llama-3-8B-Instruct",
13
+ model_kwargs={"torch_dtype": torch.bfloat16},
14
+ device="cuda",
15
+ )
16
 
17
  print("llama download successfully")
18