LovnishVerma commited on
Commit
5130d64
1 Parent(s): e755e5c

Update app.py

Browse files
Files changed (1) hide show
  1. app.py +5 -34
app.py CHANGED
@@ -1,37 +1,8 @@
1
- # Install necessary libraries
2
- # !pip install transformers
3
 
4
- from transformers import AutoTokenizer, AutoModelForCausalLM, pipeline
5
- import torch
6
-
7
- # Model and prompt details
8
  model_name = "mlabonne/llama-2-7b-guanaco"
9
- prompt = "What is a large language model?"
10
-
11
- # Load tokenizer and model
12
- tokenizer = AutoTokenizer.from_pretrained(model_name)
13
- model = AutoModelForCausalLM.from_pretrained(
14
- model_name,
15
- torch_dtype=torch.float16,
16
- device_map="auto",
17
- offload_folder="offload/folder" # Replace with the path to the offload folder
18
- )
19
-
20
- # Generate text using the provided prompt
21
- sequences = pipeline(
22
- "text-generation",
23
- model=model,
24
- tokenizer=tokenizer,
25
- device=0, # Change to the appropriate device index or "cuda" if using GPU
26
- )(
27
- f'<s>[INST] {prompt} [/INST]',
28
- do_sample=True,
29
- top_k=10,
30
- num_return_sequences=1,
31
- eos_token_id=tokenizer.eos_token_id,
32
- max_length=200,
33
- )
34
 
35
- # Print the generated text
36
- for seq in sequences:
37
- print(f"Generated Text: {seq['generated_text']}")
 
 
1
+ from transformers import AutoModelForCausalLM, AutoTokenizer
 
2
 
 
 
 
 
3
  model_name = "mlabonne/llama-2-7b-guanaco"
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
4
 
5
+ try:
6
+ model = AutoModelForCausalLM.from_pretrained(model_name)
7
+ except Exception as e:
8
+ print(f"Error loading the model: {e}")