neuraldevx commited on
Commit
efada3f
1 Parent(s): 10c7bc9

Update app.py

Browse files
Files changed (1) hide show
  1. app.py +10 -5
app.py CHANGED
@@ -5,10 +5,12 @@ import torch
5
  # Initialize the model and tokenizer
6
  model_id = "meta-llama/Meta-Llama-3-8B"
7
 
8
- @st.cache(allow_output_mutation=True)
9
  def load_model():
10
  tokenizer = AutoTokenizer.from_pretrained(model_id)
11
  model = AutoModelForCausalLM.from_pretrained(model_id, torch_dtype=torch.bfloat16)
 
 
12
  return pipeline("text-generation", model=model, tokenizer=tokenizer, device_map="auto")
13
 
14
  pipe = load_model()
@@ -22,8 +24,11 @@ user_input = st.text_area("Enter your query about stock market investments:", ""
22
  if st.button("Get Insights"):
23
  if user_input:
24
  with st.spinner("Generating insights..."):
25
- result = pipe(user_input, max_length=100, num_return_sequences=1)
26
- st.success("Here are your insights:")
27
- st.write(result[0]['generated_text'])
 
 
 
28
  else:
29
- st.error("Please enter a query.")
 
5
  # Initialize the model and tokenizer
6
  model_id = "meta-llama/Meta-Llama-3-8B"
7
 
8
+ @st.cache_resource
9
  def load_model():
10
  tokenizer = AutoTokenizer.from_pretrained(model_id)
11
  model = AutoModelForCausalLM.from_pretrained(model_id, torch_dtype=torch.bfloat16)
12
+ tokenizer.add_special_tokens({"pad_token": "<pad>"})
13
+ model.resize_token_embeddings(len(tokenizer))
14
  return pipeline("text-generation", model=model, tokenizer=tokenizer, device_map="auto")
15
 
16
  pipe = load_model()
 
24
  if st.button("Get Insights"):
25
  if user_input:
26
  with st.spinner("Generating insights..."):
27
+ try:
28
+ result = pipe(user_input, max_length=100, num_return_sequences=1, truncation=True, pad_token_id=tokenizer.pad_token_id)
29
+ st.success("Here are your insights:")
30
+ st.write(result[0]['generated_text'])
31
+ except Exception as e:
32
+ st.error(f"An error occurred: {e}")
33
  else:
34
+ st.error("Please enter a query.")