Tuana commited on
Commit
e0b936d
1 Parent(s): a5b24dd

fix caching and change model

Browse files
Files changed (2) hide show
  1. app.py +1 -1
  2. utils/haystack.py +2 -2
app.py CHANGED
@@ -25,7 +25,7 @@ if st.session_state.get("model") == None:
25
  mistral, openai = st.columns([2,1], gap="small")
26
 
27
  with mistral:
28
- st.button("Mixtral-8x7B-Instruct", on_click=mistral_pressed, type="primary")
29
 
30
  with openai:
31
  st.button("GPT-4", on_click=openai_pressed, type="primary")
 
25
  mistral, openai = st.columns([2,1], gap="small")
26
 
27
  with mistral:
28
+ st.button("Mistral-7B-Instruct-v0.2", on_click=mistral_pressed, type="primary")
29
 
30
  with openai:
31
  st.button("GPT-4", on_click=openai_pressed, type="primary")
utils/haystack.py CHANGED
@@ -4,6 +4,7 @@ from haystack.components.builders.prompt_builder import PromptBuilder
4
  from haystack.components.generators import HuggingFaceTGIGenerator, OpenAIGenerator
5
  from .hackernews_fetcher import HackernewsFetcher
6
 
 
7
  def start_haystack(key, model):
8
  prompt_template = """
9
  You will be provided one or more top HakcerNews posts, followed by their URL.
@@ -19,7 +20,7 @@ Summaries:
19
 
20
  prompt_builder = PromptBuilder(template=prompt_template)
21
  if model == "Mistral":
22
- llm = HuggingFaceTGIGenerator("mistralai/Mixtral-8x7B-Instruct-v0.1", token=key)
23
  elif model == "GPT-4":
24
  llm = OpenAIGenerator(api_key=key, model="gpt-4")
25
  fetcher = HackernewsFetcher()
@@ -34,7 +35,6 @@ Summaries:
34
  return pipe
35
 
36
 
37
- @st.cache_data(show_spinner=True)
38
  def query(top_k, _pipeline):
39
  try:
40
  run_args = {"hackernews_fetcher": {"top_k": top_k}}
 
4
  from haystack.components.generators import HuggingFaceTGIGenerator, OpenAIGenerator
5
  from .hackernews_fetcher import HackernewsFetcher
6
 
7
+ st.cache_resource
8
  def start_haystack(key, model):
9
  prompt_template = """
10
  You will be provided one or more top HakcerNews posts, followed by their URL.
 
20
 
21
  prompt_builder = PromptBuilder(template=prompt_template)
22
  if model == "Mistral":
23
+ llm = HuggingFaceTGIGenerator("mistralai/Mistral-7B-Instruct-v0.2", token=key)
24
  elif model == "GPT-4":
25
  llm = OpenAIGenerator(api_key=key, model="gpt-4")
26
  fetcher = HackernewsFetcher()
 
35
  return pipe
36
 
37
 
 
38
  def query(top_k, _pipeline):
39
  try:
40
  run_args = {"hackernews_fetcher": {"top_k": top_k}}