zayed-upal commited on
Commit
12991e6
1 Parent(s): fcb0ce9

updated models, added multi-threaded generation

Browse files
Files changed (2) hide show
  1. LLMHelper.py +6 -10
  2. app.py +6 -2
LLMHelper.py CHANGED
@@ -13,15 +13,10 @@ AVAILABLE_MODELS_GGUF = {
13
  "model_file": "mistral-7b-instruct-v0.2.Q4_K_M.gguf",
14
  "model_type": "mistral"
15
  },
16
- "TheBloke/zephyr-7B-beta-GGUF": {
17
- "model_file": "zephyr-7b-beta.Q4_K_M.gguf",
18
- "model_type": "zephyr"
19
- },
20
- "TheBloke/una-cybertron-7B-v2-GGUF": {
21
- "model_file": "una-cybertron-7b-v2-bf16.Q4_K_M.gguf",
22
- "model_type": "cybertron"
23
- },
24
-
25
  }
26
 
27
  AVAILABLE_MODELS_OPENAI = [
@@ -55,7 +50,8 @@ def generate_cover_letter_open_source(job_description, resume, selected_model, c
55
  reset=True,
56
  stream=True,
57
  # top_k=2,
58
- temperature=0.5
 
59
  )
60
 
61
  llm_response = llm(prompt)
 
13
  "model_file": "mistral-7b-instruct-v0.2.Q4_K_M.gguf",
14
  "model_type": "mistral"
15
  },
16
+ "TheBloke/LeoScorpius-7B-GGUF": {
17
+ "model_file": "leoscorpius-7b.Q4_K_M.gguf",
18
+ "model_type": "leoscorpius"
19
+ }
 
 
 
 
 
20
  }
21
 
22
  AVAILABLE_MODELS_OPENAI = [
 
50
  reset=True,
51
  stream=True,
52
  # top_k=2,
53
+ temperature=0.5,
54
+ threads=(os.cpu_count()//2)
55
  )
56
 
57
  llm_response = llm(prompt)
app.py CHANGED
@@ -44,14 +44,18 @@ def generate_response(cover_letter_gen, start_time):
44
  generated_text_placeholder.write(st.session_state.cover_letter_stream)
45
  st.write(f"generated words: {len(st.session_state.cover_letter_stream.split())}")
46
  st.write(f"generation time: {round(time.time() - start_time, 2)} seconds")
 
 
47
 
48
 
49
  if 'running' not in st.session_state:
50
  st.session_state.running = False
51
 
52
  st.session_state.cover_letter_stream = ""
53
- st.set_page_config(page_title='Cover Letter Generator', layout="wide")
54
- st.markdown("## Cover Letter Generator")
 
 
55
  info = st.expander("Information")
56
  info.write(f"This project aims to:\n"
57
  f"- Explore various open-source Large Language Models (LLMs).\n"
 
44
  generated_text_placeholder.write(st.session_state.cover_letter_stream)
45
  st.write(f"generated words: {len(st.session_state.cover_letter_stream.split())}")
46
  st.write(f"generation time: {round(time.time() - start_time, 2)} seconds")
47
+ st.write(
48
+ f"tokens per second: {round(len(st.session_state.cover_letter_stream.split())/(round(time.time() - start_time, 2)))} seconds")
49
 
50
 
51
  if 'running' not in st.session_state:
52
  st.session_state.running = False
53
 
54
  st.session_state.cover_letter_stream = ""
55
+ st.set_page_config(page_title='LLM Cover Letter Generator', layout="wide")
56
+ st.markdown("## Cover Letter Generator using Large Language Models (LLM)")
57
+ st.info("Please be patient with the open source LLM models, as they are running on a CPU on the server.\n "
58
+ "Average generation time around 5 mins.")
59
  info = st.expander("Information")
60
  info.write(f"This project aims to:\n"
61
  f"- Explore various open-source Large Language Models (LLMs).\n"