Sean-Case commited on
Commit
c2ff47a
1 Parent(s): 7339026

Put sources behind accordion, minor change to model params. Gitignore update

Browse files
Files changed (3) hide show
  1. .gitignore +3 -1
  2. app.py +5 -3
  3. chatfuncs/chatfuncs.py +2 -3
.gitignore CHANGED
@@ -3,6 +3,8 @@
3
  *.pdf
4
  *.spec
5
  *.toc
 
6
  bootstrapper.py
7
  build/*
8
- dist/*
 
 
3
  *.pdf
4
  *.spec
5
  *.toc
6
+ *.csv
7
  bootstrapper.py
8
  build/*
9
+ dist/*
10
+ Q tests/*
app.py CHANGED
@@ -95,11 +95,13 @@ def load_model(model_type, gpu_layers, gpu_config=None, cpu_config=None, torch_d
95
  #model = AutoModelForCausalLM.from_pretrained('Aryanne/Orca-Mini-3B-gguf', model_type='llama', model_file='q5_0-orca-mini-3b.gguf', **vars(gpu_config)) # **asdict(CtransRunConfig_cpu())
96
  #model = AutoModelForCausalLM.from_pretrained('Aryanne/Wizard-Orca-3B-gguf', model_type='llama', model_file='q4_1-wizard-orca-3b.gguf', **vars(gpu_config)) # **asdict(CtransRunConfig_cpu())
97
  model = AutoModelForCausalLM.from_pretrained('TheBloke/Mistral-7B-OpenOrca-GGUF', model_type='mistral', model_file='mistral-7b-openorca.Q4_K_M.gguf', **vars(gpu_config)) # **asdict(CtransRunConfig_cpu())
 
98
 
99
  except:
100
  #model = AutoModelForCausalLM.from_pretrained('Aryanne/Orca-Mini-3B-gguf', model_type='llama', model_file='q5_0-orca-mini-3b.gguf', **vars(cpu_config)) #**asdict(CtransRunConfig_gpu())
101
  #model = AutoModelForCausalLM.from_pretrained('Aryanne/Wizard-Orca-3B-gguf', model_type='llama', model_file='q4_1-wizard-orca-3b.gguf', **vars(cpu_config)) # **asdict(CtransRunConfig_cpu())
102
  model = AutoModelForCausalLM.from_pretrained('TheBloke/Mistral-7B-OpenOrca-GGUF', model_type='mistral', model_file='mistral-7b-openorca.Q4_K_M.gguf', **vars(cpu_config)) # **asdict(CtransRunConfig_cpu())
 
103
 
104
  tokenizer = []
105
 
@@ -192,12 +194,12 @@ with block:
192
  with gr.Row():
193
  chat_height = 500
194
  chatbot = gr.Chatbot(height=chat_height, avatar_images=('user.jfif', 'bot.jpg'),bubble_full_width = False, scale = 1)
195
- #sources = gr.HTML(value = "Source paragraphs with the most relevant text will appear here", height=chat_height, scale = 2)
196
- sources = gr.Markdown(value = "Source paragraphs with the most relevant text will appear here", height=chat_height, scale = 2)
197
 
198
  with gr.Row():
199
  message = gr.Textbox(
200
- label="Enter your question here.",
201
  lines=1,
202
  )
203
  with gr.Row():
 
95
  #model = AutoModelForCausalLM.from_pretrained('Aryanne/Orca-Mini-3B-gguf', model_type='llama', model_file='q5_0-orca-mini-3b.gguf', **vars(gpu_config)) # **asdict(CtransRunConfig_cpu())
96
  #model = AutoModelForCausalLM.from_pretrained('Aryanne/Wizard-Orca-3B-gguf', model_type='llama', model_file='q4_1-wizard-orca-3b.gguf', **vars(gpu_config)) # **asdict(CtransRunConfig_cpu())
97
  model = AutoModelForCausalLM.from_pretrained('TheBloke/Mistral-7B-OpenOrca-GGUF', model_type='mistral', model_file='mistral-7b-openorca.Q4_K_M.gguf', **vars(gpu_config)) # **asdict(CtransRunConfig_cpu())
98
+ #model = AutoModelForCausalLM.from_pretrained('TheBloke/MistralLite-7B-GGUF', model_type='mistral', model_file='mistrallite.Q4_K_M.gguf', **vars(gpu_config)) # **asdict(CtransRunConfig_cpu())
99
 
100
  except:
101
  #model = AutoModelForCausalLM.from_pretrained('Aryanne/Orca-Mini-3B-gguf', model_type='llama', model_file='q5_0-orca-mini-3b.gguf', **vars(cpu_config)) #**asdict(CtransRunConfig_gpu())
102
  #model = AutoModelForCausalLM.from_pretrained('Aryanne/Wizard-Orca-3B-gguf', model_type='llama', model_file='q4_1-wizard-orca-3b.gguf', **vars(cpu_config)) # **asdict(CtransRunConfig_cpu())
103
  model = AutoModelForCausalLM.from_pretrained('TheBloke/Mistral-7B-OpenOrca-GGUF', model_type='mistral', model_file='mistral-7b-openorca.Q4_K_M.gguf', **vars(cpu_config)) # **asdict(CtransRunConfig_cpu())
104
+ #model = AutoModelForCausalLM.from_pretrained('TheBloke/MistralLite-7B-GGUF', model_type='mistral', model_file='mistrallite.Q4_K_M.gguf', **vars(cpu_config)) # **asdict(CtransRunConfig_cpu())
105
 
106
  tokenizer = []
107
 
 
194
  with gr.Row():
195
  chat_height = 500
196
  chatbot = gr.Chatbot(height=chat_height, avatar_images=('user.jfif', 'bot.jpg'),bubble_full_width = False, scale = 1)
197
+ with gr.Accordion("Open this tab to see the source paragraphs used to generate the answer", open = False):
198
+ sources = gr.HTML(value = "Source paragraphs with the most relevant text will appear here", height=chat_height, scale = 2)
199
 
200
  with gr.Row():
201
  message = gr.Textbox(
202
+ label="Enter your question here",
203
  lines=1,
204
  )
205
  with gr.Row():
chatfuncs/chatfuncs.py CHANGED
@@ -84,15 +84,14 @@ top_k: int = 3
84
  top_p: float = 1
85
  repetition_penalty: float = 1.3
86
  flan_alpaca_repetition_penalty: float = 1.3
87
- tinyllama_repetition_penalty: float = 1.5
88
  last_n_tokens: int = 64
89
- max_new_tokens: int = 512
90
  seed: int = 42
91
  reset: bool = False
92
  stream: bool = True
93
  threads: int = threads
94
  batch_size:int = 256
95
- context_length:int = 4096
96
  sample = True
97
 
98
 
 
84
  top_p: float = 1
85
  repetition_penalty: float = 1.3
86
  flan_alpaca_repetition_penalty: float = 1.3
 
87
  last_n_tokens: int = 64
88
+ max_new_tokens: int = 256
89
  seed: int = 42
90
  reset: bool = False
91
  stream: bool = True
92
  threads: int = threads
93
  batch_size:int = 256
94
+ context_length:int = 2048
95
  sample = True
96
 
97