ian commited on
Commit
b999014
1 Parent(s): cbc9a17

update model

Browse files
Files changed (2) hide show
  1. app.py +8 -0
  2. flowsettings.py +12 -0
app.py CHANGED
@@ -1,5 +1,13 @@
 
1
  from ktem.main import App
2
 
 
 
 
 
 
 
 
3
  app = App()
4
  demo = app.make()
5
  demo.launch()
 
1
+ from huggingface_hub import hf_hub_download, snapshot_download
2
  from ktem.main import App
3
 
4
+ # preload model
5
+ snapshot_download("mixedbread-ai/mxbai-embed-large-v1")
6
+ hf_hub_download(
7
+ repo_id="Qwen/Qwen1.5-7B-Chat-GGUF",
8
+ filename="qwen1_5-7b-chat-q5_k_m.gguf",
9
+ )
10
+
11
  app = App()
12
  demo = app.make()
13
  demo.launch()
flowsettings.py CHANGED
@@ -124,6 +124,18 @@ if len(KH_EMBEDDINGS) < 1:
124
  "default": True,
125
  }
126
 
 
 
 
 
 
 
 
 
 
 
 
 
127
  KH_REASONINGS = ["ktem.reasoning.simple.FullQAPipeline"]
128
  KH_VLM_ENDPOINT = "{0}/openai/deployments/{1}/chat/completions?api-version={2}".format(
129
  config("AZURE_OPENAI_ENDPOINT", default=""),
 
124
  "default": True,
125
  }
126
 
127
+
128
+ KH_LLMS["qwen1.5-7B"] = {
129
+ "spec": {
130
+ "__type__": "kotaemon.llms.LlamaCppChat",
131
+ "repo_id": "Qwen/Qwen1.5-7B-Chat-GGUF",
132
+ "filename": "qwen1_5-7b-chat-q5_k_m.gguf",
133
+ },
134
+ "default": False,
135
+ "cost": 0,
136
+ }
137
+
138
+
139
  KH_REASONINGS = ["ktem.reasoning.simple.FullQAPipeline"]
140
  KH_VLM_ENDPOINT = "{0}/openai/deployments/{1}/chat/completions?api-version={2}".format(
141
  config("AZURE_OPENAI_ENDPOINT", default=""),