peichao.dong
commited on
Commit
•
3772cf0
1
Parent(s):
5c65101
update embedding config
Browse files- .gitignore +1 -0
- documents/abstract.faiss/index.faiss +0 -0
- documents/abstract.faiss/index.pkl +2 -2
- documents/business_context.faiss/index.faiss +0 -0
- documents/business_context.faiss/index.pkl +0 -0
- documents/bussiness_context/business_context.md +3 -0
- embedding.py +1 -1
- promopts.py +1 -0
- requirements.txt +3 -1
.gitignore
CHANGED
@@ -1,3 +1,4 @@
|
|
1 |
__pycache__/*
|
2 |
**/__pycache__/*
|
3 |
.DS_Store
|
|
|
|
1 |
__pycache__/*
|
2 |
**/__pycache__/*
|
3 |
.DS_Store
|
4 |
+
.idea/*
|
documents/abstract.faiss/index.faiss
CHANGED
Binary files a/documents/abstract.faiss/index.faiss and b/documents/abstract.faiss/index.faiss differ
|
|
documents/abstract.faiss/index.pkl
CHANGED
@@ -1,3 +1,3 @@
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
-
oid sha256:
|
3 |
-
size
|
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:2b5614785dff26ba4c3cb191b8aef6c88a7e90fc50f0414df5285863e58dd833
|
3 |
+
size 58026
|
documents/business_context.faiss/index.faiss
CHANGED
Binary files a/documents/business_context.faiss/index.faiss and b/documents/business_context.faiss/index.faiss differ
|
|
documents/business_context.faiss/index.pkl
CHANGED
Binary files a/documents/business_context.faiss/index.pkl and b/documents/business_context.faiss/index.pkl differ
|
|
documents/bussiness_context/business_context.md
CHANGED
@@ -14,3 +14,6 @@ FeatureConfig 用于配置某个 Feature 中控制前端展示效果的配置项
|
|
14 |
|
15 |
用户白名单圈人条件需要上传用户id的白名单,仅在白名单里的用户可以获取到相关feature
|
16 |
地理位置配置端需要设置圈定地区的地理位置编号列表,客户端请求接口是传递地理位置编号参数,位置编号匹配的数据用户可见
|
|
|
|
|
|
|
|
14 |
|
15 |
用户白名单圈人条件需要上传用户id的白名单,仅在白名单里的用户可以获取到相关feature
|
16 |
地理位置配置端需要设置圈定地区的地理位置编号列表,客户端请求接口是传递地理位置编号参数,位置编号匹配的数据用户可见
|
17 |
+
|
18 |
+
|
19 |
+
新增实验需要提供实验名称、目标、分组信息(包括分组标识、描述、比例)
|
embedding.py
CHANGED
@@ -48,7 +48,7 @@ class CustomEmbedding:
|
|
48 |
CONDENSE_QUESTION_PROMPT = PromptTemplate.from_template(_template)
|
49 |
question_generator = LLMChain(llm=llm, prompt=CONDENSE_QUESTION_PROMPT)
|
50 |
|
51 |
-
doc_chain = load_qa_chain(llm, chain_type="
|
52 |
qa = ConversationalRetrievalChain( retriever= docsearch.as_retriever(search_kwargs={"k": 1}),
|
53 |
question_generator=question_generator,
|
54 |
combine_docs_chain=doc_chain,
|
|
|
48 |
CONDENSE_QUESTION_PROMPT = PromptTemplate.from_template(_template)
|
49 |
question_generator = LLMChain(llm=llm, prompt=CONDENSE_QUESTION_PROMPT)
|
50 |
|
51 |
+
doc_chain = load_qa_chain(llm, chain_type="stuff")
|
52 |
qa = ConversationalRetrievalChain( retriever= docsearch.as_retriever(search_kwargs={"k": 1}),
|
53 |
question_generator=question_generator,
|
54 |
combine_docs_chain=doc_chain,
|
promopts.py
CHANGED
@@ -19,6 +19,7 @@ Answer: the answer I responded to the question
|
|
19 |
Thought: I know enough to explain the user story
|
20 |
Scenarios: List all possible scenarios with concrete example in Given/When/Then style
|
21 |
|
|
|
22 |
|
23 |
{history}
|
24 |
Answer:{input}"""
|
|
|
19 |
Thought: I know enough to explain the user story
|
20 |
Scenarios: List all possible scenarios with concrete example in Given/When/Then style
|
21 |
|
22 |
+
Please use Chinese! Begin!
|
23 |
|
24 |
{history}
|
25 |
Answer:{input}"""
|
requirements.txt
CHANGED
@@ -6,4 +6,6 @@ sentence_transformers
|
|
6 |
gradio_tools
|
7 |
faiss-cpu
|
8 |
tiktoken
|
9 |
-
pygpt4all
|
|
|
|
|
|
6 |
gradio_tools
|
7 |
faiss-cpu
|
8 |
tiktoken
|
9 |
+
pygpt4all
|
10 |
+
chromadb
|
11 |
+
unstructured
|