Spaces:
Sleeping
Sleeping
File size: 6,022 Bytes
ba9f995 |
1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 16 17 18 19 20 21 22 23 24 25 26 27 28 29 30 31 32 33 34 35 36 37 38 39 40 41 42 43 44 45 46 47 48 49 50 51 52 53 54 55 56 57 58 59 60 61 62 63 64 65 66 67 68 69 70 71 72 73 74 75 76 77 78 79 80 81 82 83 84 85 86 87 88 89 90 91 92 93 94 95 96 97 98 99 100 101 102 103 104 105 106 107 108 109 110 111 112 113 114 115 116 117 118 119 120 121 122 123 124 125 126 127 128 129 130 131 132 133 134 135 136 137 138 139 140 141 142 143 144 145 146 147 148 149 150 151 152 153 154 155 156 157 158 159 160 161 162 163 164 165 166 167 168 169 |
[tokens]
; Enter one/all of your API key here.
; E.g., OPENAI_API_KEY = sk-xxxxxxx
OPENAI_API_KEY = sk-proj-2JwvyIn7WoKlkbjPOYVWT3BlbkFJnGAk65YAzvPH6cEVQXmr
ANTHROPIC_API_KEY = xxxxx
TOGETHER_API_KEY = xxxxx
; if you use Meta-Llama models, you may need Huggingface token to access.
HUGGINGFACE_TOKEN = xxxxx
VERSION = 1.0.1
[directory]
; Directory for source files.
DOCS_DIR = ./data
; Directory to store embeddings and Langchain documents.
DB_DIR = ./database_store
LOCAL_MODEL_DIR = ./models
; The below parameters are optional to modify:
; --------------------------------------------
[parameters]
; Model name schema: Model Provider|Model Name|Model File. Model File is only valid for GGUF format, set None for other format.
; For example:
; OpenAI|gpt-3.5-turbo|None
; OpenAI|gpt-4|None
; Anthropic|claude-2.0|None
; Together|togethercomputer/llama-2-70b-chat|None
; HuggingFace|TheBloke/Llama-2-70B-chat-GGUF|llama-2-70b-chat.q4_K_M.gguf
; HuggingFace|meta-llama/Llama-2-70b-chat-hf|None
; The full Together.AI model list can be found in the end of this file; We currently only support quantized gguf and the full huggingface local LLMs.
MODEL_NAME = OpenAI|gpt-4-1106-preview|None
; LLM temperature
TEMPURATURE = 0
; Maximum tokens for storing chat history.
MAX_CHAT_HISTORY = 800
; Maximum tokens for LLM context for retrieved information.
MAX_LLM_CONTEXT = 1200
; Maximum tokens for LLM generation.
MAX_LLM_GENERATION = 1000
; Supported embeddings: openAIEmbeddings and hkunlpInstructorLarge.
EMBEDDING_NAME = openAIEmbeddings
; This is dependent on your GPU type.
N_GPU_LAYERS = 100
; this is depend on your GPU and CPU ram when using open source LLMs.
N_BATCH = 512
; The base (small) chunk size for first stage document retrieval.
BASE_CHUNK_SIZE = 100
; Set to 0 for no overlap.
CHUNK_OVERLAP = 0
; The final retrieval (medium) chunk size will be BASE_CHUNK_SIZE * CHUNK_SCALE.
CHUNK_SCALE = 3
WINDOW_STEPS = 3
; The # tokens of window chunk will be BASE_CHUNK_SIZE * WINDOW_SCALE.
WINDOW_SCALE = 18
; Ratio of BM25 retriever to Chroma Vectorstore retriever.
RETRIEVER_WEIGHTS = 0.5, 0.5
; Number of retrieved chunks will range from FIRST_RETRIEVAL_K to 2*FIRST_RETRIEVAL_K due to the ensemble retriever.
FIRST_RETRIEVAL_K = 3
; Number of retrieved chunks will range from SECOND_RETRIEVAL_K to 2*SECOND_RETRIEVAL_K due to the ensemble retriever.
SECOND_RETRIEVAL_K = 3
; Number of windows (large chunks) for the third retriever.
NUM_WINDOWS = 2
; (The third retrieval gets the final chunks passed to the LLM QA chain. The 'k' value is dynamic (based on MAX_LLM_CONTEXT), depending on the number of rephrased questions and retrieved documents.)
[logging]
; If you do not want to enable logging, set enabled to False.
enabled = True
level = INFO
filename = IncarnaMind.log
format = %(asctime)s [%(levelname)s] %(name)s: %(message)s
; Together.AI supported models:
; 0 Austism/chronos-hermes-13b
; 1 EleutherAI/pythia-12b-v0
; 2 EleutherAI/pythia-1b-v0
; 3 EleutherAI/pythia-2.8b-v0
; 4 EleutherAI/pythia-6.9b
; 5 Gryphe/MythoMax-L2-13b
; 6 HuggingFaceH4/starchat-alpha
; 7 NousResearch/Nous-Hermes-13b
; 8 NousResearch/Nous-Hermes-Llama2-13b
; 9 NumbersStation/nsql-llama-2-7B
; 10 OpenAssistant/llama2-70b-oasst-sft-v10
; 11 OpenAssistant/oasst-sft-4-pythia-12b-epoch-3.5
; 12 OpenAssistant/stablelm-7b-sft-v7-epoch-3
; 13 Phind/Phind-CodeLlama-34B-Python-v1
; 14 Phind/Phind-CodeLlama-34B-v2
; 15 SG161222/Realistic_Vision_V3.0_VAE
; 16 WizardLM/WizardCoder-15B-V1.0
; 17 WizardLM/WizardCoder-Python-34B-V1.0
; 18 WizardLM/WizardLM-70B-V1.0
; 19 bigcode/starcoder
; 20 databricks/dolly-v2-12b
; 21 databricks/dolly-v2-3b
; 22 databricks/dolly-v2-7b
; 23 defog/sqlcoder
; 24 garage-bAInd/Platypus2-70B-instruct
; 25 huggyllama/llama-13b
; 26 huggyllama/llama-30b
; 27 huggyllama/llama-65b
; 28 huggyllama/llama-7b
; 29 lmsys/fastchat-t5-3b-v1.0
; 30 lmsys/vicuna-13b-v1.3
; 31 lmsys/vicuna-13b-v1.5-16k
; 32 lmsys/vicuna-13b-v1.5
; 33 lmsys/vicuna-7b-v1.3
; 34 prompthero/openjourney
; 35 runwayml/stable-diffusion-v1-5
; 36 stabilityai/stable-diffusion-2-1
; 37 stabilityai/stable-diffusion-xl-base-1.0
; 38 togethercomputer/CodeLlama-13b-Instruct
; 39 togethercomputer/CodeLlama-13b-Python
; 40 togethercomputer/CodeLlama-13b
; 41 togethercomputer/CodeLlama-34b-Instruct
; 42 togethercomputer/CodeLlama-34b-Python
; 43 togethercomputer/CodeLlama-34b
; 44 togethercomputer/CodeLlama-7b-Instruct
; 45 togethercomputer/CodeLlama-7b-Python
; 46 togethercomputer/CodeLlama-7b
; 47 togethercomputer/GPT-JT-6B-v1
; 48 togethercomputer/GPT-JT-Moderation-6B
; 49 togethercomputer/GPT-NeoXT-Chat-Base-20B
; 50 togethercomputer/Koala-13B
; 51 togethercomputer/LLaMA-2-7B-32K
; 52 togethercomputer/Llama-2-7B-32K-Instruct
; 53 togethercomputer/Pythia-Chat-Base-7B-v0.16
; 54 togethercomputer/Qwen-7B-Chat
; 55 togethercomputer/Qwen-7B
; 56 togethercomputer/RedPajama-INCITE-7B-Base
; 57 togethercomputer/RedPajama-INCITE-7B-Chat
; 58 togethercomputer/RedPajama-INCITE-7B-Instruct
; 59 togethercomputer/RedPajama-INCITE-Base-3B-v1
; 60 togethercomputer/RedPajama-INCITE-Chat-3B-v1
; 61 togethercomputer/RedPajama-INCITE-Instruct-3B-v1
; 62 togethercomputer/alpaca-7b
; 63 togethercomputer/codegen2-16B
; 64 togethercomputer/codegen2-7B
; 65 togethercomputer/falcon-40b-instruct
; 66 togethercomputer/falcon-40b
; 67 togethercomputer/falcon-7b-instruct
; 68 togethercomputer/falcon-7b
; 69 togethercomputer/guanaco-13b
; 70 togethercomputer/guanaco-33b
; 71 togethercomputer/guanaco-65b
; 72 togethercomputer/guanaco-7b
; 73 togethercomputer/llama-2-13b-chat
; 74 togethercomputer/llama-2-13b
; 75 togethercomputer/llama-2-70b-chat
; 76 togethercomputer/llama-2-70b
; 77 togethercomputer/llama-2-7b-chat
; 78 togethercomputer/llama-2-7b
; 79 togethercomputer/mpt-30b-chat
; 80 togethercomputer/mpt-30b-instruct
; 81 togethercomputer/mpt-30b
; 82 togethercomputer/mpt-7b-chat
; 83 togethercomputer/mpt-7b
; 84 togethercomputer/replit-code-v1-3b
; 85 upstage/SOLAR-0-70b-16bit
; 86 wavymulder/Analog-Diffusion |