base_model: tokyotech-llm/Swallow-7b-instruct-v0.1 #tokyotech-llm/Swallow-7b-instruct-v0.1 gate_mode: hidden # one of "hidden", "cheap_embed", or "random" dtype: bfloat16 # output dtype (float32, float16, or bfloat16) ## (optional) experts_per_token: 2 experts: - source_model: tokyotech-llm/Swallow-7b-instruct-v0.1 #okyotech-llm/Swallow-7b-instruct-v0.1 positive_prompts: - "日本語でのマルチターンの質問回答の専門家です。" - source_model: elyza/ELYZA-japanese-Llama-2-7b-fast-instruct #https://huggingface.co/elyza/ELYZA-japanese-Llama-2-7b-fast-instruct positive_prompts: - "日本語でのQ&Aの専門家です。" - source_model: elyza/ELYZA-japanese-CodeLlama-7b-instruct #https://huggingface.co/elyza/ELYZA-japanese-CodeLlama-7b-instruct positive_prompts: - "コーディングの専門家です。" - source_model: meta-math/MetaMath-7B-V1.0 #https://huggingface.co/meta-math/MetaMath-7B-V1.0 positive_prompts: - "Math(数学)の専門家です。"