{ | |
"model_lib": "mistral", | |
"local_id": "mistral", | |
"conv_template": "mistral_default", | |
"temperature": 0.7, | |
"repetition_penalty": 1.0, | |
"conv_config": { | |
"seps": [ | |
" " | |
], | |
"stop_tokens": [ | |
2 | |
], | |
"offset": 0, | |
"messages": [ | |
], | |
"stop_str": "</s>", | |
"roles": [ | |
"[INST]", | |
"[/INST]" | |
], | |
"role_msg_sep": ": ", | |
"role_empty_sep": ": ", | |
"system": "<s>[INST] A chat between a curious user and an artificial intelligence research data management assistant. The assistant gives helpful, detailed, and polite answers to the user's questions. [/INST]", | |
"add_bos": true, | |
"name": "Llama2" | |
}, | |
"top_p": 0.95, | |
"mean_gen_len": 128, | |
"max_gen_len": 512, | |
"num_shards": 1, | |
"shift_fill_factor": 0.3, | |
"tokenizer_files": [ | |
"tokenizer.json", | |
"tokenizer.model" | |
], | |
"model_category": "mistral", | |
"model_name": "Mistral-7B-Instruct-v0.1", | |
"vocab_size": 32000, | |
"sliding_window": 4096, | |
"prefill_chunk_size": 4096 | |
} |