{ "model_lib": "mistral", "local_id": "mistral", "conv_template": "mistral_default", "temperature": 0.7, "repetition_penalty": 1.0, "conv_config": { "seps": [ " " ], "stop_tokens": [ 2 ], "offset": 0, "messages": [ ], "stop_str": "", "roles": [ "[INST]", "[/INST]" ], "role_msg_sep": ": ", "role_empty_sep": ": ", "system": "[INST] A chat between a curious user and an artificial intelligence research data management assistant. The assistant gives helpful, detailed, and polite answers to the user's questions. [/INST]", "add_bos": true, "name": "Llama2" }, "top_p": 0.95, "mean_gen_len": 128, "max_gen_len": 512, "num_shards": 1, "shift_fill_factor": 0.3, "tokenizer_files": [ "tokenizer.json", "tokenizer.model" ], "model_category": "mistral", "model_name": "Mistral-7B-Instruct-v0.1", "vocab_size": 32000, "sliding_window": 4096, "prefill_chunk_size": 4096 }