{ "model_lib": "zephyr-7b-beta-chatRDM-q4f32_1", "local_id": "zephyr-7b-beta-q4f32_1", "conv_template": "mistral_default", "temperature": 0.7, "repetition_penalty": 1.0, "top_p": 0.95, "mean_gen_len": 128, "max_gen_len": 512, "num_shards": 1, "use_presharded_weights": false, "shift_fill_factor": 0.3, "conv_config": { "seps": [ " " ], "stop_tokens": [ 2 ], "offset": 0, "messages": [ ], "stop_str": "", "roles": [ "<|user|>", "<|assistant|>" ], "role_msg_sep": " ", "role_empty_sep": " ", "system": "<|system|> A chat between a curious user and an artificial intelligence research data management assistant. The assistant gives helpful, detailed, and polite answers to the user's questions. [/INST]", "add_bos": true, "name": "Llama2" }, "tokenizer_files": [ "tokenizer.json", "added_tokens.json", "tokenizer.model" ], "model_category": "mistral", "model_name": "zephyr-7b-beta", "vocab_size": 32000, "prefill_chunk_size": 4096, "sliding_window": 4096 }