LLaMA-65B-HF / generation_config.json
Neko-Institute-of-Science's picture
LLaMA: convert_llama_weights_to_hf.py as of c0f99b4d2ec73090595914dde4c16da207e21d73
dff3a11
raw
history blame
137 Bytes
{
"_from_model_config": true,
"bos_token_id": 1,
"eos_token_id": 2,
"pad_token_id": 0,
"transformers_version": "4.28.0.dev0"
}