{ "model_class": "llama2", "model_size": "7b", "architectures": [ "Llama2ForCausalLM" ], "dim": 4096, "n_layers": 32, "n_heads": 32, "n_kv_heads": null, "vocab_size": 32000, "multiple_of": 256, "ffn_dim_multiplier": null, "norm_eps": 1e-5, "max_batch_size": 32, "max_seq_len": 2048, "bos_token_id":1, "eos_token_id":2, "pad_token_id":-1, "torch_dtype": "float16", "pretraining_base": "llama2-7b-2t-tokens", "model_repo_folder_path": "llama2-7b-base" }