Xenova HF staff commited on
Commit
d2331a3
1 Parent(s): a59c683

Update config.json

Browse files
Files changed (1) hide show
  1. config.json +3 -4
config.json CHANGED
@@ -4,10 +4,6 @@
4
  "Phi3ForCausalLM"
5
  ],
6
  "attention_dropout": 0.0,
7
- "auto_map": {
8
- "AutoConfig": "configuration_phi3.Phi3Config",
9
- "AutoModelForCausalLM": "modeling_phi3.Phi3ForCausalLM"
10
- },
11
  "bos_token_id": 1,
12
  "embd_pdrop": 0.0,
13
  "eos_token_id": 32000,
@@ -132,6 +128,9 @@
132
  "tie_word_embeddings": false,
133
  "torch_dtype": "bfloat16",
134
  "transformers_version": "4.43.3",
 
 
 
135
  "use_cache": true,
136
  "attention_bias": false,
137
  "vocab_size": 32064
 
4
  "Phi3ForCausalLM"
5
  ],
6
  "attention_dropout": 0.0,
 
 
 
 
7
  "bos_token_id": 1,
8
  "embd_pdrop": 0.0,
9
  "eos_token_id": 32000,
 
128
  "tie_word_embeddings": false,
129
  "torch_dtype": "bfloat16",
130
  "transformers_version": "4.43.3",
131
+ "transformers.js_config": {
132
+ "kv_cache_dtype": "float16"
133
+ },
134
  "use_cache": true,
135
  "attention_bias": false,
136
  "vocab_size": 32064