File size: 657 Bytes
6c4fcb5
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
_name_or_path: ./models/dolphin-2_6-phi-2
activation_function: gelu_new
architectures:
  - PhiForCausalLM
attn_pdrop: 0.0
auto_map:
  AutoConfig: configuration_phi.PhiConfig
  AutoModel: modeling_phi.PhiForCausalLM
  AutoModelForCausalLM: modeling_phi.PhiForCausalLM
embd_pdrop: 0.0
flash_attn: false
flash_rotary: false
fused_dense: false
img_processor: null
initializer_range: 0.02
layer_norm_epsilon: 1e-05
model_type: phi-msft
n_embd: 2560
n_head: 32
n_head_kv: null
n_inner: null
n_layer: 32
n_positions: 2048
resid_pdrop: 0.1
rotary_dim: 32
tie_word_embeddings: false
torch_dtype: float16
transformers_version: 4.36.2
use_cache: true
vocab_size: 51200