File size: 308 Bytes
f3cdba6
 
 
 
 
 
 
 
 
 
 
 
 
 
1
2
3
4
5
6
7
8
9
10
11
12
13
14
{
	"sentence_encoder_config": {
		"num_hidden_layers": 2,
		"hidden_size": 4096,
		"rms_norm_eps": 1e-05,
		"attention_dropout": 0.0,
		"num_attention_heads": 32,
		"num_key_value_heads": 8,
		"max_position_embeddings": 256,
		"rope_theta": 500000.0,
		"intermediate_size": 14336,
		"hidden_act": "silu"
	}
}