File size: 419 Bytes
6aced58
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
from dataclasses import dataclass

@dataclass
class LlamaConfig:
    hidden_size: int = 576
    num_attention_heads: int = 16
    num_key_value_heads: int = 4
    num_hidden_layers: int = 30
    intermediate_size: int = 1536
    hidden_act: str = "silu"
    rms_norm_eps: float = 1e-5
    vocab_size: int = 49152
    max_position_embeddings: int = 8192
    rope_theta: int = 100000
    tie_word_embeddings: bool = False