arnocandel commited on
Commit
f55dcbf
·
1 Parent(s): 8d8266c

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +32 -1
README.md CHANGED
@@ -20,4 +20,35 @@ Try it live on our [h2oGPT demo](https://gpt.h2o.ai) with side-by-side LLM compa
20
 
21
  See how it compares to other models on our [LLM Leaderboard](https://evalgpt.ai/)!
22
 
23
- See more at [H2O.ai](https://h2o.ai/)
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
20
 
21
  See how it compares to other models on our [LLM Leaderboard](https://evalgpt.ai/)!
22
 
23
+ See more at [H2O.ai](https://h2o.ai/)
24
+
25
+
26
+ ## Model Architecture
27
+ ```
28
+ LlamaForCausalLM(
29
+ (model): LlamaModel(
30
+ (embed_tokens): Embedding(32000, 4096, padding_idx=0)
31
+ (layers): ModuleList(
32
+ (0-31): 32 x LlamaDecoderLayer(
33
+ (self_attn): LlamaAttention(
34
+ (q_proj): Linear(in_features=4096, out_features=4096, bias=False)
35
+ (k_proj): Linear(in_features=4096, out_features=4096, bias=False)
36
+ (v_proj): Linear(in_features=4096, out_features=4096, bias=False)
37
+ (o_proj): Linear(in_features=4096, out_features=4096, bias=False)
38
+ (rotary_emb): LlamaRotaryEmbedding()
39
+ )
40
+ (mlp): LlamaMLP(
41
+ (gate_proj): Linear(in_features=4096, out_features=11008, bias=False)
42
+ (up_proj): Linear(in_features=4096, out_features=11008, bias=False)
43
+ (down_proj): Linear(in_features=11008, out_features=4096, bias=False)
44
+ (act_fn): SiLUActivation()
45
+ )
46
+ (input_layernorm): LlamaRMSNorm()
47
+ (post_attention_layernorm): LlamaRMSNorm()
48
+ )
49
+ )
50
+ (norm): LlamaRMSNorm()
51
+ )
52
+ (lm_head): Linear(in_features=4096, out_features=32000, bias=False)
53
+ )
54
+ ```