arnocandel commited on
Commit
d205b10
1 Parent(s): fafcff4

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +33 -1
README.md CHANGED
@@ -20,4 +20,36 @@ Try it live on our [h2oGPT demo](https://gpt.h2o.ai) with side-by-side LLM compa
20
 
21
  See how it compares to other models on our [LLM Leaderboard](https://evalgpt.ai/)!
22
 
23
- See more at [H2O.ai](https://h2o.ai/)
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
20
 
21
  See how it compares to other models on our [LLM Leaderboard](https://evalgpt.ai/)!
22
 
23
+ See more at [H2O.ai](https://h2o.ai/)
24
+
25
+
26
+ ## Model Architecture
27
+
28
+ ```
29
+ LlamaForCausalLM(
30
+ (model): LlamaModel(
31
+ (embed_tokens): Embedding(32000, 5120, padding_idx=0)
32
+ (layers): ModuleList(
33
+ (0-39): 40 x LlamaDecoderLayer(
34
+ (self_attn): LlamaAttention(
35
+ (q_proj): Linear(in_features=5120, out_features=5120, bias=False)
36
+ (k_proj): Linear(in_features=5120, out_features=5120, bias=False)
37
+ (v_proj): Linear(in_features=5120, out_features=5120, bias=False)
38
+ (o_proj): Linear(in_features=5120, out_features=5120, bias=False)
39
+ (rotary_emb): LlamaRotaryEmbedding()
40
+ )
41
+ (mlp): LlamaMLP(
42
+ (gate_proj): Linear(in_features=5120, out_features=13824, bias=False)
43
+ (up_proj): Linear(in_features=5120, out_features=13824, bias=False)
44
+ (down_proj): Linear(in_features=13824, out_features=5120, bias=False)
45
+ (act_fn): SiLUActivation()
46
+ )
47
+ (input_layernorm): LlamaRMSNorm()
48
+ (post_attention_layernorm): LlamaRMSNorm()
49
+ )
50
+ )
51
+ (norm): LlamaRMSNorm()
52
+ )
53
+ (lm_head): Linear(in_features=5120, out_features=32000, bias=False)
54
+ )
55
+ ```