Diogo-V commited on
Commit
4d37030
1 Parent(s): 2f07ab8

Upload learned parameters for llama3 in bit 8

Browse files
This view is limited to 50 files because it contains too many changes.   See raw diff
Files changed (50) hide show
  1. params/llama3/8/norm_nlr/comb_12/init/lm_head/_0.pt +3 -0
  2. params/llama3/8/norm_nlr/comb_12/init/lm_head/_1.pt +3 -0
  3. params/llama3/8/norm_nlr/comb_12/init/lm_head/_s.pt +3 -0
  4. params/llama3/8/norm_nlr/comb_12/init/model.layers.0.mlp.down_proj/_0.pt +3 -0
  5. params/llama3/8/norm_nlr/comb_12/init/model.layers.0.mlp.down_proj/_1.pt +3 -0
  6. params/llama3/8/norm_nlr/comb_12/init/model.layers.0.mlp.down_proj/_s.pt +3 -0
  7. params/llama3/8/norm_nlr/comb_12/init/model.layers.0.mlp.gate_proj/_0.pt +3 -0
  8. params/llama3/8/norm_nlr/comb_12/init/model.layers.0.mlp.gate_proj/_1.pt +3 -0
  9. params/llama3/8/norm_nlr/comb_12/init/model.layers.0.mlp.gate_proj/_s.pt +3 -0
  10. params/llama3/8/norm_nlr/comb_12/init/model.layers.0.mlp.up_proj/_0.pt +3 -0
  11. params/llama3/8/norm_nlr/comb_12/init/model.layers.0.mlp.up_proj/_1.pt +3 -0
  12. params/llama3/8/norm_nlr/comb_12/init/model.layers.0.mlp.up_proj/_s.pt +3 -0
  13. params/llama3/8/norm_nlr/comb_12/init/model.layers.0.self_attn.k_proj/_0.pt +3 -0
  14. params/llama3/8/norm_nlr/comb_12/init/model.layers.0.self_attn.k_proj/_1.pt +3 -0
  15. params/llama3/8/norm_nlr/comb_12/init/model.layers.0.self_attn.k_proj/_s.pt +3 -0
  16. params/llama3/8/norm_nlr/comb_12/init/model.layers.0.self_attn.o_proj/_0.pt +3 -0
  17. params/llama3/8/norm_nlr/comb_12/init/model.layers.0.self_attn.o_proj/_1.pt +3 -0
  18. params/llama3/8/norm_nlr/comb_12/init/model.layers.0.self_attn.o_proj/_s.pt +3 -0
  19. params/llama3/8/norm_nlr/comb_12/init/model.layers.0.self_attn.q_proj/_0.pt +3 -0
  20. params/llama3/8/norm_nlr/comb_12/init/model.layers.0.self_attn.q_proj/_1.pt +3 -0
  21. params/llama3/8/norm_nlr/comb_12/init/model.layers.0.self_attn.q_proj/_s.pt +3 -0
  22. params/llama3/8/norm_nlr/comb_12/init/model.layers.0.self_attn.v_proj/_0.pt +3 -0
  23. params/llama3/8/norm_nlr/comb_12/init/model.layers.0.self_attn.v_proj/_1.pt +3 -0
  24. params/llama3/8/norm_nlr/comb_12/init/model.layers.0.self_attn.v_proj/_s.pt +3 -0
  25. params/llama3/8/norm_nlr/comb_12/init/model.layers.1.mlp.down_proj/_0.pt +3 -0
  26. params/llama3/8/norm_nlr/comb_12/init/model.layers.1.mlp.down_proj/_1.pt +3 -0
  27. params/llama3/8/norm_nlr/comb_12/init/model.layers.1.mlp.down_proj/_s.pt +3 -0
  28. params/llama3/8/norm_nlr/comb_12/init/model.layers.1.mlp.gate_proj/_0.pt +3 -0
  29. params/llama3/8/norm_nlr/comb_12/init/model.layers.1.mlp.gate_proj/_1.pt +3 -0
  30. params/llama3/8/norm_nlr/comb_12/init/model.layers.1.mlp.gate_proj/_s.pt +3 -0
  31. params/llama3/8/norm_nlr/comb_12/init/model.layers.1.mlp.up_proj/_0.pt +3 -0
  32. params/llama3/8/norm_nlr/comb_12/init/model.layers.1.mlp.up_proj/_1.pt +3 -0
  33. params/llama3/8/norm_nlr/comb_12/init/model.layers.1.mlp.up_proj/_s.pt +3 -0
  34. params/llama3/8/norm_nlr/comb_12/init/model.layers.1.self_attn.k_proj/_0.pt +3 -0
  35. params/llama3/8/norm_nlr/comb_12/init/model.layers.1.self_attn.k_proj/_1.pt +3 -0
  36. params/llama3/8/norm_nlr/comb_12/init/model.layers.1.self_attn.k_proj/_s.pt +3 -0
  37. params/llama3/8/norm_nlr/comb_12/init/model.layers.1.self_attn.o_proj/_0.pt +3 -0
  38. params/llama3/8/norm_nlr/comb_12/init/model.layers.1.self_attn.o_proj/_1.pt +3 -0
  39. params/llama3/8/norm_nlr/comb_12/init/model.layers.1.self_attn.o_proj/_s.pt +3 -0
  40. params/llama3/8/norm_nlr/comb_12/init/model.layers.1.self_attn.q_proj/_0.pt +3 -0
  41. params/llama3/8/norm_nlr/comb_12/init/model.layers.1.self_attn.q_proj/_1.pt +3 -0
  42. params/llama3/8/norm_nlr/comb_12/init/model.layers.1.self_attn.q_proj/_s.pt +3 -0
  43. params/llama3/8/norm_nlr/comb_12/init/model.layers.1.self_attn.v_proj/_0.pt +3 -0
  44. params/llama3/8/norm_nlr/comb_12/init/model.layers.1.self_attn.v_proj/_1.pt +3 -0
  45. params/llama3/8/norm_nlr/comb_12/init/model.layers.1.self_attn.v_proj/_s.pt +3 -0
  46. params/llama3/8/norm_nlr/comb_12/init/model.layers.10.mlp.down_proj/_0.pt +3 -0
  47. params/llama3/8/norm_nlr/comb_12/init/model.layers.10.mlp.down_proj/_1.pt +3 -0
  48. params/llama3/8/norm_nlr/comb_12/init/model.layers.10.mlp.down_proj/_s.pt +3 -0
  49. params/llama3/8/norm_nlr/comb_12/init/model.layers.10.mlp.gate_proj/_0.pt +3 -0
  50. params/llama3/8/norm_nlr/comb_12/init/model.layers.10.mlp.gate_proj/_1.pt +3 -0
params/llama3/8/norm_nlr/comb_12/init/lm_head/_0.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:d62010371e35ea949ac33ba9632d175f94a6545cc92b475143489f66ecb3da56
3
+ size 514115
params/llama3/8/norm_nlr/comb_12/init/lm_head/_1.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:904337ebfa597681bfbbfa8b559ba2095597cbcf4e26241efd09a9b2855889d8
3
+ size 514115
params/llama3/8/norm_nlr/comb_12/init/lm_head/_s.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:227ab84230c9a7ae14436c928491ee58b5c1dec31e1337d05d09968ccf46b0ea
3
+ size 514115
params/llama3/8/norm_nlr/comb_12/init/model.layers.0.mlp.down_proj/_0.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:0f4a1e9323d91b3fa5ae5b222540b13dab9a3ad2f0e844d31230394be4a02a9f
3
+ size 17475
params/llama3/8/norm_nlr/comb_12/init/model.layers.0.mlp.down_proj/_1.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:679200fba0cf9ed65b9a32d946a395f67aba822eb86452b2f457a79f1d991e0f
3
+ size 17475
params/llama3/8/norm_nlr/comb_12/init/model.layers.0.mlp.down_proj/_s.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:2d56a3010463ef90843f163c38502bc221f948828182434b35da85d08c507bbc
3
+ size 17475
params/llama3/8/norm_nlr/comb_12/init/model.layers.0.mlp.gate_proj/_0.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:3274ce7047df96b5a30997b53a05cfa70cc18bf45d9cabb4f173cf0024ee2bec
3
+ size 58435
params/llama3/8/norm_nlr/comb_12/init/model.layers.0.mlp.gate_proj/_1.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:d4010e3f95de9ecdf2aa79f4c65ea39d4fca4f8adc1f32149a9326c411a6e05b
3
+ size 58435
params/llama3/8/norm_nlr/comb_12/init/model.layers.0.mlp.gate_proj/_s.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:096e8565d60628abc2e09dfaab878994a7fbd527dcf5459d951bd93452bf8817
3
+ size 58435
params/llama3/8/norm_nlr/comb_12/init/model.layers.0.mlp.up_proj/_0.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:2e97e579461e97242d9310d9eec28be047a46bb82ff7fdaaf4ae3c196110b566
3
+ size 58435
params/llama3/8/norm_nlr/comb_12/init/model.layers.0.mlp.up_proj/_1.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:da62d2ce96c91202ec62150858dd830d63db5e1dc4bc1e0791eac195b9b24f5d
3
+ size 58435
params/llama3/8/norm_nlr/comb_12/init/model.layers.0.mlp.up_proj/_s.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:096e8565d60628abc2e09dfaab878994a7fbd527dcf5459d951bd93452bf8817
3
+ size 58435
params/llama3/8/norm_nlr/comb_12/init/model.layers.0.self_attn.k_proj/_0.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:bb4716528436dfb0654bfc339c7bfd7dea234fff1c6257e602eaa0c048bf27f9
3
+ size 5187
params/llama3/8/norm_nlr/comb_12/init/model.layers.0.self_attn.k_proj/_1.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:c93ffb9727abae5932e182fa9a8c8ab806fcfc2d7462b17d8b3ad6e620bf958a
3
+ size 5187
params/llama3/8/norm_nlr/comb_12/init/model.layers.0.self_attn.k_proj/_s.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:7c05175971967d6605a179b1bb181fb3ab764a53703e2497801c4b7e5ebfade9
3
+ size 5187
params/llama3/8/norm_nlr/comb_12/init/model.layers.0.self_attn.o_proj/_0.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:5cb04c2af75fc6d840f0ace32dd0666995989d3a49582a847f698bf5f2c41caa
3
+ size 17475
params/llama3/8/norm_nlr/comb_12/init/model.layers.0.self_attn.o_proj/_1.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:fa544c5930834d2ff10075d529e43bb8067baf8b9cf40eaa25345bc204d3d291
3
+ size 17475
params/llama3/8/norm_nlr/comb_12/init/model.layers.0.self_attn.o_proj/_s.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:2d56a3010463ef90843f163c38502bc221f948828182434b35da85d08c507bbc
3
+ size 17475
params/llama3/8/norm_nlr/comb_12/init/model.layers.0.self_attn.q_proj/_0.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:bfd2604800d414436d879c8b910a6682870263cf0d364048037be87916291d2b
3
+ size 17475
params/llama3/8/norm_nlr/comb_12/init/model.layers.0.self_attn.q_proj/_1.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:5a1aa3608aafecde6c30c632180c82006e209c91709343944e7767917c855e2d
3
+ size 17475
params/llama3/8/norm_nlr/comb_12/init/model.layers.0.self_attn.q_proj/_s.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:2d56a3010463ef90843f163c38502bc221f948828182434b35da85d08c507bbc
3
+ size 17475
params/llama3/8/norm_nlr/comb_12/init/model.layers.0.self_attn.v_proj/_0.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:a06e364e0a9d15403193e2e0190da6e337b4afe54ade64bcdd382b06290b4913
3
+ size 5187
params/llama3/8/norm_nlr/comb_12/init/model.layers.0.self_attn.v_proj/_1.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:643ebae93814c18f0d966ab18cfd780a63ab8f3c9f31b43aafa5a5a0750b7b30
3
+ size 5187
params/llama3/8/norm_nlr/comb_12/init/model.layers.0.self_attn.v_proj/_s.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:7c05175971967d6605a179b1bb181fb3ab764a53703e2497801c4b7e5ebfade9
3
+ size 5187
params/llama3/8/norm_nlr/comb_12/init/model.layers.1.mlp.down_proj/_0.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:c6dfae9cbec368d7bd42ae8cb15e4a20c62d78d40124c808daff2fda5fe972d9
3
+ size 17475
params/llama3/8/norm_nlr/comb_12/init/model.layers.1.mlp.down_proj/_1.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:4892301f4116e2c75a386aa8dfd7fb5ec645f1570af4c83ee1d0b5d05d59b76a
3
+ size 17475
params/llama3/8/norm_nlr/comb_12/init/model.layers.1.mlp.down_proj/_s.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:2d56a3010463ef90843f163c38502bc221f948828182434b35da85d08c507bbc
3
+ size 17475
params/llama3/8/norm_nlr/comb_12/init/model.layers.1.mlp.gate_proj/_0.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:a5becf80ee5d42797b23bd6b8d636e78d9736614cd8517a4a6cad760f4d00b6d
3
+ size 58435
params/llama3/8/norm_nlr/comb_12/init/model.layers.1.mlp.gate_proj/_1.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:f984cf9e9d5d0486638851d98efb859402eb8db99b123839ff210ea8943377c8
3
+ size 58435
params/llama3/8/norm_nlr/comb_12/init/model.layers.1.mlp.gate_proj/_s.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:096e8565d60628abc2e09dfaab878994a7fbd527dcf5459d951bd93452bf8817
3
+ size 58435
params/llama3/8/norm_nlr/comb_12/init/model.layers.1.mlp.up_proj/_0.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:367063f6e9dacf4eeda06a6022838c740729cb67cb8067b14a6e59e162a6336b
3
+ size 58435
params/llama3/8/norm_nlr/comb_12/init/model.layers.1.mlp.up_proj/_1.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:56b4884878730f8ccee9e811c9c1c9a3d9f5a68e7dd58f5eaba15e61d4c07103
3
+ size 58435
params/llama3/8/norm_nlr/comb_12/init/model.layers.1.mlp.up_proj/_s.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:096e8565d60628abc2e09dfaab878994a7fbd527dcf5459d951bd93452bf8817
3
+ size 58435
params/llama3/8/norm_nlr/comb_12/init/model.layers.1.self_attn.k_proj/_0.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:a89f2a6fc831a7d4e136ac5988d82420d8d2d2ab887ee8031d788781fd312eb2
3
+ size 5187
params/llama3/8/norm_nlr/comb_12/init/model.layers.1.self_attn.k_proj/_1.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:5fece8d039f54b27c05b441dcf3fb3bc5e83ebca5123dc91ee67b03bdf6f944f
3
+ size 5187
params/llama3/8/norm_nlr/comb_12/init/model.layers.1.self_attn.k_proj/_s.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:7c05175971967d6605a179b1bb181fb3ab764a53703e2497801c4b7e5ebfade9
3
+ size 5187
params/llama3/8/norm_nlr/comb_12/init/model.layers.1.self_attn.o_proj/_0.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:7281dfa1c1c828b510d2563bd4072c0c30d92dd6e828ba856982596167346a4b
3
+ size 17475
params/llama3/8/norm_nlr/comb_12/init/model.layers.1.self_attn.o_proj/_1.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:5c5fb8bdf840f756fb3741b520df7d724005c62df3fe91173eefadc6d760809d
3
+ size 17475
params/llama3/8/norm_nlr/comb_12/init/model.layers.1.self_attn.o_proj/_s.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:2d56a3010463ef90843f163c38502bc221f948828182434b35da85d08c507bbc
3
+ size 17475
params/llama3/8/norm_nlr/comb_12/init/model.layers.1.self_attn.q_proj/_0.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:410598afab91dd46c8eb06c8889e9071e48e735c9ef8a88b0addd3bf6a9973d5
3
+ size 17475
params/llama3/8/norm_nlr/comb_12/init/model.layers.1.self_attn.q_proj/_1.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:59be746e994e20ae89a23febfe8d0ec2334f77d5c09750051431fd305e11aacc
3
+ size 17475
params/llama3/8/norm_nlr/comb_12/init/model.layers.1.self_attn.q_proj/_s.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:2d56a3010463ef90843f163c38502bc221f948828182434b35da85d08c507bbc
3
+ size 17475
params/llama3/8/norm_nlr/comb_12/init/model.layers.1.self_attn.v_proj/_0.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:6c3b8f1b0012b20f8d2fbce1d717b23591f5d173ffd4c0f0ed787d46fae2866d
3
+ size 5187
params/llama3/8/norm_nlr/comb_12/init/model.layers.1.self_attn.v_proj/_1.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:dac76cc0dc8a24d45ed6dd72e7df96595c685babe7a069ea04fd196fcdb9807d
3
+ size 5187
params/llama3/8/norm_nlr/comb_12/init/model.layers.1.self_attn.v_proj/_s.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:7c05175971967d6605a179b1bb181fb3ab764a53703e2497801c4b7e5ebfade9
3
+ size 5187
params/llama3/8/norm_nlr/comb_12/init/model.layers.10.mlp.down_proj/_0.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:0e49ed7335aa2abfeb922b5af3f26d1f70feb3c26686e7bed41a852d701467d1
3
+ size 17475
params/llama3/8/norm_nlr/comb_12/init/model.layers.10.mlp.down_proj/_1.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:fba2574942416108980a55c48758319a54e6d2ed734a22fcc8959f0b2740cd09
3
+ size 17475
params/llama3/8/norm_nlr/comb_12/init/model.layers.10.mlp.down_proj/_s.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:2d56a3010463ef90843f163c38502bc221f948828182434b35da85d08c507bbc
3
+ size 17475
params/llama3/8/norm_nlr/comb_12/init/model.layers.10.mlp.gate_proj/_0.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:26c96d1c897a2d3573966358e6ff0d36d3ac8db0d739e954b9c3d347cb56d77d
3
+ size 58435
params/llama3/8/norm_nlr/comb_12/init/model.layers.10.mlp.gate_proj/_1.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:2553af7a3ab9c143ad189c859ff04620cc7933dd9161ed4c9b959937565f5f43
3
+ size 58435