NeMo
nvidia
jiaqiz commited on
Commit
6df829b
1 Parent(s): dc48b99

Add files using large-upload tool

Browse files
Files changed (26) hide show
  1. .gitattributes +25 -0
  2. model_weights/model.decoder.layers.self_attention.linear_proj.weight/10.0.0 +3 -0
  3. model_weights/model.decoder.layers.self_attention.linear_proj.weight/11.0.0 +3 -0
  4. model_weights/model.decoder.layers.self_attention.linear_proj.weight/2.0.2 +3 -0
  5. model_weights/model.decoder.layers.self_attention.linear_proj.weight/25.0.0 +3 -0
  6. model_weights/model.decoder.layers.self_attention.linear_proj.weight/3.0.3 +3 -0
  7. model_weights/model.decoder.layers.self_attention.linear_proj.weight/31.0.5 +3 -0
  8. model_weights/model.decoder.layers.self_attention.linear_proj.weight/34.0.2 +3 -0
  9. model_weights/model.decoder.layers.self_attention.linear_proj.weight/34.0.7 +3 -0
  10. model_weights/model.decoder.layers.self_attention.linear_proj.weight/35.0.1 +3 -0
  11. model_weights/model.decoder.layers.self_attention.linear_proj.weight/38.0.5 +3 -0
  12. model_weights/model.decoder.layers.self_attention.linear_proj.weight/52.0.0 +3 -0
  13. model_weights/model.decoder.layers.self_attention.linear_proj.weight/52.0.3 +3 -0
  14. model_weights/model.decoder.layers.self_attention.linear_proj.weight/55.0.1 +3 -0
  15. model_weights/model.decoder.layers.self_attention.linear_proj.weight/57.0.0 +3 -0
  16. model_weights/model.decoder.layers.self_attention.linear_proj.weight/63.0.7 +3 -0
  17. model_weights/model.decoder.layers.self_attention.linear_proj.weight/64.0.4 +3 -0
  18. model_weights/model.decoder.layers.self_attention.linear_proj.weight/64.0.5 +3 -0
  19. model_weights/model.decoder.layers.self_attention.linear_proj.weight/67.0.4 +3 -0
  20. model_weights/model.decoder.layers.self_attention.linear_proj.weight/68.0.4 +3 -0
  21. model_weights/model.decoder.layers.self_attention.linear_proj.weight/7.0.4 +3 -0
  22. model_weights/model.decoder.layers.self_attention.linear_proj.weight/8.0.3 +3 -0
  23. model_weights/model.decoder.layers.self_attention.linear_proj.weight/85.0.3 +3 -0
  24. model_weights/model.decoder.layers.self_attention.linear_proj.weight/88.0.7 +3 -0
  25. model_weights/model.decoder.layers.self_attention.linear_proj.weight/90.0.1 +3 -0
  26. model_weights/model.output_layer.weight/2.0 +3 -0
.gitattributes CHANGED
@@ -2670,3 +2670,28 @@ model_weights/model.decoder.layers.self_attention.linear_proj.weight/57.0.5 filt
2670
  model_weights/model.decoder.layers.self_attention.linear_proj.weight/0.0.7 filter=lfs diff=lfs merge=lfs -text
2671
  model_weights/model.decoder.layers.self_attention.linear_proj.weight/21.0.5 filter=lfs diff=lfs merge=lfs -text
2672
  model_weights/model.decoder.layers.self_attention.linear_proj.weight/46.0.7 filter=lfs diff=lfs merge=lfs -text
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
2670
  model_weights/model.decoder.layers.self_attention.linear_proj.weight/0.0.7 filter=lfs diff=lfs merge=lfs -text
2671
  model_weights/model.decoder.layers.self_attention.linear_proj.weight/21.0.5 filter=lfs diff=lfs merge=lfs -text
2672
  model_weights/model.decoder.layers.self_attention.linear_proj.weight/46.0.7 filter=lfs diff=lfs merge=lfs -text
2673
+ model_weights/model.decoder.layers.self_attention.linear_proj.weight/57.0.0 filter=lfs diff=lfs merge=lfs -text
2674
+ model_weights/model.decoder.layers.self_attention.linear_proj.weight/10.0.0 filter=lfs diff=lfs merge=lfs -text
2675
+ model_weights/model.decoder.layers.self_attention.linear_proj.weight/11.0.0 filter=lfs diff=lfs merge=lfs -text
2676
+ model_weights/model.decoder.layers.self_attention.linear_proj.weight/52.0.3 filter=lfs diff=lfs merge=lfs -text
2677
+ model_weights/model.decoder.layers.self_attention.linear_proj.weight/38.0.5 filter=lfs diff=lfs merge=lfs -text
2678
+ model_weights/model.decoder.layers.self_attention.linear_proj.weight/52.0.0 filter=lfs diff=lfs merge=lfs -text
2679
+ model_weights/model.decoder.layers.self_attention.linear_proj.weight/3.0.3 filter=lfs diff=lfs merge=lfs -text
2680
+ model_weights/model.decoder.layers.self_attention.linear_proj.weight/63.0.7 filter=lfs diff=lfs merge=lfs -text
2681
+ model_weights/model.decoder.layers.self_attention.linear_proj.weight/34.0.2 filter=lfs diff=lfs merge=lfs -text
2682
+ model_weights/model.decoder.layers.self_attention.linear_proj.weight/34.0.7 filter=lfs diff=lfs merge=lfs -text
2683
+ model_weights/model.decoder.layers.self_attention.linear_proj.weight/90.0.1 filter=lfs diff=lfs merge=lfs -text
2684
+ model_weights/model.decoder.layers.self_attention.linear_proj.weight/35.0.1 filter=lfs diff=lfs merge=lfs -text
2685
+ model_weights/model.decoder.layers.self_attention.linear_proj.weight/67.0.4 filter=lfs diff=lfs merge=lfs -text
2686
+ model_weights/model.decoder.layers.self_attention.linear_proj.weight/8.0.3 filter=lfs diff=lfs merge=lfs -text
2687
+ model_weights/model.decoder.layers.self_attention.linear_proj.weight/85.0.3 filter=lfs diff=lfs merge=lfs -text
2688
+ model_weights/model.decoder.layers.self_attention.linear_proj.weight/64.0.5 filter=lfs diff=lfs merge=lfs -text
2689
+ model_weights/model.decoder.layers.self_attention.linear_proj.weight/25.0.0 filter=lfs diff=lfs merge=lfs -text
2690
+ model_weights/model.output_layer.weight/2.0 filter=lfs diff=lfs merge=lfs -text
2691
+ model_weights/model.decoder.layers.self_attention.linear_proj.weight/2.0.2 filter=lfs diff=lfs merge=lfs -text
2692
+ model_weights/model.decoder.layers.self_attention.linear_proj.weight/7.0.4 filter=lfs diff=lfs merge=lfs -text
2693
+ model_weights/model.decoder.layers.self_attention.linear_proj.weight/88.0.7 filter=lfs diff=lfs merge=lfs -text
2694
+ model_weights/model.decoder.layers.self_attention.linear_proj.weight/68.0.4 filter=lfs diff=lfs merge=lfs -text
2695
+ model_weights/model.decoder.layers.self_attention.linear_proj.weight/55.0.1 filter=lfs diff=lfs merge=lfs -text
2696
+ model_weights/model.decoder.layers.self_attention.linear_proj.weight/64.0.4 filter=lfs diff=lfs merge=lfs -text
2697
+ model_weights/model.decoder.layers.self_attention.linear_proj.weight/31.0.5 filter=lfs diff=lfs merge=lfs -text
model_weights/model.decoder.layers.self_attention.linear_proj.weight/10.0.0 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:ef1dd466b86c507addf3b31ba74d154b88506bf11953c107a8361453bb872c40
3
+ size 84934656
model_weights/model.decoder.layers.self_attention.linear_proj.weight/11.0.0 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:661992af14b381acdbc9a64055ed8d34d5073f82bbcc6515efc040a83a6d90fd
3
+ size 84934656
model_weights/model.decoder.layers.self_attention.linear_proj.weight/2.0.2 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:cdedede2a71aed0d4cd84b4613ee7e78fabe49d5a44c313e9f0f879828acb818
3
+ size 84934656
model_weights/model.decoder.layers.self_attention.linear_proj.weight/25.0.0 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:e2d903c7e0c9c5442a72c60ff756b6a94c8b4efc01ccf83cc3a7a3679d087b3c
3
+ size 84934656
model_weights/model.decoder.layers.self_attention.linear_proj.weight/3.0.3 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:a51d7e7fbf755f60db5d95dcf3c0360a4a86e8a684508e93ddde7ef1fce25888
3
+ size 84934656
model_weights/model.decoder.layers.self_attention.linear_proj.weight/31.0.5 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:1bd4f241c7666964d001d58a5c3ad52f104c87808109e9d35d2957af2d99a7e9
3
+ size 84934656
model_weights/model.decoder.layers.self_attention.linear_proj.weight/34.0.2 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:92d9e474b1577ec55ccc7b891b4365c66ca7046e6db8370d8d614ba25f539083
3
+ size 84934656
model_weights/model.decoder.layers.self_attention.linear_proj.weight/34.0.7 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:466109a477e949ebab88eed3e511a46d5d46f8c0329fdb8963e676ed6706ff01
3
+ size 84934656
model_weights/model.decoder.layers.self_attention.linear_proj.weight/35.0.1 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:426e63a9d561c1e5e1742970c65cbb87a43ef1d415674486b9a328aa0b6960c4
3
+ size 84934656
model_weights/model.decoder.layers.self_attention.linear_proj.weight/38.0.5 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:214d8cd1b3575e5a50fd92b812c6a49a81d00f34bee46e1afeb1fa7feea491d2
3
+ size 84934656
model_weights/model.decoder.layers.self_attention.linear_proj.weight/52.0.0 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:39c6adaad093f23d4f9f3675d01be238d2a36676d97255b7e9ec91be5666397c
3
+ size 84934656
model_weights/model.decoder.layers.self_attention.linear_proj.weight/52.0.3 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:086ad795bc21895a3ca087d138f7004b2fc6ad195e96587523eb2d94b48239e2
3
+ size 84934656
model_weights/model.decoder.layers.self_attention.linear_proj.weight/55.0.1 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:0cd52beed725bd6cd4a20494140967c3aa3163974d5f654f901090decd7e1b46
3
+ size 84934656
model_weights/model.decoder.layers.self_attention.linear_proj.weight/57.0.0 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:402d7575fd94a9fd64dbbec9d9a28ea5206f160151aaa302c40633cf197f78e5
3
+ size 84934656
model_weights/model.decoder.layers.self_attention.linear_proj.weight/63.0.7 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:f8d1459ea52a3e5311af92471d32ed563ffc2ed53a226028fcd514581d53d4b5
3
+ size 84934656
model_weights/model.decoder.layers.self_attention.linear_proj.weight/64.0.4 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:9e8950a827472a5ee9d39a7844518ccc3d9a6e29de57d269f9d76ddfe8ffacf4
3
+ size 84934656
model_weights/model.decoder.layers.self_attention.linear_proj.weight/64.0.5 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:cb90675652a2efd41d55a4f725b1032f4ab60de9da1b28265e0cdb43783949b5
3
+ size 84934656
model_weights/model.decoder.layers.self_attention.linear_proj.weight/67.0.4 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:d495a8e0a81d78515c596c872117768db6e6a582e5c5676ece6996cd9a28a887
3
+ size 84934656
model_weights/model.decoder.layers.self_attention.linear_proj.weight/68.0.4 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:8e5ef442faaad53b4157f8f4d9e7e605a1f4cee7dd757d36e3438c788807924d
3
+ size 84934656
model_weights/model.decoder.layers.self_attention.linear_proj.weight/7.0.4 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:fcb3c6083f30a017e8a712392d33dc7fc063d8e41ef03028d3b6a3c96ca8efa7
3
+ size 84934656
model_weights/model.decoder.layers.self_attention.linear_proj.weight/8.0.3 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:c16f35bc04a166747870da93fc8b70f3e2bc1f77d0d01dd41472ecedbaba203d
3
+ size 84934656
model_weights/model.decoder.layers.self_attention.linear_proj.weight/85.0.3 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:254e6d82762e54e7c02992f611bb767b87fde41c99a0fd5b21f9826127b3fd47
3
+ size 84934656
model_weights/model.decoder.layers.self_attention.linear_proj.weight/88.0.7 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:20e4988ff874808c9e75b13aa1d5cb4f56d9b53d864bf143f9b760a218a55500
3
+ size 84934656
model_weights/model.decoder.layers.self_attention.linear_proj.weight/90.0.1 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:85b77492925385becfbc077cc1f2eb8021144f250e8814292240b8bc5566aa4b
3
+ size 84934656
model_weights/model.output_layer.weight/2.0 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:f0d1b78628c8349c84767a2ca87ce6b1b987d96f039dc951bb5ff23abc8b9c4d
3
+ size 1179648000