NeMo
nvidia
jiaqiz commited on
Commit
ba90ea2
·
verified ·
1 Parent(s): 770d3d4

Add files using large-upload tool

Browse files
Files changed (1) hide show
  1. .gitattributes +25 -0
.gitattributes CHANGED
@@ -1432,3 +1432,28 @@ model_weights/model.decoder.layers.self_attention.linear_qkv.weight/4.1.0 filter
1432
  model_weights/model.decoder.layers.self_attention.linear_qkv.weight/70.6.0 filter=lfs diff=lfs merge=lfs -text
1433
  model_weights/model.decoder.layers.self_attention.linear_qkv.weight/15.4.0 filter=lfs diff=lfs merge=lfs -text
1434
  model_weights/model.decoder.layers.self_attention.linear_qkv.weight/35.3.0 filter=lfs diff=lfs merge=lfs -text
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1432
  model_weights/model.decoder.layers.self_attention.linear_qkv.weight/70.6.0 filter=lfs diff=lfs merge=lfs -text
1433
  model_weights/model.decoder.layers.self_attention.linear_qkv.weight/15.4.0 filter=lfs diff=lfs merge=lfs -text
1434
  model_weights/model.decoder.layers.self_attention.linear_qkv.weight/35.3.0 filter=lfs diff=lfs merge=lfs -text
1435
+ model_weights/model.decoder.layers.self_attention.linear_qkv.weight/70.2.0 filter=lfs diff=lfs merge=lfs -text
1436
+ model_weights/model.decoder.layers.self_attention.linear_qkv.weight/68.6.0 filter=lfs diff=lfs merge=lfs -text
1437
+ model_weights/model.decoder.layers.self_attention.linear_qkv.weight/67.7.0 filter=lfs diff=lfs merge=lfs -text
1438
+ model_weights/model.decoder.layers.self_attention.linear_qkv.weight/87.4.0 filter=lfs diff=lfs merge=lfs -text
1439
+ model_weights/model.decoder.layers.self_attention.linear_qkv.weight/52.4.0 filter=lfs diff=lfs merge=lfs -text
1440
+ model_weights/model.decoder.layers.self_attention.linear_qkv.weight/44.1.0 filter=lfs diff=lfs merge=lfs -text
1441
+ model_weights/model.decoder.layers.self_attention.linear_qkv.weight/52.7.0 filter=lfs diff=lfs merge=lfs -text
1442
+ model_weights/model.decoder.layers.self_attention.linear_qkv.weight/60.2.0 filter=lfs diff=lfs merge=lfs -text
1443
+ model_weights/model.decoder.layers.self_attention.linear_qkv.weight/72.6.0 filter=lfs diff=lfs merge=lfs -text
1444
+ model_weights/model.decoder.layers.self_attention.linear_qkv.weight/30.1.0 filter=lfs diff=lfs merge=lfs -text
1445
+ model_weights/model.decoder.layers.self_attention.linear_qkv.weight/64.2.0 filter=lfs diff=lfs merge=lfs -text
1446
+ model_weights/model.decoder.layers.self_attention.linear_qkv.weight/80.6.0 filter=lfs diff=lfs merge=lfs -text
1447
+ model_weights/model.decoder.layers.self_attention.linear_qkv.weight/30.2.0 filter=lfs diff=lfs merge=lfs -text
1448
+ model_weights/model.decoder.layers.self_attention.linear_qkv.weight/80.3.0 filter=lfs diff=lfs merge=lfs -text
1449
+ model_weights/model.decoder.layers.self_attention.linear_qkv.weight/24.5.0 filter=lfs diff=lfs merge=lfs -text
1450
+ model_weights/model.decoder.layers.self_attention.linear_qkv.weight/54.3.0 filter=lfs diff=lfs merge=lfs -text
1451
+ model_weights/model.decoder.layers.self_attention.linear_qkv.weight/19.1.0 filter=lfs diff=lfs merge=lfs -text
1452
+ model_weights/model.decoder.layers.self_attention.linear_qkv.weight/38.3.0 filter=lfs diff=lfs merge=lfs -text
1453
+ model_weights/model.decoder.layers.self_attention.linear_qkv.weight/86.0.0 filter=lfs diff=lfs merge=lfs -text
1454
+ model_weights/model.decoder.layers.self_attention.linear_qkv.weight/24.4.0 filter=lfs diff=lfs merge=lfs -text
1455
+ model_weights/model.decoder.layers.self_attention.linear_qkv.weight/44.2.0 filter=lfs diff=lfs merge=lfs -text
1456
+ model_weights/model.decoder.layers.self_attention.linear_qkv.weight/10.1.0 filter=lfs diff=lfs merge=lfs -text
1457
+ model_weights/model.decoder.layers.self_attention.linear_qkv.weight/57.7.0 filter=lfs diff=lfs merge=lfs -text
1458
+ model_weights/model.decoder.layers.self_attention.linear_qkv.weight/82.1.0 filter=lfs diff=lfs merge=lfs -text
1459
+ model_weights/model.decoder.layers.self_attention.linear_qkv.weight/90.1.0 filter=lfs diff=lfs merge=lfs -text