Upload folder using huggingface_hub
Browse files- .gitattributes +12 -0
- compiled/1633cf904f95b0038ce0.neff +3 -0
- compiled/45b1cdb7f1e25dfd80ee.neff +3 -0
- compiled/52c54ebdcc7728297417.neff +3 -0
- compiled/638f6e4774811f7e52c9.neff +3 -0
- compiled/73af1aa692fdc898383d.neff +3 -0
- compiled/9c074ee591ecac3f4135.neff +3 -0
- compiled/a3315ff272b9cdcd2b81.neff +3 -0
- compiled/a6add5d4178acd6bd44c.neff +3 -0
- compiled/a96751337a7365cf9f26.neff +3 -0
- compiled/ab4b2301092efdfcb139.neff +3 -0
- compiled/b1e52489b7020535cc04.neff +3 -0
- compiled/b7cc1a385fe0e918eb4e.neff +3 -0
- config.json +53 -0
- generation_config.json +14 -0
.gitattributes
CHANGED
@@ -33,3 +33,15 @@ saved_model/**/* filter=lfs diff=lfs merge=lfs -text
|
|
33 |
*.zip filter=lfs diff=lfs merge=lfs -text
|
34 |
*.zst filter=lfs diff=lfs merge=lfs -text
|
35 |
*tfevents* filter=lfs diff=lfs merge=lfs -text
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
33 |
*.zip filter=lfs diff=lfs merge=lfs -text
|
34 |
*.zst filter=lfs diff=lfs merge=lfs -text
|
35 |
*tfevents* filter=lfs diff=lfs merge=lfs -text
|
36 |
+
compiled/1633cf904f95b0038ce0.neff filter=lfs diff=lfs merge=lfs -text
|
37 |
+
compiled/45b1cdb7f1e25dfd80ee.neff filter=lfs diff=lfs merge=lfs -text
|
38 |
+
compiled/52c54ebdcc7728297417.neff filter=lfs diff=lfs merge=lfs -text
|
39 |
+
compiled/638f6e4774811f7e52c9.neff filter=lfs diff=lfs merge=lfs -text
|
40 |
+
compiled/73af1aa692fdc898383d.neff filter=lfs diff=lfs merge=lfs -text
|
41 |
+
compiled/9c074ee591ecac3f4135.neff filter=lfs diff=lfs merge=lfs -text
|
42 |
+
compiled/a3315ff272b9cdcd2b81.neff filter=lfs diff=lfs merge=lfs -text
|
43 |
+
compiled/a6add5d4178acd6bd44c.neff filter=lfs diff=lfs merge=lfs -text
|
44 |
+
compiled/a96751337a7365cf9f26.neff filter=lfs diff=lfs merge=lfs -text
|
45 |
+
compiled/ab4b2301092efdfcb139.neff filter=lfs diff=lfs merge=lfs -text
|
46 |
+
compiled/b1e52489b7020535cc04.neff filter=lfs diff=lfs merge=lfs -text
|
47 |
+
compiled/b7cc1a385fe0e918eb4e.neff filter=lfs diff=lfs merge=lfs -text
|
compiled/1633cf904f95b0038ce0.neff
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:08f1bac81a97ab48324698787ff678f66e26aeb6df5464fde1934c6bd0b2a617
|
3 |
+
size 1373184
|
compiled/45b1cdb7f1e25dfd80ee.neff
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:4fd7f61588f385c6074bf1b4f63ccec138dc058f57cf6613f767951ba8181058
|
3 |
+
size 1403904
|
compiled/52c54ebdcc7728297417.neff
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:8202a29a164b9df106b00efbea72ece46e68bbdef2524a7c46c50948eb39e9d1
|
3 |
+
size 1557504
|
compiled/638f6e4774811f7e52c9.neff
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:30f466c151d82a8ab9e22e9d6e8116d29e50023cfa51e9c0d24b28e964b60314
|
3 |
+
size 1393664
|
compiled/73af1aa692fdc898383d.neff
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:14b1347cd07f30044ec23b608f4fa7610ee3266f7002860fec29fffe0e15dc6c
|
3 |
+
size 1362944
|
compiled/9c074ee591ecac3f4135.neff
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:504a0966e58fa1cf7e3ad30f7a53c25bd92b6bd87474374b726b14317450ed5b
|
3 |
+
size 10005504
|
compiled/a3315ff272b9cdcd2b81.neff
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:3c19a646f3703ab869a15cd213de4ff5ba1d701b48a91a389b7dd09ee8bcc588
|
3 |
+
size 1567744
|
compiled/a6add5d4178acd6bd44c.neff
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:5361a1706e2619295e7f815f73fe878332e3321b42bfd987986eff6d12cb9e14
|
3 |
+
size 2038784
|
compiled/a96751337a7365cf9f26.neff
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:2599f9939f052e81f0a32c97256fde5a46959c860311d34cc498e9203d893918
|
3 |
+
size 4250624
|
compiled/ab4b2301092efdfcb139.neff
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:a4e78f202610e078a5886c01be99b42fe0e42a4a8cb37b79c4e8f265a7b4ff30
|
3 |
+
size 1721344
|
compiled/b1e52489b7020535cc04.neff
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:08135ce51918ccbc05c8d5820708468156e47c0b976905a147dd2c37efe38423
|
3 |
+
size 1465344
|
compiled/b7cc1a385fe0e918eb4e.neff
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:76081b0e48dbce122c28416c68d6367a4e4a81886373e42b86cd457ab9bc4f2c
|
3 |
+
size 1383424
|
config.json
ADDED
@@ -0,0 +1,53 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
{
|
2 |
+
"_name_or_path": "vietphuon/Llama-3.2-1B-Instruct-alpaca-then-quizgen-16bit",
|
3 |
+
"architectures": [
|
4 |
+
"LlamaForCausalLM"
|
5 |
+
],
|
6 |
+
"attention_bias": false,
|
7 |
+
"attention_dropout": 0.0,
|
8 |
+
"bos_token_id": 128000,
|
9 |
+
"eos_token_id": [
|
10 |
+
128001,
|
11 |
+
128008,
|
12 |
+
128009
|
13 |
+
],
|
14 |
+
"head_dim": 64,
|
15 |
+
"hidden_act": "silu",
|
16 |
+
"hidden_size": 2048,
|
17 |
+
"initializer_range": 0.02,
|
18 |
+
"intermediate_size": 8192,
|
19 |
+
"max_position_embeddings": 131072,
|
20 |
+
"mlp_bias": false,
|
21 |
+
"model_type": "llama",
|
22 |
+
"neuron": {
|
23 |
+
"auto_cast_type": "fp16",
|
24 |
+
"batch_size": 1,
|
25 |
+
"checkpoint_id": "vietphuon/Llama-3.2-1B-Instruct-alpaca-then-quizgen-16bit",
|
26 |
+
"checkpoint_revision": "a01e6bf5efa30f940147bf0f1d038e32af8d46d6",
|
27 |
+
"compiler_type": "neuronx-cc",
|
28 |
+
"compiler_version": "2.15.128.0+56dc5a86",
|
29 |
+
"num_cores": 2,
|
30 |
+
"sequence_length": 4096,
|
31 |
+
"task": "text-generation"
|
32 |
+
},
|
33 |
+
"num_attention_heads": 32,
|
34 |
+
"num_hidden_layers": 16,
|
35 |
+
"num_key_value_heads": 8,
|
36 |
+
"pad_token_id": 128004,
|
37 |
+
"pretraining_tp": 1,
|
38 |
+
"rms_norm_eps": 1e-05,
|
39 |
+
"rope_scaling": {
|
40 |
+
"factor": 32.0,
|
41 |
+
"high_freq_factor": 4.0,
|
42 |
+
"low_freq_factor": 1.0,
|
43 |
+
"original_max_position_embeddings": 8192,
|
44 |
+
"rope_type": "llama3"
|
45 |
+
},
|
46 |
+
"rope_theta": 500000.0,
|
47 |
+
"tie_word_embeddings": true,
|
48 |
+
"torch_dtype": "float16",
|
49 |
+
"transformers_version": "4.43.2",
|
50 |
+
"unsloth_version": "2024.10.3",
|
51 |
+
"use_cache": true,
|
52 |
+
"vocab_size": 128256
|
53 |
+
}
|
generation_config.json
ADDED
@@ -0,0 +1,14 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
{
|
2 |
+
"bos_token_id": 128000,
|
3 |
+
"do_sample": true,
|
4 |
+
"eos_token_id": [
|
5 |
+
128001,
|
6 |
+
128008,
|
7 |
+
128009
|
8 |
+
],
|
9 |
+
"max_length": 131072,
|
10 |
+
"pad_token_id": 128004,
|
11 |
+
"temperature": 0.6,
|
12 |
+
"top_p": 0.9,
|
13 |
+
"transformers_version": "4.43.2"
|
14 |
+
}
|