Upload folder using huggingface_hub
Browse files- adapter_0.pt +3 -0
- adapter_1.pt +3 -0
- adapter_config.json +1 -0
- adapter_model.bin +3 -0
- config.json +1 -0
- hf_model_0001_0.pt +3 -0
- hf_model_0001_1.pt +3 -0
- hf_model_0002_0.pt +3 -0
- hf_model_0002_1.pt +3 -0
- hf_model_0003_0.pt +3 -0
- hf_model_0003_1.pt +3 -0
- hf_model_0004_0.pt +3 -0
- hf_model_0004_1.pt +3 -0
- recipe_state.pt +3 -0
adapter_0.pt
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:83c15b4dd4fa28eac3290b7ac92a2e90e3055c5bc617a2f2033edef475eda7ac
|
3 |
+
size 9500346
|
adapter_1.pt
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:b55ddd6c570924cc63447a03925a156ae1a984cd1d3c893106c72a74039812b4
|
3 |
+
size 9500346
|
adapter_config.json
ADDED
@@ -0,0 +1 @@
|
|
|
|
|
1 |
+
{"r": 8, "lora_alpha": 8, "target_modules": ["q_proj", "k_proj", "v_proj"], "peft_type": "LORA"}
|
adapter_model.bin
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:94b8aa91027fb9d0f90357d394c2f5f5901476b9d08136b8d2e166dbd7d7ae67
|
3 |
+
size 9506506
|
config.json
ADDED
@@ -0,0 +1 @@
|
|
|
|
|
1 |
+
{"architectures": ["LlamaForCausalLM"], "attention_bias": false, "attention_dropout": 0.0, "bos_token_id": 128000, "eos_token_id": [128001, 128008, 128009], "hidden_act": "silu", "hidden_size": 4096, "initializer_range": 0.02, "intermediate_size": 14336, "max_position_embeddings": 131072, "mlp_bias": false, "model_type": "llama", "num_attention_heads": 32, "num_hidden_layers": 32, "num_key_value_heads": 8, "pretraining_tp": 1, "rms_norm_eps": 1e-05, "rope_scaling": {"factor": 8.0, "low_freq_factor": 1.0, "high_freq_factor": 4.0, "original_max_position_embeddings": 8192, "rope_type": "llama3"}, "rope_theta": 500000.0, "tie_word_embeddings": false, "torch_dtype": "bfloat16", "transformers_version": "4.42.3", "use_cache": true, "vocab_size": 128256}
|
hf_model_0001_0.pt
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:22483e1a2a024b9fb1a53a15e57b874317c1e1361cc2d96940187856a0f7fde9
|
3 |
+
size 4976716084
|
hf_model_0001_1.pt
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:eca16bfb56e884c9ed1b43f32118dfc238eb25b1f97f2ca8173ae85eece23caa
|
3 |
+
size 4976716084
|
hf_model_0002_0.pt
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:b3aa67034c0dbc35f4456f9954637acac628d1b00debf98e4236f5fc6a4e83d7
|
3 |
+
size 4999824766
|
hf_model_0002_1.pt
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:c59042ad0bac79f724c2161a93f60a42cc262bae313a526ed72bc6411582d8ec
|
3 |
+
size 4999824766
|
hf_model_0003_0.pt
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:1da210113437d2e3a3553ad8825aa597b7b1bc87d26906bd80deccff08aa428b
|
3 |
+
size 4915937322
|
hf_model_0003_1.pt
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:73f7153b67585712801ea36e1a83dff6d6e27703f8eacc482db86ae7db504475
|
3 |
+
size 4915937322
|
hf_model_0004_0.pt
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:13e98ba8ca7f4e285361e1d2496145c0b8f19802f3c4ee6d6a599bf3b9663e78
|
3 |
+
size 1168140692
|
hf_model_0004_1.pt
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:499834ffbb1bcdae8957a2ac03db65af08a69f41470487af0308ecc926c678e7
|
3 |
+
size 1168140692
|
recipe_state.pt
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:d1e61e331a37e6f09d6f9bf97296364171abc4b689c49e054ac08d16c12ade2e
|
3 |
+
size 19038214
|