Upload folder using huggingface_hub
Browse files
.gitattributes
CHANGED
@@ -37,3 +37,8 @@ part_aa filter=lfs diff=lfs merge=lfs -text
|
|
37 |
part_ab filter=lfs diff=lfs merge=lfs -text
|
38 |
part_ac filter=lfs diff=lfs merge=lfs -text
|
39 |
adapted_model_dynamic_attention-Q5_K_M.gguf filter=lfs diff=lfs merge=lfs -text
|
|
|
|
|
|
|
|
|
|
|
|
37 |
part_ab filter=lfs diff=lfs merge=lfs -text
|
38 |
part_ac filter=lfs diff=lfs merge=lfs -text
|
39 |
adapted_model_dynamic_attention-Q5_K_M.gguf filter=lfs diff=lfs merge=lfs -text
|
40 |
+
solo-llama-3-maal-mlp-koen-8b-Q3_K_M.gguf filter=lfs diff=lfs merge=lfs -text
|
41 |
+
solo-llama-3-maal-mlp-koen-8b-Q4_K_M.gguf filter=lfs diff=lfs merge=lfs -text
|
42 |
+
solo-llama-3-maal-mlp-koen-8b-Q5_K_M.gguf filter=lfs diff=lfs merge=lfs -text
|
43 |
+
solo-llama-3-maal-mlp-koen-8b-Q6_K.gguf filter=lfs diff=lfs merge=lfs -text
|
44 |
+
solo-llama-3-maal-mlp-koen-8b-Q8_0.gguf filter=lfs diff=lfs merge=lfs -text
|
Modelfile_Q5_K_M
CHANGED
@@ -1,4 +1,4 @@
|
|
1 |
-
FROM
|
2 |
TEMPLATE """{{ if .System }}<|start_header_id|>system<|end_header_id|>
|
3 |
|
4 |
{{ .System }}<|eot_id|>{{ end }}{{ if .Prompt }}<|start_header_id|>user<|end_header_id|>
|
|
|
1 |
+
FROM solo-llama-3-maal-mlp-koen-8b-Q5_K_M.gguf
|
2 |
TEMPLATE """{{ if .System }}<|start_header_id|>system<|end_header_id|>
|
3 |
|
4 |
{{ .System }}<|eot_id|>{{ end }}{{ if .Prompt }}<|start_header_id|>user<|end_header_id|>
|
config.json
ADDED
@@ -0,0 +1,29 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
{
|
2 |
+
"_name_or_path": "maum-ai/Llama-3-MAAL-8B-Instruct-v0.1",
|
3 |
+
"architectures": [
|
4 |
+
"LlamaForCausalLM"
|
5 |
+
],
|
6 |
+
"attention_bias": false,
|
7 |
+
"attention_dropout": 0.0,
|
8 |
+
"bos_token_id": 128000,
|
9 |
+
"eos_token_id": 128001,
|
10 |
+
"hidden_act": "silu",
|
11 |
+
"hidden_size": 4096,
|
12 |
+
"initializer_range": 0.02,
|
13 |
+
"intermediate_size": 14336,
|
14 |
+
"max_position_embeddings": 8192,
|
15 |
+
"mlp_bias": false,
|
16 |
+
"model_type": "llama",
|
17 |
+
"num_attention_heads": 32,
|
18 |
+
"num_hidden_layers": 32,
|
19 |
+
"num_key_value_heads": 8,
|
20 |
+
"pretraining_tp": 1,
|
21 |
+
"rms_norm_eps": 1e-05,
|
22 |
+
"rope_scaling": null,
|
23 |
+
"rope_theta": 500000.0,
|
24 |
+
"tie_word_embeddings": false,
|
25 |
+
"torch_dtype": "bfloat16",
|
26 |
+
"transformers_version": "4.41.0",
|
27 |
+
"use_cache": false,
|
28 |
+
"vocab_size": 145039
|
29 |
+
}
|
solo-llama-3-maal-mlp-koen-8b-Q3_K_M.gguf
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:de4d2e6a1acace8666cf5643158a527ae0182785f14b3f79ffede4488299e535
|
3 |
+
size 4105321856
|
solo-llama-3-maal-mlp-koen-8b-Q4_K_M.gguf
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:caf8f1668fa1cbd90c544a11249726e59fe8942e7d47b7d527d4ce288f5340cc
|
3 |
+
size 5016268128
|
solo-llama-3-maal-mlp-koen-8b-Q5_K_M.gguf
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:1acd536b4123837aee2f43ffde8a697f842be5ab4d789ab6787a7887291c4bb3
|
3 |
+
size 5837114208
|
solo-llama-3-maal-mlp-koen-8b-Q6_K.gguf
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:92ebf8f6f99cbc2bad92532abb21b5fd27dd547eecd1b23367c3ac601933cede
|
3 |
+
size 6709263168
|
solo-llama-3-maal-mlp-koen-8b-Q8_0.gguf
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:dab512712e30332146d40fff18b0345ad024c78d7ebddfe1b760c73d04625982
|
3 |
+
size 8687325056
|