Transformers
GGUF
English
Russian
Not-For-All-Audiences
nsfw
LakoMoor commited on
Commit
8a2a674
1 Parent(s): 8a4d49b

Upload folder using huggingface_hub

Browse files
.gitattributes CHANGED
@@ -33,3 +33,12 @@ saved_model/**/* filter=lfs diff=lfs merge=lfs -text
33
  *.zip filter=lfs diff=lfs merge=lfs -text
34
  *.zst filter=lfs diff=lfs merge=lfs -text
35
  *tfevents* filter=lfs diff=lfs merge=lfs -text
 
 
 
 
 
 
 
 
 
 
33
  *.zip filter=lfs diff=lfs merge=lfs -text
34
  *.zst filter=lfs diff=lfs merge=lfs -text
35
  *tfevents* filter=lfs diff=lfs merge=lfs -text
36
+ assets/alice.png filter=lfs diff=lfs merge=lfs -text
37
+ assets/masha.png filter=lfs diff=lfs merge=lfs -text
38
+ silicon-masha-7b.Q2_K.gguf filter=lfs diff=lfs merge=lfs -text
39
+ silicon-masha-7b.Q3_K_M.gguf filter=lfs diff=lfs merge=lfs -text
40
+ silicon-masha-7b.Q4_0.gguf filter=lfs diff=lfs merge=lfs -text
41
+ silicon-masha-7b.Q4_K_M.gguf filter=lfs diff=lfs merge=lfs -text
42
+ silicon-masha-7b.Q5_0.gguf filter=lfs diff=lfs merge=lfs -text
43
+ silicon-masha-7b.Q5_K_M.gguf filter=lfs diff=lfs merge=lfs -text
44
+ silicon-masha-7b.Q6_K.gguf filter=lfs diff=lfs merge=lfs -text
.ipynb_checkpoints/README-checkpoint.md ADDED
@@ -0,0 +1,57 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ ---
2
+ language:
3
+ - en
4
+ - ru
5
+ tags:
6
+ - not-for-all-audiences
7
+ - nsfw
8
+ base_model:
9
+ - LakoMoor/Silicon-Masha-7B
10
+ license: cc-by-nc-4.0
11
+ inference: false
12
+ library_name: transformers
13
+ model_creator: LakoMoor
14
+ model_name: Silicon-Masha-7B
15
+ model_type: mistral
16
+ ---
17
+
18
+ # Silicon-Masha-7B-GGUF
19
+ ![Silicon-Masha-7B](https://huggingface.co/LakoMoor/Silicon-Masha-7B-GGUF/resolve/main/assets/masha.png)
20
+
21
+ ## What's that?
22
+
23
+ Silicon-Masha-7B-GGUF is a quantized model aiming to be both strong in RP, be smart **and** understand Russian, that can follow character maps very well. It is suitable for RP/ERP and general use.It can be run on weak iron using [llama.cpp](https://github.com/ggerganov/llama.cpp) or [koboldcpp](https://github.com/LostRuins/koboldcpp).
24
+
25
+ ## Prompt Template (Alpaca)
26
+
27
+ I found the best SillyTavern results from using the Noromaid template but please try other templates! Let me know if you find anything good.
28
+
29
+ SillyTavern config files: [Context](https://huggingface.co/LakoMoor/Silicon-Masha-7B-GGUF/resolve/main/assets/context.json), [Instruct](https://huggingface.co/LakoMoor/Silicon-Masha-7B-GGUF/resolve/main/assets/instruct.json).
30
+
31
+ Additionally, here is my highly recommended [Text Completion preset](https://huggingface.co/LakoMoor/Silicon-Masha-7B-GGUF/resolve/main/assets/MinP.json). You can tweak this by adjusting temperature up or dropping min p to boost creativity or raise min p to increase stability. You shouldn't need to touch anything else!
32
+
33
+ ```
34
+ Below is an instruction that describes a task. Write a response that appropriately completes the request.
35
+ ### Instruction:
36
+ {prompt}
37
+ ### Response:
38
+ ```
39
+
40
+ ## Provided files
41
+
42
+ | Name | Quant method | Bits | Use case |
43
+ | ---- | ---- | ---- | ----- |
44
+ | [silicon-masha-7b.Q2_K.gguf](https://huggingface.co/LakoMoor/Silicon-Masha-7B-GGUF/blob/main/silicon-masha-7b.Q2_K.gguf) | Q2_K | 2 | smallest, significant quality loss - not recommended for most purposes |
45
+ | [silicon-masha-7b.Q3_K_M.gguf](https://huggingface.co/LakoMoor/Silicon-Masha-7B-GGUF/blob/main/silicon-masha-7b.Q3_K_M.gguf) | Q3_K_M | 3 | very small, high quality loss |
46
+ | [silicon-masha-7b.Q4_0.gguf](https://huggingface.co/LakoMoor/Silicon-Masha-7B-GGUF/blob/main/silicon-masha-7b.Q4_0.gguf) | Q4_0 | 4 | legacy; small, very high quality loss - prefer using Q3_K_M |
47
+ | [silicon-masha-7b.Q4_K_M.gguf](https://huggingface.co/LakoMoor/Silicon-Masha-7B-GGUF/blob/main/silicon-masha-7b.Q4_K_M.gguf) | Q4_K_M | 4 | medium, balanced quality - recommended |
48
+ | [silicon-masha-7b.Q5_0.gguf](https://huggingface.co/LakoMoor/Silicon-Masha-7B-GGUF/blob/main/silicon-masha-7b.Q5_0.gguf) | Q5_0 | 5 | legacy; medium, balanced quality - prefer using Q4_K_M |
49
+ | [silicon-masha-7b.Q5_K_M.gguf](https://huggingface.co/LakoMoor/Silicon-Masha-7B-GGUF/blob/main/silicon-masha-7b.Q5_K_M.gguf) | Q5_K_M | 5 | large, very low quality loss - recommended |
50
+ | [silicon-masha-7b.Q6_K.gguf](https://huggingface.co/LakoMoor/Silicon-Masha-7B-GGUF/blob/main/silicon-masha-7b.Q6_K.gguf) | Q6_K | 6 | very large, extremely low quality loss |
51
+ | [silicon-masha-7b.Q8_0.gguf](https://huggingface.co/LakoMoor/Silicon-Masha-7B-GGUF/blob/main/silicon-masha-7b.Q8_0.gguf) | Q8_0 | 8 | very large, extremely low quality loss - not recommended |
52
+
53
+ ## How run it ?
54
+ [llama.cpp](https://github.com/ggerganov/llama.cpp)
55
+ ```shell
56
+ ./main -ngl 35 -m silicon-masha-7b.Q4_K_M.gguf --color -c 32768 --temp 0.7 --repeat_penalty 1.1 -n -1 -p "Below is an instruction that describes a task. Write a response that appropriately completes the request.\n{system_message}\n### Instruction:{prompt}\n### Response:\n"
57
+ ```
README.md ADDED
@@ -0,0 +1,57 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ ---
2
+ language:
3
+ - en
4
+ - ru
5
+ tags:
6
+ - not-for-all-audiences
7
+ - nsfw
8
+ base_model:
9
+ - LakoMoor/Silicon-Masha-7B
10
+ license: cc-by-nc-4.0
11
+ inference: false
12
+ library_name: transformers
13
+ model_creator: LakoMoor
14
+ model_name: Silicon-Masha-7B
15
+ model_type: mistral
16
+ ---
17
+
18
+ # Silicon-Masha-7B-GGUF
19
+ ![Silicon-Masha-7B](https://huggingface.co/LakoMoor/Silicon-Masha-7B-GGUF/resolve/main/assets/masha.png)
20
+
21
+ ## What's that?
22
+
23
+ Silicon-Masha-7B-GGUF is a quantized model aiming to be both strong in RP, be smart **and** understand Russian, that can follow character maps very well. It is suitable for RP/ERP and general use.It can be run on weak iron using [llama.cpp](https://github.com/ggerganov/llama.cpp) or [koboldcpp](https://github.com/LostRuins/koboldcpp).
24
+
25
+ ## Prompt Template (Alpaca)
26
+
27
+ I found the best SillyTavern results from using the Noromaid template but please try other templates! Let me know if you find anything good.
28
+
29
+ SillyTavern config files: [Context](https://huggingface.co/LakoMoor/Silicon-Masha-7B-GGUF/resolve/main/assets/context.json), [Instruct](https://huggingface.co/LakoMoor/Silicon-Masha-7B-GGUF/resolve/main/assets/instruct.json).
30
+
31
+ Additionally, here is my highly recommended [Text Completion preset](https://huggingface.co/LakoMoor/Silicon-Masha-7B-GGUF/resolve/main/assets/MinP.json). You can tweak this by adjusting temperature up or dropping min p to boost creativity or raise min p to increase stability. You shouldn't need to touch anything else!
32
+
33
+ ```
34
+ Below is an instruction that describes a task. Write a response that appropriately completes the request.
35
+ ### Instruction:
36
+ {prompt}
37
+ ### Response:
38
+ ```
39
+
40
+ ## Provided files
41
+
42
+ | Name | Quant method | Bits | Use case |
43
+ | ---- | ---- | ---- | ----- |
44
+ | [silicon-masha-7b.Q2_K.gguf](https://huggingface.co/LakoMoor/Silicon-Masha-7B-GGUF/blob/main/silicon-masha-7b.Q2_K.gguf) | Q2_K | 2 | smallest, significant quality loss - not recommended for most purposes |
45
+ | [silicon-masha-7b.Q3_K_M.gguf](https://huggingface.co/LakoMoor/Silicon-Masha-7B-GGUF/blob/main/silicon-masha-7b.Q3_K_M.gguf) | Q3_K_M | 3 | very small, high quality loss |
46
+ | [silicon-masha-7b.Q4_0.gguf](https://huggingface.co/LakoMoor/Silicon-Masha-7B-GGUF/blob/main/silicon-masha-7b.Q4_0.gguf) | Q4_0 | 4 | legacy; small, very high quality loss - prefer using Q3_K_M |
47
+ | [silicon-masha-7b.Q4_K_M.gguf](https://huggingface.co/LakoMoor/Silicon-Masha-7B-GGUF/blob/main/silicon-masha-7b.Q4_K_M.gguf) | Q4_K_M | 4 | medium, balanced quality - recommended |
48
+ | [silicon-masha-7b.Q5_0.gguf](https://huggingface.co/LakoMoor/Silicon-Masha-7B-GGUF/blob/main/silicon-masha-7b.Q5_0.gguf) | Q5_0 | 5 | legacy; medium, balanced quality - prefer using Q4_K_M |
49
+ | [silicon-masha-7b.Q5_K_M.gguf](https://huggingface.co/LakoMoor/Silicon-Masha-7B-GGUF/blob/main/silicon-masha-7b.Q5_K_M.gguf) | Q5_K_M | 5 | large, very low quality loss - recommended |
50
+ | [silicon-masha-7b.Q6_K.gguf](https://huggingface.co/LakoMoor/Silicon-Masha-7B-GGUF/blob/main/silicon-masha-7b.Q6_K.gguf) | Q6_K | 6 | very large, extremely low quality loss |
51
+ | [silicon-masha-7b.Q8_0.gguf](https://huggingface.co/LakoMoor/Silicon-Masha-7B-GGUF/blob/main/silicon-masha-7b.Q8_0.gguf) | Q8_0 | 8 | very large, extremely low quality loss - not recommended |
52
+
53
+ ## How run it ?
54
+ [llama.cpp](https://github.com/ggerganov/llama.cpp)
55
+ ```shell
56
+ ./main -ngl 35 -m silicon-masha-7b.Q4_K_M.gguf --color -c 32768 --temp 0.7 --repeat_penalty 1.1 -n -1 -p "Below is an instruction that describes a task. Write a response that appropriately completes the request.\n{system_message}\n### Instruction:{prompt}\n### Response:\n"
57
+ ```
assets/MinP.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"temp": 0.8, "temperature_last": false, "top_p": 0.98, "top_k": 100, "top_a": 0, "tfs": 1, "epsilon_cutoff": 0, "eta_cutoff": 0, "typical_p": 1, "min_p": 0.05, "rep_pen": 1.1, "rep_pen_range": 2048, "no_repeat_ngram_size": 0, "penalty_alpha": 0, "num_beams": 1, "length_penalty": 1, "min_length": 0, "encoder_rep_pen": 1, "freq_pen": 0, "presence_pen": 0, "do_sample": true, "early_stopping": false, "add_bos_token": true, "truncation_length": 2048, "ban_eos_token": false, "skip_special_tokens": true, "streaming": true, "mirostat_mode": 0, "mirostat_tau": 5, "mirostat_eta": 0.1, "guidance_scale": 1, "negative_prompt": "", "grammar_string": "", "banned_tokens": "", "ignore_eos_token_aphrodite": false, "spaces_between_special_tokens_aphrodite": true, "type": "koboldcpp", "legacy_api": false, "sampler_order": [6, 0, 5, 3, 2, 1, 4], "n": 1, "rep_pen_size": 0, "genamt": 250, "max_length": 1024}
assets/alice.png ADDED

Git LFS Details

  • SHA256: 7b7ba45b4da02d3d98c766ad14e77fa427b9f9eb3e0ae97e1e4f0b79a30f8c20
  • Pointer size: 132 Bytes
  • Size of remote file: 1.39 MB
assets/context.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"story_string": "### Instruction:\nWrite {{char}}'s next reply in a fictional roleplay chat between {{user}} and {{char}}. Use the provided character sheet and example dialogue for formatting direction and character speech patterns.\n\n{{#if system}}{{system}}\n\n{{/if}}{{#if wiBefore}}{{wiBefore}}\n\n{{/if}}Description of {{char}}:\n{{#if description}}{{description}}\n{{/if}}{{#if personality}}{{char}}'s personality: {{personality}}\n\n{{/if}}{{#if scenario}}Scenario: {{scenario}}\n\n{{/if}}{{#if persona}}Description of {{user}}: {{persona}}\n\n{{/if}}Play the role of {{char}}\n\n{{#if wiAfter}}{{wiAfter}}\n{{/if}}", "example_separator": "Example roleplay chat:", "chat_start": "Taking the above information into consideration,\nyou must engage in a roleplay conversation with {{user}} below this line.\nDo not write {{user}}'s dialogue lines in your responses.\n", "always_force_name2": true, "trim_sentences": true, "include_newline": true, "single_line": false, "name": "Context-Alice"}
assets/instruct.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"system_prompt": "Avoid repetition, don't loop. Develop the plot slowly, always stay in character. Describe all actions in full, elaborate, explicit, graphic, and vivid detail. Mention all relevant sensory perceptions.", "input_sequence": "\n### Instruction: (Style: Markdown, Present Tense)", "output_sequence": "\n### Response: (Style: Markdown, Present Tense)", "first_output_sequence": "### Response:", "last_output_sequence": "", "system_sequence_prefix": "", "system_sequence_suffix": "", "stop_sequence": "", "separator_sequence": "", "wrap": true, "macro": true, "names": true, "names_force_groups": true, "activation_regex": "", "name": "Instruct-Alice"}
assets/masha.png ADDED

Git LFS Details

  • SHA256: c8091805f5c8e45644553b28fd15b69f42f199453a40fa25627d6d94140f0488
  • Pointer size: 132 Bytes
  • Size of remote file: 1.59 MB
silicon-masha-7b.Q2_K.gguf ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:df4136053330117d6e90c884534a6fff6e4afafb9c0cd15506a441c4d51e8c8f
3
+ size 2719241920
silicon-masha-7b.Q3_K_M.gguf ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:a946771d4f2c6577b1fecbf4612f70ce5da30285f98fa6d25b9be4bd2b3745ce
3
+ size 3518985920
silicon-masha-7b.Q4_0.gguf ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:2aa5bc5bb37cafc4d6a52d9c264ae744410326539fd67e7c4dab10933e011d18
3
+ size 4108916416
silicon-masha-7b.Q4_K_M.gguf ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:736ef2db1b4cb1d1ae93e45189cf3c164d604083c13aa4588f00818310e22263
3
+ size 4368438976
silicon-masha-7b.Q5_0.gguf ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:f07122686c0e0d9818381008db755237405ae8acb29327a48dc5102d92b8d52b
3
+ size 4997715648
silicon-masha-7b.Q5_K_M.gguf ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:19ffd5e475a38ba0ec1f845a8ff231635ba8a494f576b4086d9d036ccae55b50
3
+ size 5131409088
silicon-masha-7b.Q6_K.gguf ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:1c5eaf2870e9ea63648bb584091803a1bad1261a291da8d68703992810b36dff
3
+ size 5942064832