invisietch commited on
Commit
8a17a96
·
verified ·
1 Parent(s): a082964

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +67 -35
README.md CHANGED
@@ -4,54 +4,86 @@ tags:
4
  - not-for-all-audiences
5
  - axolotl
6
  - qlora
7
- - llama-cpp
8
- - gguf-my-repo
9
  language:
10
  - en
11
  license: other
12
- base_model: invisietch/MiS-Firefly-v0.2-22B
13
  ---
14
 
15
- # invisietch/MiS-Firefly-v0.2-22B-Q6_K-GGUF
16
- This model was converted to GGUF format from [`invisietch/MiS-Firefly-v0.2-22B`](https://huggingface.co/invisietch/MiS-Firefly-v0.2-22B) using llama.cpp via the ggml.ai's [GGUF-my-repo](https://huggingface.co/spaces/ggml-org/gguf-my-repo) space.
17
- Refer to the [original model card](https://huggingface.co/invisietch/MiS-Firefly-v0.2-22B) for more details on the model.
18
 
19
- ## Use with llama.cpp
20
- Install llama.cpp through brew (works on Mac and Linux)
21
 
22
- ```bash
23
- brew install llama.cpp
 
 
 
 
 
24
 
25
- ```
26
- Invoke the llama.cpp server or the CLI.
27
 
28
- ### CLI:
29
- ```bash
30
- llama-cli --hf-repo invisietch/MiS-Firefly-v0.2-22B-Q6_K-GGUF --hf-file mis-firefly-v0.2-22b-q6_k.gguf -p "The meaning to life and the universe is"
31
- ```
32
 
33
- ### Server:
34
- ```bash
35
- llama-server --hf-repo invisietch/MiS-Firefly-v0.2-22B-Q6_K-GGUF --hf-file mis-firefly-v0.2-22b-q6_k.gguf -c 2048
36
- ```
37
 
38
- Note: You can also use this checkpoint directly through the [usage steps](https://github.com/ggerganov/llama.cpp?tab=readme-ov-file#usage) listed in the Llama.cpp repo as well.
 
39
 
40
- Step 1: Clone llama.cpp from GitHub.
41
- ```
42
- git clone https://github.com/ggerganov/llama.cpp
43
- ```
44
 
45
- Step 2: Move into the llama.cpp folder and build it with `LLAMA_CURL=1` flag along with other hardware-specific flags (for ex: LLAMA_CUDA=1 for Nvidia GPUs on Linux).
46
- ```
47
- cd llama.cpp && LLAMA_CURL=1 make
48
- ```
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
49
 
50
- Step 3: Run inference through the main binary.
51
- ```
52
- ./llama-cli --hf-repo invisietch/MiS-Firefly-v0.2-22B-Q6_K-GGUF --hf-file mis-firefly-v0.2-22b-q6_k.gguf -p "The meaning to life and the universe is"
53
- ```
54
- or
55
  ```
56
- ./llama-server --hf-repo invisietch/MiS-Firefly-v0.2-22B-Q6_K-GGUF --hf-file mis-firefly-v0.2-22b-q6_k.gguf -c 2048
57
  ```
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
4
  - not-for-all-audiences
5
  - axolotl
6
  - qlora
 
 
7
  language:
8
  - en
9
  license: other
 
10
  ---
11
 
12
+ <div align="center">
13
+ <b style="font-size: 36px;">MiS-Firefly-v0.2-22B (Q6_K)</b>
 
14
 
15
+ <img src="https://huggingface.co/invisietch/MiS-Firefly-v0.2-22B/resolve/main/header.png" style="width:60%">
 
16
 
17
+ <b>HF</b> :
18
+ <a href="https://huggingface.co/invisietch/MiS-Firefly-v0.2-22B">FP16</a>
19
+ &vert;
20
+ <b>GGUF</b> :
21
+ <a href="https://huggingface.co/invisietch/MiS-Firefly-v0.2-22B-Q6_K-GGUF">Q6_K</a> &middot;
22
+ <a href="https://huggingface.co/invisietch/MiS-Firefly-v0.2-22B-Q4_K_M-GGUF">Q4_K_M</a>
23
+ </div>
24
 
25
+ # Model Details
 
26
 
27
+ **This is a fix for the quantization issue in Firefly v0.1.**
 
 
 
28
 
29
+ Firefly is a Mistral Small 22B finetune designed for creative writing and roleplay. The model is largely uncensored and should support
30
+ context up to 32,768 tokens.
 
 
31
 
32
+ The model has been tested in various roleplay scenarios up to 16k context, as well as in a role as an assistant. It shows a broad
33
+ competency &amp; coherence across various scenarios.
34
 
35
+ Special thanks to <a href="https://huggingface.co/SicariusSicariiStuff">SicariusSicariiStuff</a> for bouncing ideas back &amp; forth on
36
+ training, and <a href="https://huggingface.co/SytanSD">SytanSD</a> for quants.
 
 
37
 
38
+ # Feedback
39
+
40
+ I appreciate all feedback on any of my models, you can use:
41
+
42
+ * [My Discord server](https://discord.gg/AJwZuu7Ncx) - requires Discord.
43
+ * [The Community tab](https://huggingface.co/invisietch/MiS-Firefly-v0.2-22B/discussions) - requires HF login.
44
+ * Discord DMs to **invisietch**.
45
+
46
+ Your feedback is how I improve these models for future versions.
47
+
48
+ # Disclaimer
49
+
50
+ This model is extensively uncensored. It can generate explicit, disturbing or offensive responses. Use responsibly. I am not responsible for
51
+ your use of this model.
52
+
53
+ This model is a finetune of Mistral Small 22B (2409) and usage must follow the terms of Mistral's license. By downloading this model, you
54
+ agree not to use it for commercial purposes unless you have a valid Mistral commercial license. See [the base model card](https://huggingface.co/mistralai/Mistral-Small-Instruct-2409)
55
+ for more details.
56
+
57
+ # Prompting Format
58
+
59
+ I'd recommend Mistral v2v3 prompting format:
60
 
 
 
 
 
 
61
  ```
62
+ <s>[INST] User message here.[/INST] Bot response here</s>[INST] User message 2 here.
63
  ```
64
+
65
+ # Sampler Settings
66
+
67
+ I'm running the following sampler settings but this is an RC and they may not be optimal.
68
+
69
+ - **Temperature:** Dynamic 0.7-1.1
70
+ - **Min-P:** 0.07
71
+ - **Rep Pen:** 1.08
72
+ - **Rep Pen Range:** 1536
73
+ - **XTC:** 0.1/0.15
74
+
75
+ If you get completely incoherent responses, feel free to use these as a starting point.
76
+
77
+ # Training Strategy
78
+
79
+ I started with a finetune of Mistral Small 22B which had been trained on the Gutenberg dataset: [nbeerbower/Mistral-Small-Gutenberg-Doppel-22B](https://huggingface.co/nbeerbower/Mistral-Small-Gutenberg-Doppel-22B).
80
+
81
+ The first stage of my training was a single epoch at low LR over a 474 million token text completion dataset.
82
+
83
+ I followed this up with a coherence, decensorship & roleplay finetune over a 172 million token instruct dataset over two epochs.
84
+
85
+ I did a slerp merge of epoch 1 into epoch 2 at a light weight which resolved the name-spelling issues on quantized versions of Firefly v0.1.
86
+
87
+ Total training time was about 32hrs on 4x Nvidia A100 80GB.
88
+
89
+ <img src="https://raw.githubusercontent.com/axolotl-ai-cloud/axolotl/main/image/axolotl-badge-web.png" alt="Built with Axolotl" width="200" height="32"/>