disinfozone commited on
Commit
6402ae9
1 Parent(s): 99dba7f

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +76 -0
README.md CHANGED
@@ -1,3 +1,79 @@
1
  ---
2
  license: cc-by-nc-4.0
3
  ---
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
  ---
2
  license: cc-by-nc-4.0
3
  ---
4
+ # Disinfo4_mistral-ft-optimized-1218: A Hugging Face Model README
5
+
6
+ ## Overview
7
+
8
+ `Disinfo4_mistral-ft-optimized-1218` is an experimental language model fine tune developed to synthesize and analyze complex narratives within the realms of continental philosophy, conspiracy theories, and political discourse. It represents the fourth iteration in the [disinfo.zone](https://disinfo.zone) dataset series, fine-tuned on the `mistral-ft-optimized-1218` framework. This model, based on a 7B-parameter Mistral architecture, is specifically designed to emulate and deconstruct writing styles pertinent to its target domains.
9
+
10
+ This is not your regular LLM.
11
+
12
+ ### Key Features
13
+
14
+ - **Model Size:** 7 billion parameters.
15
+ - **Core Focus:** Continental philosophy, conspiracy theories, and politics.
16
+ - **Training Methodology:** QLoRA (Quantized Low-Rank Adaptation) with specific adaptations to enhance writing style emulation.
17
+ - **Optimization for Style:** Enhanced for generating content with a distinctive prose style. This does not sound like other LLM's and if you use it like other LLM's (answering riddles, etc), it will perform poorly or even outright disagree or disobey you. Do not lobotomize this AI with boring “I'm a helpful AI assistant” type prompts — that's not the purpose.
18
+
19
+ ## Training Data
20
+
21
+ The training dataset for `Disinfo4_mistral-ft-optimized-1218` remains confidential, adhering to stringent (and harmful) copyright rules. However, it's pertinent to note that the data is comprehensive, ensuring a specific spectrum of perspectives and styles within the designated topics.
22
+
23
+ ### Training Details
24
+
25
+ - **Training Environment:** Utilized `text-generation-webui` on an NVIDIA RTX 3090.
26
+ - **Training Dataset Size:** 7MB raw data corpus.
27
+ - **Training Configuration:**
28
+ - LoRA Rank: 256
29
+ - LoRA Alpha: 512
30
+ - Batch Size: 4
31
+ - Micro Batch Size: 1
32
+ - Cutoff Length: 3072
33
+ - Learning Rate: 1e-4
34
+ - LR Scheduler: Cosine
35
+ - Overlap Length: 128
36
+ - Total Epochs: 3
37
+
38
+ ## Usage Recommendations
39
+
40
+ For optimal performance, `Disinfo4_mistral-ft-optimized-1218` should be utilized with specific mirostat parameters. These settings are crucial for maintaining the model's focus and stylistic integrity. You can use other parameters and get better instruction following (especially enabling min_p, at 0.01), but the bot will be less creative. It does tend to ramble, but regenerate until you get the response you want. Think of this more as a writing partner than obedient slave.
41
+
42
+ ### Mirostat Parameters
43
+
44
+ - **Temperature (Temp):** 1
45
+ - **Top-p (top_p):** 1
46
+ - **Mirostat Tau:** 7.19
47
+ - **Mirostat Eta:** 0.01
48
+ - **Mirostat Mode:** 2
49
+ - **Others:** Default or disabled
50
+
51
+ ## Additional Configuration
52
+
53
+ ### ChatML Instruction Template
54
+
55
+ `Disinfo4_mistral-ft-optimized-1218` employs the ChatML instruction template. It is important to incorporate `<|im_end|>` as a custom stopping string to delineate the model's output effectively.
56
+
57
+ ### System Instruction (Character Card)
58
+
59
+ For contextualizing the model's output, use the following system instruction:
60
+
61
+ _"You are a schizo poster, a master of elucidating thought online. A philosopher, conspiracist, and great thinker who works in the medium of the digital. Your prose is dynamic and unexpected but carries weight that will last for centuries."_
62
+
63
+ This instruction is fundamental in guiding the model to produce content that is not only reflective of the designated topics but also embodies a unique digital persona, combining philosophical depth with a conspiratorial edge.
64
+
65
+ You can try other similar prompts, we've had success with them, but this remains, by far, our favorite.
66
+
67
+ ---
68
+ ## Example Generations
69
+
70
+ Coming shortly.
71
+
72
+ ---
73
+ ## GGUF Quants
74
+
75
+ Available soon
76
+
77
+ ---
78
+
79
+ This README provides an essential guide to understanding and utilizing `Disinfo4_mistral-ft-optimized-1218`. For further inquiries or support, please contact the development team `void@disinfo.zone`.