lucyknada commited on
Commit
cb0d7bb
1 Parent(s): 23edc8f

Create README.md

Browse files
Files changed (1) hide show
  1. README.md +166 -0
README.md ADDED
@@ -0,0 +1,166 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ ---
2
+ license: other
3
+ license_name: mrl
4
+ language:
5
+ - en
6
+ tags:
7
+ - chat
8
+ pipeline_tag: text-generation
9
+ library_name: transformers
10
+ ---
11
+
12
+ ![image/png](https://cdn-uploads.huggingface.co/production/uploads/658a46cbfb9c2bdfae75b3a6/WvQykcYiK13x7sMI93T6e.png)
13
+ ## This repo contains GGUF quants of the model. If you need the original weights, please find them [here](https://huggingface.co/anthracite-org/magnum-v4-22b).
14
+
15
+ This is a series of models designed to replicate the prose quality of the Claude 3 models, specifically Sonnet and Opus.
16
+
17
+
18
+
19
+ This model is fine-tuned on top of [Mistral-Small-Instruct-2409](https://huggingface.co/mistralai/Mistral-Small-Instruct-2409).
20
+
21
+ ## Prompting
22
+ A typical input would look like this:
23
+
24
+ ```py
25
+ <s>[INST] SYSTEM MESSAGE
26
+ USER MESSAGE[/INST] ASSISTANT MESSAGE</s>[INST] USER MESSAGE[/INST]
27
+ ```
28
+
29
+ ## SillyTavern templates
30
+
31
+ Below are Instruct and Context templates for use within SillyTavern.
32
+
33
+ <details><summary>context template</summary>
34
+
35
+ ```yaml
36
+ default SillyTavern template works fine
37
+ ```
38
+
39
+ </details><br>
40
+ <details><summary>instruct template</summary>
41
+
42
+ ```yaml
43
+ default SillyTavern template works fine
44
+ ```
45
+
46
+ </details><br>
47
+
48
+ ## Axolotl config
49
+
50
+ <details><summary>See axolotl config</summary>
51
+
52
+ ```yaml
53
+ base_model: /workspace/models/Mistral-Small-Instruct-2409
54
+ model_type: AutoModelForCausalLM
55
+ tokenizer_type: AutoTokenizer
56
+
57
+ hub_model_id: anthracite-core/magnum-v4-22b-r4
58
+ hub_strategy: "all_checkpoints"
59
+ push_dataset_to_hub:
60
+ hf_use_auth_token: true
61
+
62
+ plugins:
63
+ - axolotl.integrations.liger.LigerPlugin
64
+ liger_rope: true
65
+ liger_rms_norm: true
66
+ liger_swiglu: true
67
+ #liger_cross_entropy: true
68
+ liger_fused_linear_cross_entropy: true
69
+
70
+ load_in_8bit: false
71
+ load_in_4bit: false
72
+ strict: false
73
+
74
+ datasets:
75
+ - path: anthracite-core/c2_logs_32k_mistral-v3_v1.2_no_system
76
+ type: custommistralv2v3
77
+ - path: anthracite-core/kalo-opus-instruct-22k-no-refusal-no-system
78
+ type: custommistralv2v3
79
+ - path: anthracite-core/kalo-opus-instruct-3k-filtered-no-system
80
+ type: custommistralv2v3
81
+ - path: anthracite-org/nopm_claude_writing_fixed
82
+ type: custommistralv2v3
83
+ - path: anthracite-core/kalo_opus_misc_240827_no_system
84
+ type: custommistralv2v3
85
+ - path: anthracite-core/kalo_misc_part2_no_system
86
+ type: custommistralv2v3
87
+ #chat_template: mistral_v2v3
88
+ shuffle_merged_datasets: true
89
+ #default_system_message: "You are an assistant that responds to the user."
90
+ dataset_prepared_path: /workspace/data/magnum-22b-data
91
+ val_set_size: 0.0
92
+ output_dir: /workspace/data/22b-r4-fft-out
93
+
94
+ sequence_len: 32768
95
+ sample_packing: true
96
+ pad_to_sequence_len: true
97
+
98
+ adapter:
99
+ lora_model_dir:
100
+ lora_r:
101
+ lora_alpha:
102
+ lora_dropout:
103
+ lora_target_linear:
104
+ lora_fan_in_fan_out:
105
+
106
+ wandb_project: 22b-magnum-fft
107
+ wandb_entity:
108
+ wandb_watch:
109
+ wandb_name: v4-r4-attempt-01
110
+ wandb_log_model:
111
+
112
+ gradient_accumulation_steps: 2
113
+ micro_batch_size: 1
114
+ num_epochs: 2
115
+ optimizer: adamw_bnb_8bit
116
+ lr_scheduler: cosine
117
+ learning_rate: 0.000004
118
+
119
+ train_on_inputs: false
120
+ group_by_length: false
121
+ bf16: auto
122
+ fp16:
123
+ tf32: false
124
+
125
+ gradient_checkpointing: true
126
+ early_stopping_patience:
127
+ resume_from_checkpoint:
128
+ local_rank:
129
+ logging_steps: 1
130
+ xformers_attention:
131
+ flash_attention: true
132
+
133
+ warmup_steps: 40
134
+ evals_per_epoch:
135
+ eval_table_size:
136
+ eval_max_new_tokens:
137
+ saves_per_epoch: 2
138
+ debug:
139
+ deepspeed: deepspeed_configs/zero3_bf16.json
140
+ weight_decay: 0.1
141
+ fsdp:
142
+ fsdp_config:
143
+ special_tokens:
144
+ ```
145
+ </details><br>
146
+
147
+ ## Credits
148
+ We'd like to thank Recursal / Featherless for sponsoring the compute for this train, Featherless has been hosting our Magnum models since the first 72 B and has given thousands of people access to our models and helped us grow.
149
+
150
+ We would also like to thank all members of Anthracite who made this finetune possible.
151
+
152
+ ## Datasets
153
+ - [anthracite-core/c2_logs_32k_mistral-v3_v1.2_no_system](https://huggingface.co/datasets/anthracite-core/c2_logs_32k_mistral-v3_v1.2_no_system)
154
+ - [anthracite-core/kalo-opus-instruct-22k-no-refusal-no-system](https://huggingface.co/datasets/anthracite-core/kalo-opus-instruct-22k-no-refusal-no-system)
155
+ - [anthracite-core/kalo-opus-instruct-3k-filtered-no-system](https://huggingface.co/datasets/anthracite-core/kalo-opus-instruct-3k-filtered-no-system)
156
+ - [anthracite-org/nopm_claude_writing_fixed](https://huggingface.co/datasets/anthracite-org/nopm_claude_writing_fixed)
157
+ - [anthracite-core/kalo_opus_misc_240827_no_system](https://huggingface.co/datasets/anthracite-core/kalo_opus_misc_240827_no_system)
158
+ - [anthracite-core/kalo_misc_part2_no_system](https://huggingface.co/datasets/anthracite-core/kalo_misc_part2_no_system)
159
+
160
+ ## Training
161
+ The training was done for 2 epochs. We used 8x[H100s](https://www.nvidia.com/en-us/data-center/h100/) GPUs graciously provided by [Recursal AI](https://recursal.ai/) / [Featherless AI](https://featherless.ai/) for the full-parameter fine-tuning of the model.
162
+
163
+ [<img src="https://raw.githubusercontent.com/OpenAccess-AI-Collective/axolotl/main/image/axolotl-badge-web.png" alt="Built with Axolotl" width="200" height="32"/>](https://github.com/OpenAccess-AI-Collective/axolotl)
164
+
165
+ ## Safety
166
+ ...