Update README.md
Browse files
README.md
CHANGED
@@ -54,7 +54,7 @@ base_model: /workspace/models/Mistral-Small-Instruct-2409
|
|
54 |
model_type: AutoModelForCausalLM
|
55 |
tokenizer_type: AutoTokenizer
|
56 |
|
57 |
-
hub_model_id: anthracite-
|
58 |
hub_strategy: "all_checkpoints"
|
59 |
push_dataset_to_hub:
|
60 |
hf_use_auth_token: true
|
@@ -72,17 +72,17 @@ load_in_4bit: false
|
|
72 |
strict: false
|
73 |
|
74 |
datasets:
|
75 |
-
- path: anthracite-
|
76 |
type: custommistralv2v3
|
77 |
-
- path: anthracite-
|
78 |
type: custommistralv2v3
|
79 |
-
- path: anthracite-
|
80 |
type: custommistralv2v3
|
81 |
- path: anthracite-org/nopm_claude_writing_fixed
|
82 |
type: custommistralv2v3
|
83 |
-
- path: anthracite-
|
84 |
type: custommistralv2v3
|
85 |
-
- path: anthracite-
|
86 |
type: custommistralv2v3
|
87 |
#chat_template: mistral_v2v3
|
88 |
shuffle_merged_datasets: true
|
@@ -150,12 +150,12 @@ We'd like to thank Recursal / Featherless for sponsoring the compute for this tr
|
|
150 |
We would also like to thank all members of Anthracite who made this finetune possible.
|
151 |
|
152 |
## Datasets
|
153 |
-
- [anthracite-
|
154 |
-
- [anthracite-
|
155 |
-
- [anthracite-
|
156 |
- [anthracite-org/nopm_claude_writing_fixed](https://huggingface.co/datasets/anthracite-org/nopm_claude_writing_fixed)
|
157 |
-
- [anthracite-
|
158 |
-
- [anthracite-
|
159 |
|
160 |
## Training
|
161 |
The training was done for 2 epochs. We used 8x[H100s](https://www.nvidia.com/en-us/data-center/h100/) GPUs graciously provided by [Recursal AI](https://recursal.ai/) / [Featherless AI](https://featherless.ai/) for the full-parameter fine-tuning of the model.
|
|
|
54 |
model_type: AutoModelForCausalLM
|
55 |
tokenizer_type: AutoTokenizer
|
56 |
|
57 |
+
hub_model_id: anthracite-org/magnum-v4-22b-r4
|
58 |
hub_strategy: "all_checkpoints"
|
59 |
push_dataset_to_hub:
|
60 |
hf_use_auth_token: true
|
|
|
72 |
strict: false
|
73 |
|
74 |
datasets:
|
75 |
+
- path: anthracite-org/c2_logs_32k_mistral-v3_v1.2_no_system
|
76 |
type: custommistralv2v3
|
77 |
+
- path: anthracite-org/kalo-opus-instruct-22k-no-refusal-no-system
|
78 |
type: custommistralv2v3
|
79 |
+
- path: anthracite-org/kalo-opus-instruct-3k-filtered-no-system
|
80 |
type: custommistralv2v3
|
81 |
- path: anthracite-org/nopm_claude_writing_fixed
|
82 |
type: custommistralv2v3
|
83 |
+
- path: anthracite-org/kalo_opus_misc_240827_no_system
|
84 |
type: custommistralv2v3
|
85 |
+
- path: anthracite-org/kalo_misc_part2_no_system
|
86 |
type: custommistralv2v3
|
87 |
#chat_template: mistral_v2v3
|
88 |
shuffle_merged_datasets: true
|
|
|
150 |
We would also like to thank all members of Anthracite who made this finetune possible.
|
151 |
|
152 |
## Datasets
|
153 |
+
- [anthracite-org/c2_logs_32k_mistral-v3_v1.2_no_system](https://huggingface.co/datasets/anthracite-org/c2_logs_32k_mistral-v3_v1.2_no_system)
|
154 |
+
- [anthracite-org/kalo-opus-instruct-22k-no-refusal-no-system](https://huggingface.co/datasets/anthracite-org/kalo-opus-instruct-22k-no-refusal-no-system)
|
155 |
+
- [anthracite-org/kalo-opus-instruct-3k-filtered-no-system](https://huggingface.co/datasets/anthracite-org/kalo-opus-instruct-3k-filtered-no-system)
|
156 |
- [anthracite-org/nopm_claude_writing_fixed](https://huggingface.co/datasets/anthracite-org/nopm_claude_writing_fixed)
|
157 |
+
- [anthracite-org/kalo_opus_misc_240827_no_system](https://huggingface.co/datasets/anthracite-org/kalo_opus_misc_240827_no_system)
|
158 |
+
- [anthracite-org/kalo_misc_part2_no_system](https://huggingface.co/datasets/anthracite-org/kalo_misc_part2_no_system)
|
159 |
|
160 |
## Training
|
161 |
The training was done for 2 epochs. We used 8x[H100s](https://www.nvidia.com/en-us/data-center/h100/) GPUs graciously provided by [Recursal AI](https://recursal.ai/) / [Featherless AI](https://featherless.ai/) for the full-parameter fine-tuning of the model.
|