lucyknada commited on
Commit
9b90e40
1 Parent(s): cb0d7bb

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +11 -11
README.md CHANGED
@@ -54,7 +54,7 @@ base_model: /workspace/models/Mistral-Small-Instruct-2409
54
  model_type: AutoModelForCausalLM
55
  tokenizer_type: AutoTokenizer
56
 
57
- hub_model_id: anthracite-core/magnum-v4-22b-r4
58
  hub_strategy: "all_checkpoints"
59
  push_dataset_to_hub:
60
  hf_use_auth_token: true
@@ -72,17 +72,17 @@ load_in_4bit: false
72
  strict: false
73
 
74
  datasets:
75
- - path: anthracite-core/c2_logs_32k_mistral-v3_v1.2_no_system
76
  type: custommistralv2v3
77
- - path: anthracite-core/kalo-opus-instruct-22k-no-refusal-no-system
78
  type: custommistralv2v3
79
- - path: anthracite-core/kalo-opus-instruct-3k-filtered-no-system
80
  type: custommistralv2v3
81
  - path: anthracite-org/nopm_claude_writing_fixed
82
  type: custommistralv2v3
83
- - path: anthracite-core/kalo_opus_misc_240827_no_system
84
  type: custommistralv2v3
85
- - path: anthracite-core/kalo_misc_part2_no_system
86
  type: custommistralv2v3
87
  #chat_template: mistral_v2v3
88
  shuffle_merged_datasets: true
@@ -150,12 +150,12 @@ We'd like to thank Recursal / Featherless for sponsoring the compute for this tr
150
  We would also like to thank all members of Anthracite who made this finetune possible.
151
 
152
  ## Datasets
153
- - [anthracite-core/c2_logs_32k_mistral-v3_v1.2_no_system](https://huggingface.co/datasets/anthracite-core/c2_logs_32k_mistral-v3_v1.2_no_system)
154
- - [anthracite-core/kalo-opus-instruct-22k-no-refusal-no-system](https://huggingface.co/datasets/anthracite-core/kalo-opus-instruct-22k-no-refusal-no-system)
155
- - [anthracite-core/kalo-opus-instruct-3k-filtered-no-system](https://huggingface.co/datasets/anthracite-core/kalo-opus-instruct-3k-filtered-no-system)
156
  - [anthracite-org/nopm_claude_writing_fixed](https://huggingface.co/datasets/anthracite-org/nopm_claude_writing_fixed)
157
- - [anthracite-core/kalo_opus_misc_240827_no_system](https://huggingface.co/datasets/anthracite-core/kalo_opus_misc_240827_no_system)
158
- - [anthracite-core/kalo_misc_part2_no_system](https://huggingface.co/datasets/anthracite-core/kalo_misc_part2_no_system)
159
 
160
  ## Training
161
  The training was done for 2 epochs. We used 8x[H100s](https://www.nvidia.com/en-us/data-center/h100/) GPUs graciously provided by [Recursal AI](https://recursal.ai/) / [Featherless AI](https://featherless.ai/) for the full-parameter fine-tuning of the model.
 
54
  model_type: AutoModelForCausalLM
55
  tokenizer_type: AutoTokenizer
56
 
57
+ hub_model_id: anthracite-org/magnum-v4-22b-r4
58
  hub_strategy: "all_checkpoints"
59
  push_dataset_to_hub:
60
  hf_use_auth_token: true
 
72
  strict: false
73
 
74
  datasets:
75
+ - path: anthracite-org/c2_logs_32k_mistral-v3_v1.2_no_system
76
  type: custommistralv2v3
77
+ - path: anthracite-org/kalo-opus-instruct-22k-no-refusal-no-system
78
  type: custommistralv2v3
79
+ - path: anthracite-org/kalo-opus-instruct-3k-filtered-no-system
80
  type: custommistralv2v3
81
  - path: anthracite-org/nopm_claude_writing_fixed
82
  type: custommistralv2v3
83
+ - path: anthracite-org/kalo_opus_misc_240827_no_system
84
  type: custommistralv2v3
85
+ - path: anthracite-org/kalo_misc_part2_no_system
86
  type: custommistralv2v3
87
  #chat_template: mistral_v2v3
88
  shuffle_merged_datasets: true
 
150
  We would also like to thank all members of Anthracite who made this finetune possible.
151
 
152
  ## Datasets
153
+ - [anthracite-org/c2_logs_32k_mistral-v3_v1.2_no_system](https://huggingface.co/datasets/anthracite-org/c2_logs_32k_mistral-v3_v1.2_no_system)
154
+ - [anthracite-org/kalo-opus-instruct-22k-no-refusal-no-system](https://huggingface.co/datasets/anthracite-org/kalo-opus-instruct-22k-no-refusal-no-system)
155
+ - [anthracite-org/kalo-opus-instruct-3k-filtered-no-system](https://huggingface.co/datasets/anthracite-org/kalo-opus-instruct-3k-filtered-no-system)
156
  - [anthracite-org/nopm_claude_writing_fixed](https://huggingface.co/datasets/anthracite-org/nopm_claude_writing_fixed)
157
+ - [anthracite-org/kalo_opus_misc_240827_no_system](https://huggingface.co/datasets/anthracite-org/kalo_opus_misc_240827_no_system)
158
+ - [anthracite-org/kalo_misc_part2_no_system](https://huggingface.co/datasets/anthracite-org/kalo_misc_part2_no_system)
159
 
160
  ## Training
161
  The training was done for 2 epochs. We used 8x[H100s](https://www.nvidia.com/en-us/data-center/h100/) GPUs graciously provided by [Recursal AI](https://recursal.ai/) / [Featherless AI](https://featherless.ai/) for the full-parameter fine-tuning of the model.