AuriAetherwiing commited on
Commit
30fe13b
1 Parent(s): cc0ce82

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +67 -66
README.md CHANGED
@@ -1,7 +1,20 @@
1
  ---
2
  library_name: transformers
3
  license: other
 
 
4
  base_model: Qwen/Qwen2.5-72B
 
 
 
 
 
 
 
 
 
 
 
5
  tags:
6
  - generated_from_trainer
7
  model-index:
@@ -9,8 +22,59 @@ model-index:
9
  results: []
10
  ---
11
 
12
- <!-- This model card has been generated automatically according to the information the Trainer had access to. You
13
- should probably proofread and complete it, then remove this comment. -->
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
14
 
15
  [<img src="https://raw.githubusercontent.com/axolotl-ai-cloud/axolotl/main/image/axolotl-badge-web.png" alt="Built with Axolotl" width="200" height="32"/>](https://github.com/axolotl-ai-cloud/axolotl)
16
  <details><summary>See axolotl config</summary>
@@ -425,67 +489,4 @@ weight_decay: 0.12
425
  # fsdp_mixed_precision: BF16 # Added
426
  ```
427
 
428
- </details><br>
429
-
430
- # EVA-Qwen2.5-72B-SFFT-v0.2
431
-
432
- This model is a fine-tuned version of [Qwen/Qwen2.5-72B](https://huggingface.co/Qwen/Qwen2.5-72B) on the None dataset.
433
- It achieves the following results on the evaluation set:
434
- - Loss: 1.0002
435
-
436
- ## Model description
437
-
438
- More information needed
439
-
440
- ## Intended uses & limitations
441
-
442
- More information needed
443
-
444
- ## Training and evaluation data
445
-
446
- More information needed
447
-
448
- ## Training procedure
449
-
450
- ### Training hyperparameters
451
-
452
- The following hyperparameters were used during training:
453
- - learning_rate: 3e-05
454
- - train_batch_size: 1
455
- - eval_batch_size: 1
456
- - seed: 42
457
- - distributed_type: multi-GPU
458
- - num_devices: 8
459
- - gradient_accumulation_steps: 8
460
- - total_train_batch_size: 64
461
- - total_eval_batch_size: 8
462
- - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
463
- - lr_scheduler_type: cosine
464
- - lr_scheduler_warmup_steps: 20
465
- - num_epochs: 3
466
-
467
- ### Training results
468
-
469
- | Training Loss | Epoch | Step | Validation Loss |
470
- |:-------------:|:------:|:----:|:---------------:|
471
- | 1.3381 | 0.0078 | 1 | 1.2881 |
472
- | 1.0803 | 0.2493 | 32 | 0.9745 |
473
- | 0.8733 | 0.4985 | 64 | 0.9413 |
474
- | 0.9273 | 0.7478 | 96 | 0.9224 |
475
- | 0.8413 | 0.9971 | 128 | 0.9133 |
476
- | 0.5342 | 1.2493 | 160 | 0.9449 |
477
- | 0.5184 | 1.4985 | 192 | 0.9414 |
478
- | 0.5177 | 1.7478 | 224 | 0.9300 |
479
- | 0.5494 | 1.9971 | 256 | 0.9273 |
480
- | 0.303 | 2.2449 | 288 | 0.9968 |
481
- | 0.314 | 2.4946 | 320 | 1.0009 |
482
- | 0.3022 | 2.7444 | 352 | 1.0004 |
483
- | 0.2901 | 2.9941 | 384 | 1.0002 |
484
-
485
-
486
- ### Framework versions
487
-
488
- - Transformers 4.45.1
489
- - Pytorch 2.3.1+cu121
490
- - Datasets 2.21.0
491
- - Tokenizers 0.20.3
 
1
  ---
2
  library_name: transformers
3
  license: other
4
+ license_name: qwen
5
+ license_link: https://huggingface.co/Qwen/Qwen2.5-72B-Instruct/blob/main/LICENSE
6
  base_model: Qwen/Qwen2.5-72B
7
+ datasets:
8
+ - anthracite-org/kalo-opus-instruct-22k-no-refusal
9
+ - Nopm/Opus_WritingStruct
10
+ - Gryphe/Sonnet3.5-SlimOrcaDedupCleaned
11
+ - Gryphe/Sonnet3.5-Charcard-Roleplay
12
+ - Gryphe/ChatGPT-4o-Writing-Prompts
13
+ - Epiculous/Synthstruct-Gens-v1.1-Filtered-n-Cleaned
14
+ - Epiculous/SynthRP-Gens-v1.1-Filtered-n-Cleaned
15
+ - nothingiisreal/Reddit-Dirty-And-WritingPrompts
16
+ - allura-org/Celeste-1.x-data-mixture
17
+ - cognitivecomputations/dolphin-2.9.3
18
  tags:
19
  - generated_from_trainer
20
  model-index:
 
22
  results: []
23
  ---
24
 
25
+
26
+
27
+ # EVA Qwen2.5-72B v0.2
28
+
29
+ <p>
30
+ A RP/storywriting specialist model, full-parameter finetune of Qwen2.5-72B on mixture of synthetic and natural data.<br>
31
+ It uses Celeste 70B 0.1 data mixture, greatly expanding it to improve versatility, creativity and "flavor" of the resulting model.<br>
32
+ </p>
33
+
34
+ <p>Dedicated to Nev.</p>
35
+
36
+ <p><b>Version notes for 0.2</b>: Optimized training hyperparameters and increased sequence length. Better instruction following deeper into context and less repetition.</p>
37
+
38
+ <p>
39
+ <p>Prompt format is ChatML.</p><br>
40
+ <h3>Recommended sampler values:</h3>
41
+ <ul>
42
+ <li>Temperature: 0.8</li>
43
+ <li>Min-P: 0.05</li>
44
+ <li>Top-A: 0.3</li>
45
+ <li>Repetition Penalty: 1.03</li>
46
+ </ul>
47
+
48
+ <h3>Recommended SillyTavern preset (via CalamitousFelicitousness):</h3>
49
+ <ul><li>Master import</li></ul>
50
+
51
+ </p>
52
+
53
+ <p>
54
+ <br>
55
+ <h3>
56
+ Training data:
57
+ </h3>
58
+ <ul>
59
+ <li>Celeste 70B 0.1 data mixture minus Opus Instruct subset. See that model's <a href=https://huggingface.co/nothingiisreal/L3.1-70B-Celeste-V0.1-BF16>card</a> for details.</li>
60
+ <li>Kalomaze's Opus_Instruct_25k dataset, filtered for refusals.</li>
61
+ <li>A subset (1k rows) of ChatGPT-4o-WritingPrompts by Gryphe</li>
62
+ <li>A subset (2k rows) of Sonnet3.5-Charcards-Roleplay by Gryphe</li>
63
+ <li>Synthstruct and SynthRP datasets by Epiculous</li>
64
+ <li>A subset from Dolphin-2.9.3, including filtered version of not_samantha and a small subset of systemchat.</li>
65
+ </ul>
66
+ <h3>
67
+ Training time and hardware:
68
+ </h3>
69
+ <ul><li>17 hours on 8xH100 SXM</a></li></ul><br>
70
+ </p>
71
+ <p>Model was created by Kearm, Auri and Cahvay.</p>
72
+ <h4>Special thanks:</h4><ul>
73
+ <li>to Cahvay for his work on investigating and reprocessing the corrupted dataset, removing the single biggest source of data poisoning.</li>
74
+ <li>to Gryphe, Lemmy, Kalomaze, Nopm, Epiculous and CognitiveComputations for the data</li>
75
+ <li>and to Allura-org for support, feedback, beta-testing and doing quality control of EVA models.</li></ul>
76
+
77
+
78
 
79
  [<img src="https://raw.githubusercontent.com/axolotl-ai-cloud/axolotl/main/image/axolotl-badge-web.png" alt="Built with Axolotl" width="200" height="32"/>](https://github.com/axolotl-ai-cloud/axolotl)
80
  <details><summary>See axolotl config</summary>
 
489
  # fsdp_mixed_precision: BF16 # Added
490
  ```
491
 
492
+ </details><br>