Svak commited on
Commit
dee3d7a
1 Parent(s): abf2a24

Create README.md

Browse files
Files changed (1) hide show
  1. README.md +124 -0
README.md ADDED
@@ -0,0 +1,124 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ ---
2
+ library_name: transformers
3
+ license: llama3
4
+ base_model: meta-llama/Llama-3.3-70B-Instruct
5
+ tags:
6
+ - generated_from_trainer
7
+ model-index:
8
+ - name: L3.3-70B-Euryale-v2.3
9
+ results: []
10
+ ---
11
+
12
+ ![eury](https://huggingface.co/Sao10K/L3.3-70B-Euryale-v2.3/resolve/main/Eury.png)
13
+
14
+
15
+ # This quant was made for and by [Infermatic.ai](https://infermatic.ai/)
16
+
17
+ [Sao10K/L3.3-70B-Euryale-v2.3](https://huggingface.co/Sao10K/L3.3-70B-Euryale-v2.3)
18
+
19
+ Copy of the original card
20
+
21
+ # L3.3-70B-Euryale-v2.3
22
+
23
+ A direct replacement / successor to Euryale v2.2, not Hanami-x1, though it is slightly better than them in my opinion.
24
+
25
+ This is entirely trained on top of Llama 3.3 Instruct, not Lora-extracted which is all the rage.
26
+
27
+ Recommended Model Settings | *Look, I just use these, they work fine enough. I don't even know how DRY or other meme samplers work. Your system prompt matters more anyway.*
28
+ ```
29
+ Prompt Format: Llama-3-Instruct
30
+ Temperature: 1.1
31
+ min_p: 0.1
32
+ ```
33
+
34
+
35
+ Future-ish plans:
36
+ <br>\- Further refine the Datasets used for quality, more secondary chats, more creative-related domains.
37
+ <br>\- Work on my other incomplete projects. About half a dozen on the backburner for a while now.
38
+
39
+ Special thanks to my wallet for funding this, my juniors who share a single braincell between them, and my current national service.
40
+
41
+ Have a good day, don't shit yourselves friends. I had a nasty call today.
42
+
43
+ Also sorry for the inactivity. Life was in the way. It still is, just less so, for now. Burnout is a thing, huh?
44
+
45
+ https://sao10k.carrd.co/ for contact.
46
+
47
+ ---
48
+
49
+ [<img src="https://raw.githubusercontent.com/axolotl-ai-cloud/axolotl/main/image/axolotl-badge-web.png" alt="Built with Axolotl" width="200" height="32"/>](https://github.com/axolotl-ai-cloud/axolotl)
50
+ <details><summary>See axolotl config</summary>
51
+
52
+ axolotl version: `0.5.2`
53
+ ```yaml
54
+ base_model: meta-llama/Llama-3.3-70B-Instruct
55
+ model_type: AutoModelForCausalLM
56
+ tokenizer_type: AutoTokenizer
57
+
58
+ load_in_8bit: false
59
+ load_in_4bit: false
60
+ strict: false
61
+ sequence_len: 16384
62
+ bf16: auto
63
+ fp16:
64
+ tf32: false
65
+ flash_attention: true
66
+
67
+ adapter: lora
68
+ lora_model_dir:
69
+ lora_r: 128
70
+ lora_alpha: 16
71
+ lora_dropout: 0.1
72
+ lora_target_linear: true
73
+ lora_fan_in_fan_out:
74
+ peft_use_rslora: true
75
+
76
+ # Data
77
+ dataset_prepared_path: last_run_prepared
78
+ datasets:
79
+ - path: datasets/amoral-full-sys-prompt.json # Unalignment Data - Cleaned Up from Original, Split to its own file
80
+ type: customllama3
81
+ - path: datasets/mimi-superfix-RP-filtered-fixed.json # RP / Creative-Instruct Data
82
+ type: customllama3
83
+ - path: datasets/hespera-smartshuffle.json # Hesperus-v2-Instruct Data
84
+ type: customllama3
85
+ warmup_steps: 15
86
+
87
+ plugins:
88
+ - axolotl.integrations.liger.LigerPlugin
89
+ liger_rope: true
90
+ liger_rms_norm: true
91
+ liger_layer_norm: true
92
+ liger_glu_activation: true
93
+ liger_fused_linear_cross_entropy: true
94
+
95
+ # Iterations
96
+ num_epochs: 1
97
+
98
+ # Batching
99
+ gradient_accumulation_steps: 4
100
+ micro_batch_size: 1
101
+ gradient_checkpointing: "unsloth"
102
+
103
+ # Optimizer
104
+ optimizer: paged_ademamix_8bit
105
+ lr_scheduler: cosine
106
+ learning_rate: 0.000004
107
+ weight_decay: 0.1
108
+ max_grad_norm: 25.0
109
+
110
+ # Iterations
111
+ num_epochs: 1
112
+
113
+ # Misc
114
+ deepspeed: ./deepspeed_configs/zero3_bf16.json
115
+ ```
116
+
117
+ </details><br>
118
+
119
+ ---
120
+
121
+ ```
122
+ Art by てぃあ
123
+ https://www.pixiv.net/en/users/724263
124
+ ```