File size: 802 Bytes
dbe1b18
4cad2f9
 
 
 
 
0600ac3
531f854
 
 
 
 
 
 
dbe1b18
0600ac3
531f854
0600ac3
 
 
251b16b
531f854
4cad2f9
531f854
 
 
 
4cad2f9
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
---
base_model:
- Qwen/Qwen2.5-72B-Instruct
- huihui-ai/Qwen2.5-72B-Instruct-abliterated
- Qwen/Qwen2.5-72B
- spow12/ChatWaifu_72B_v2.2
license: mit
datasets:
- arcee-ai/EvolKit-75K
- SkunkworksAI/reasoning-0.01
- berkeley-nest/Nectar
- Nexusflow/VirusTotalAgentic
- allenai/WildChat-1M-Full
- Magpie-Align/Magpie-LlamaCoT-250K
---

Experimental commander model V1. 

Named it Zelensky in order to troll Uncle Elon on twitter over how bad Grok-2 is.

Training process, low 1 epoch learning rate and evolutionary-merged with the 3 other models(listed on modelcard)

Process repeated multiple times on 8x AMD Mi300 192GB gpus while also running gpqa_diamond_zeroshot on LM_Eval harness.

Thank you Vultr https://www.vultr.com/register/ for sponsoring the compute.


Qwen License still applies by default.