Qwen2.5-7B-Gutenberg-KTO

This model is fine tuned over gutenberg datasets using kto strategy. It's my first time to use kto strategy, and I'm not sure how the model actually performs.

Compared to those large companies which remove accessories such as charger and cables from packages, I have achieved real environment protection by truly reducing energy consumption, rather than shifting costs to consumers.

Checkout GGUF here: Orion-zhen/Qwen2.5-7B-Gutenberg-KTO-Q6_K-GGUF

Details

Platform

I randomly grabbed some rubbish from a second-hand market and built a PC

I carefully selected various dedicated hardwares and constructed an incomparable home server, which I entitled the Great Server:

  • CPU: Intel Core i3-4160
  • Memory: 8G DDR3, single channel
  • GPU: Tesla P4, TDP 75W, boasting its Eco friendly energy consumption
  • Disk: 1TB M.2 NVME, PCIe 4.0

Training

To practice the eco-friendly training, I utilized various methods, including adam-mini, qlora and unsloth, to minimize VRAM and energy usage, as well as accelerating training speed.

Train log

training_loss

training_eval_loss

Downloads last month
32
Safetensors
Model size
7.62B params
Tensor type
BF16
·
Inference Examples
Unable to determine this model's library. Check the docs .

Model tree for Orion-zhen/Qwen2.5-7B-Gutenberg-KTO

Base model

Qwen/Qwen2.5-7B
Finetuned
(2)
this model
Quantizations
2 models

Dataset used to train Orion-zhen/Qwen2.5-7B-Gutenberg-KTO

Collection including Orion-zhen/Qwen2.5-7B-Gutenberg-KTO