Edit model card

Uploaded model

  • Computed sponsored by: Arroc ECS Denmark and Nvidia through Danish Data Science Community
  • Developed by: ThatsGroes
  • License: apache-2.0
  • Finetuned from model : AI-Sweden-Models/Llama-3-8B-instruct

This llama model was trained 2x faster with Unsloth and Huggingface's TRL library.

{'train_runtime': 215438.6486, 'train_samples_per_second': 3.141, 'train_steps_per_second': 0.393, 'train_loss': 1.035243785800245, 'epoch': 4.0}

[codecarbon INFO @ 07:52:43] Energy consumed for RAM : 11.292402 kWh. RAM Power : 188.78840446472168 W [codecarbon INFO @ 07:52:43] Energy consumed for all GPUs : 17.520012 kWh. Total GPU Power : 245.77458591976836 W [codecarbon INFO @ 07:52:43] Energy consumed for all CPUs : 2.543341 kWh. Total CPU Power : 42.5 W [codecarbon INFO @ 07:52:43] 31.355754 kWh of electricity used since the beginning.

We ended up using 65.56 GB GPU memory (82.84%), of which 49.83 GB (62.97%) was used for LoRa.

Downloads last month
35
Safetensors
Model size
8.03B params
Tensor type
BF16
·
Inference Examples
This model does not have enough activity to be deployed to Inference API (serverless) yet. Increase its social visibility and check back later, or deploy to Inference Endpoints (dedicated) instead.

Model tree for ThatsGroes/Llama-3.1-8B-Instruct-SkoleGPT-DaSlimOrca-4e

Finetuned
(15)
this model
Quantizations
1 model

Datasets used to train ThatsGroes/Llama-3.1-8B-Instruct-SkoleGPT-DaSlimOrca-4e