Coder1.8-ORPO-TEST

Model Description

Test model for ORPO finetune method, trained on ~20k code examples for 1 epoch on 2 x A40 cards with 4-bit QLora (lora rank=lora alpha=16).

Disclaimer

This is a test model and may generate incorrect responses. Use at your own risk.

Train Details

  • Base: Qwen1.5-1.8B
  • Training Data: ~20k code examples
  • Epochs: 1
  • Method: ORPO
  • Hardware: 2 x A40
  • Quantization: 4-bit QLora
  • Lora Rank/Alpha: 16

Limitations

Limited training data and quantization may impact performance.

Join the Discussion

Have questions or feedback? Join our Discord server Here.

Open LLM Leaderboard Evaluation Results

Detailed results can be found here

Metric Value
Avg. 45.76
AI2 Reasoning Challenge (25-Shot) 38.82
HellaSwag (10-Shot) 60.48
MMLU (5-Shot) 46.70
TruthfulQA (0-shot) 41.38
Winogrande (5-shot) 59.75
GSM8k (5-shot) 27.45
Downloads last month
16
Safetensors
Model size
1.84B params
Tensor type
BF16
·
Inference Examples
This model does not have enough activity to be deployed to Inference API (serverless) yet. Increase its social visibility and check back later, or deploy to Inference Endpoints (dedicated) instead.

Dataset used to train raincandy-u/Coder1.8-ORPO-TEST

Evaluation results