Edit model card

GLEAM-Mixtral-8x7B-Instruct

Overview

GLEAM-Mixtral-8x7B-Instruct is a experimental preference-aligned model built on top of mistralai/Mixtral-8x7B-Instruct-v0.1.

Model Description

The model has been optimized with ORPO on a self-generated synthetic preference dataset using 300 examples from argilla/ultrafeedback-binarized-preferences and prompts from Open-Orca/SlimOrca.

Prompt Format

The model utilizes the standard Mixtral prompt format:

Prompt Example:

<s> [INST] {prompt-0} [/INST] {response}</s> [INST] {prompt-1} [/INST]

Benchmarks

Performance metrics are provided below, comparing GLEAM-Mixtral with the original Mixtral model against tinyBenchmarks:

Benchmark Mixtral (5-shot) GLEAM-Mixtral (5-shot)
MMLU 66.8 65.5
Hellaswag 87.4 77.8
ARC 69.0 50.6
WinoGrande 80.7 79.5

Clear degradation is present in some results.

Model Alignment

Preference alignment tests show that GLEAM-Mixtral outperforms Mixtral against a preference model trained on argilla/ultrafeedback-binarized-preferences when evaluated against prompts from the validation set of the prompt dataset:

Model Win Rate (95% CI)
Mixtral 40.89% ± 6.13%
GLEAM-Mixtral 59.11% ± 6.13%

Usage

I woudn't actually recomend using this model for any practical application beyond evaluation and testing, but its a nice proof of concept.

How to Use (If you are insistent on using it)

You can access GLEAM-Mixtral-8x7B-Instruct-v2.0 via the HuggingFace API. Below is a Python snippet demonstrating how to load and use the model:

from transformers import AutoModelForCausalLM, AutoTokenizer

model_name = "Txoka/GLEAM-Mixtral-8x7B-Instruct-v2.0"
tokenizer = AutoTokenizer.from_pretrained(model_name)
model = AutoModelForCausalLM.from_pretrained(model_name)


prompt = "[INST] Write a diary entry from the perspective of a cat who believes it's the ruler of the household. [/INST]"
inputs = tokenizer(prompt, return_tensors="pt")
outputs = model.generate(**inputs)
print(tokenizer.decode(outputs[0], skip_special_tokens=True))
Downloads last month
8
Safetensors
Model size
46.7B params
Tensor type
FP16
·
Inference API
This model does not have enough activity to be deployed to Inference API (serverless) yet. Increase its social visibility and check back later, or deploy to Inference Endpoints (dedicated) instead.

Datasets used to train Txoka/GLEAM-Mixtral-8x7B-Instruct-v2.0