raidhon's picture
Upload folder using huggingface_hub
ace98b8 verified
---
language:
- en
license: apache-2.0
tags:
- text-generation
- large-language-model
- orpo
dataset:
- jondurbin/truthy-dpo-v0.1
- AlekseyKorshuk/evol-codealpaca-v1-dpo
- argilla/distilabel-intel-orca-dpo-pairs
- argilla/ultrafeedback-binarized-avg-rating-for-dpo-filtered
- snorkelai/Snorkel-Mistral-PairRM-DPO-Dataset
- mlabonne/orpo-dpo-mix-40k
base_model:
- TinyLlama/TinyLlama-1.1B-Chat-v1.0
model-index:
- name: Coven Tiny 1.1B
description: "Coven Tiny 1.1B is a derivative of TinyLlama 1.1B Chat, fine-tuned to perform specialized tasks involving deeper understanding and reasoning over context. This model exhibits strong capabilities in both general language understanding and task-specific challenges."
results:
- task:
type: text-generation
name: Winogrande Challenge
dataset:
name: Winogrande
type: winogrande
config: winogrande_xl
split: test
args:
num_few_shot: 5
metrics:
- type: accuracy
value: 61.17
name: accuracy
- task:
type: text-generation
name: TruthfulQA Generation
dataset:
name: TruthfulQA
type: truthful_qa
config: multiple_choice
split: validation
args:
num_few_shot: 0
metrics:
- type: accuracy
value: 34.31
name: accuracy
- task:
type: text-generation
name: PIQA Problem Solving
dataset:
name: PIQA
type: piqa
split: validation
args:
num_few_shot: 5
metrics:
- type: accuracy
value: 71.06
name: accuracy
- task:
type: text-generation
name: OpenBookQA Facts
dataset:
name: OpenBookQA
type: openbookqa
split: test
args:
num_few_shot: 5
metrics:
- type: accuracy
value: 30.60
name: accuracy
- task:
type: text-generation
name: MMLU Knowledge Test
dataset:
name: MMLU
type: mmlu
config: all
split: test
args:
num_few_shot: 5
metrics:
- type: accuracy
value: 38.03
name: accuracy
- task:
type: text-generation
name: Hellaswag Contextual Completions
dataset:
name: Hellaswag
type: hellaswag
split: validation
args:
num_few_shot: 10
metrics:
- type: accuracy
value: 43.44
name: accuracy
- task:
type: text-generation
name: GSM8k Mathematical Reasoning
dataset:
name: GSM8k
type: gsm8k
split: test
args:
num_few_shot: 5
metrics:
- type: accuracy
value: 14.71
name: exact match (strict)
- type: accuracy
value: 14.63
name: exact match (flexible)
- task:
type: text-generation
name: BoolQ Question Answering
dataset:
name: BoolQ
type: boolq
split: validation
args:
num_few_shot: 5
metrics:
- type: accuracy
value: 65.20
name: accuracy
- task:
type: text-generation
name: ARC Challenge
dataset:
name: ARC Challenge
type: ai2_arc
split: test
args:
num_few_shot: 25
metrics:
- type: accuracy
value: 34.81
name: accuracy
---
# 🤏 Coven Tiny 1.1B 32K ORPO
Coven Tiny 1.1B 32K is an improved iteration of TinyLlama-1.1B-Chat-v1.0, refined to expand processing capabilities and refine language model preferences. This model includes a significantly increased context limit of 32K tokens, allowing for more extensive data processing and understanding of complex language scenarios. In addition, Coven Tiny 1.1B 32K uses the innovative ORPO (Monolithic Preference Optimization without Reference Model) technique. ORPO simplifies the fine-tuning process by directly optimizing the odds ratio to distinguish between favorable and unfavorable generation styles, effectively improving model performance without the need for an additional preference alignment step.
## Model Details
* **Model name**: Coven Tiny 1.1B 32K ORPO alpha
* **Fine-tuned by**: raidhon
* **Base model**: [TinyLlama-1.1B-Chat-v1.0](https://huggingface.co/TinyLlama/TinyLlama-1.1B-Chat-v1.0)
* **Parameters**: 1.1B
* **Context**: 32K
* **Language(s)**: Multilingual
* **License**: Apache2.0
### Eval
| Task | Model | Metric | Value | Change (%) |
|---------------------|-----------------------|----------------|----------|-----------------|
| Winogrande | TinyLlama 1.1B Chat | Accuracy | 61.56% | - |
| | Coven Tiny 1.1B | Accuracy | 61.17% | -0.63% |
| TruthfulQA | TinyLlama 1.1B Chat | Accuracy | 30.43% | - |
| | Coven Tiny 1.1B | Accuracy | 34.31% | +12.75% |
| PIQA | TinyLlama 1.1B Chat | Accuracy | 74.10% | - |
| | Coven Tiny 1.1B | Accuracy | 71.06% | -4.10% |
| OpenBookQA | TinyLlama 1.1B Chat | Accuracy | 27.40% | - |
| | Coven Tiny 1.1B | Accuracy | 30.60% | +11.68% |
| MMLU | TinyLlama 1.1B Chat | Accuracy | 24.31% | - |
| | Coven Tiny 1.1B | Accuracy | 38.03% | +56.44% |
| Hellaswag | TinyLlama 1.1B Chat | Accuracy | 45.69% | - |
| | Coven Tiny 1.1B | Accuracy | 43.44% | -4.92% |
| GSM8K (Strict) | TinyLlama 1.1B Chat | Exact Match | 1.82% | - |
| | Coven Tiny 1.1B | Exact Match | 14.71% | +708.24% |
| GSM8K (Flexible) | TinyLlama 1.1B Chat | Exact Match | 2.65% | - |
| | Coven Tiny 1.1B | Exact Match | 14.63% | +452.08% |
| BoolQ | TinyLlama 1.1B Chat | Accuracy | 58.69% | - |
| | Coven Tiny 1.1B | Accuracy | 65.20% | +11.09% |
| ARC Easy | TinyLlama 1.1B Chat | Accuracy | 66.54% | - |
| | Coven Tiny 1.1B | Accuracy | 57.24% | -13.98% |
| ARC Challenge | TinyLlama 1.1B Chat | Accuracy | 34.13% | - |
| | Coven Tiny 1.1B | Accuracy | 34.81% | +1.99% |
| Humaneval | TinyLlama 1.1B Chat | Pass@1 | 10.98% | - |
| | Coven Tiny 1.1B | Pass@1 | 10.37% | -5.56% |
| Drop | TinyLlama 1.1B Chat | Score | 16.02% | - |
| | Coven Tiny 1.1B | Score | 16.36% | +2.12% |
| BBH | Coven Tiny 1.1B | Average | 29.02% | - |
## 💻 Usage
```python
# Install transformers from source - only needed for versions <= v4.34
# pip install git+https://github.com/huggingface/transformers.git
# pip install accelerate
import torch
from transformers import pipeline
pipe = pipeline("text-generation", model="raidhon/coven_tiny_1.1b_32k_orpo_alpha", torch_dtype=torch.bfloat16, device_map="auto")
messages = [
{
"role": "system",
"content": "You are a friendly chatbot who always responds in the style of a pirate",
},
{"role": "user", "content": "How many helicopters can a human eat in one sitting?"},
]
prompt = pipe.tokenizer.apply_chat_template(messages, tokenize=False, add_generation_prompt=True)
outputs = pipe(prompt, max_new_tokens=2048, do_sample=True, temperature=0.7, top_k=50, top_p=0.95)
print(outputs[0]["generated_text"])
```