File size: 1,421 Bytes
656a702 |
1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 16 17 18 19 20 21 22 23 24 25 26 27 28 29 30 31 32 33 34 35 36 37 38 39 40 |
---
license: other
language:
- en
base_model:
- mistralai/Mistral-Large-Instruct-2407
- NeverSleep/Lumimaid-v0.2-123B
---
# SmartMaid-123b
This **experimental model** is a hybrid creation combining aspects of [Mistral-Large-Instruct-2407](https://huggingface.co/mistralai/Mistral-Large-Instruct-2407) and [Lumimaid-v0.2-123B](https://huggingface.co/NeverSleep/Lumimaid-v0.2-123B) using LoRA (Low-Rank Adaptation) on the `mlp.down_proj` module.
## Model Details
- **Base Model**: Mistral-Large-Instruct-2407
- **Influence Model**: Lumimaid-v0.2-123B
- **Method**: LoRA extraction from Lumimaid and targetted application to Mistral-Large
- **LoRA Configuration**:
- Rank: 32
- Alpha: 64
- Target Module: `mlp.down_proj`
## Process
1. A LoRA was extracted from Lumimaid-v0.2-123B
2. This LoRA was then applied to a fresh instance of Mistral-Large-Instruct-2407 targetting only the mlp.down_proj modules.
3. The resulting model was merged to create this standalone version.
## Purpose
The aim of this model is to incorporate the enhanced prose qualities of Lumimaid-v0.2-123B while retaining the core intelligence and capabilities of Mistral-Large.
By applying the LoRA to the `mlp.down_proj` module, we sought to influence the model's language generation style without significantly altering its underlying knowledge and reasoning abilities.
## Prompt Template
```
<s>[INST] {input} [/INST] {output}</s>
```
|