--- language: - en - fr - de - es - it - pt - zh - ja - ru - ko license: other license_name: mrl license_link: https://mistral.ai/licenses/MRL-0.1.md --- # Writer-Large-2411-v2.1 EXL2-Quant of [gghfez/Writer-Large-2411-v2.1](https://huggingface.co/gghfez/Writer-Large-2411-v2.1-exl2-4.5bpw) Creative-Writing Control-Vectors available here: [gghfez/Writer-Large-2411-v2.1-control-vectors](https://huggingface.co/gghfez/Writer-Large-2411-v2.1-control-vectors) ## Overview This model is built on Mistral-Large-Instruct-2411 and optimized for creative writing purposes. The base model excels at following instructions and handling details in long context when using the [new prompt template](https://huggingface.co/gghfez/Mistral-Large-Instruct-2411/blob/main/tokenizer_config.json#L6177). ### Key Improvements - Reduced positivity bias - Reduced AI tropes and repetitive language patterns in story generation - Enhanced performance with longer context stories (multiple chapters) and roleplay sessions - Improved steering capabilities for roleplay via [OOC] instructions - Better handling of "group chat" scenarios ## Usage ### Prompt Template **The model requires a system prompt in the Mistral-V7 format.** If you omit [`SYSTEM_PROMPT] [/SYSTEM_PROMPT]`, the model: - May not follow instructions properly at short contexts - Can become repetitive at longer contexts Example: ```python [SYSTEM_PROMPT]You are an award winning writer. Assist the user.[/SYSTEM_PROMPT][INST] Write the opening chapter of ... [/INST] ``` ### SillyTavern Integration Story String: ```python [SYSTEM_PROMPT] {{#if system}}{{system}}[/SYSTEM_PROMPT] [INST] {{/if}}{{#if wiBefore}}{{wiBefore}} {{/if}}{{#if description}}{{description}} {{/if}}{{#if personality}}{{personality}} {{/if}}{{#if scenario}}{{scenario}} {{/if}}{{#if wiAfter}}{{wiAfter}} {{/if}}{{#if persona}}{{persona}} {{/if}}{{trim}}[/INST] Understood. ``` For response steering, use `[OOC]` commands, e.g.: - `[OOC] Have them interrupted by a loud explosion in a nearby factory` - `[OOC] Have her refuse to sell it and suggest another merchant instead` ## Technical Details ### Training - QLoRA training at 32768 context - Merged with [gghfez/Mistral-Large-Instruct-2411](https://huggingface.co/gghfez/Mistral-Large-Instruct-2411) at bf16 - [jukofyork/Creative writing control vectors](https://huggingface.co/jukofyork/creative-writing-control-vectors-v3.0) were applied during synthetic dataset generation - Includes standard assistant instruct data for long-context stability - Note: Performance on code tasks may be reduced compared to base model - Note: No attempt was made to remove 'Name-Slop', so you'll still encounter Lily and Elara if you don't specify character names ### Context Length - Base model: 131,072 tokens - Training range: 1024-32728 tokens - Training context window: 32768 tokens ## Testing Environments Tested with exllamav2 4.5bpw on: - [tabbyAPI](https://github.com/theroyallab/tabbyAPI) + [MikuPad](https://github.com/lmg-anon/mikupad) - [tabbyAPI](https://github.com/theroyallab/tabbyAPI) + [SillyTavern](https://github.com/SillyTavern/SillyTavern) - [exui](https://github.com/turboderp/exui)