Model Card: Pygmalion-2-13b-SuperCOT-weighted
This is an experimental weighted merge between:
Quantizations provided by us and TheBloke:
The merge was performed by a gradient merge script (apply-lora-weight-ltl.py) from zaraki-tools by Zaraki.
Thanks to Zaraki for the inspiration and help.
This merge differs from the previous Pyg-2-SuperCOT merges. The first iteration of the SuperCOT loras were used here since it performed better than SuperCOT2.
The SuperCOT lora was merged with the following layer weights (basically 50/50. The exact ratio is 0.51)
0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0.5,1,1,1,1,1,1,1,1,1,1,1,1,1,1,1,1,1,1,1,1
Here is an image to help visualize this merge. The light blue is Pygmalion-2-13b and the light green is the SuperCOT lora:
Usage:
Since this is an experimental weight merge between Pygmalion-2 and SuperCOT, the following instruction formats should work:
Metharme:
<|system|>This is a text adventure game. Describe the scenario to the user and give him three options to pick from on each turn.<|user|>Start!<|model|>
Alpaca:
### Instruction:
Your instruction or question here.
### Response:
Bias, Risks, and Limitations
The model will show biases similar to those observed in niche roleplaying forums on the Internet, besides those exhibited by the base model. It is not intended for supplying factual information or advice in any form.
In addition, this merge is experimental from our own testing. Your results may vary.
Training Details
This model is merged and can be reproduced using the tools mentioned above. Please refer to all provided links for extra model-specific details.
- Downloads last month
- 16