Edit model card

Model Card for PygKiCOTlion

PygKiCOTlion is a a lora merge of Pygmalion-2-13b-SuperCOT + Kimiko v2

Model Details

Q: "Why do you do this?!"
A: Was bored.

Model Description

  • Developed by: KaraKaraWitch (Merge), kaiokendev (Original SuperCOT LoRA), nRuaif (Kimiko v2 LoRA), kingbri (Pygmalion 2 13b SuperCOT)
  • Model type: Decoder only
  • License: LLaMA2 (PygKiCOTlion), SuperCOT (MIT), Kimiko v2 (CC BY-NC-SA (?))
  • Finetuned from model [optional]: LLaMA2

Model Sources [optional]

Uses

YYMV.

Direct Use

Usage:

Since this is a merge between Pygmalion 2 13b SuperCOT and Kimiko v2, the following instruction formats should work:

Metharme:

<|system|>Your system prompt goes here.<|user|>Are you alive?<|model|>

Alpaca:

### Instruction:
Your instruction or question here.
### Response:

Bias, Risks, and Limitations

YMMV. This is untested territory.

Testing Feedbakc

Notes from KaraKaraWitch:

  • The model feels weirdly loopy compared to MythKiCOTlion at lower temps.
  • Higher temps the model tries to venture out of it's comfort zone at the cost of making it not stick to the model card as close as expected.

Training Details

N/A. Refer to the respective LoRa's and models.

Downloads last month
8
Inference Examples
This model does not have enough activity to be deployed to Inference API (serverless) yet. Increase its social visibility and check back later, or deploy to Inference Endpoints (dedicated) instead.

Model tree for KaraKaraWitch/PygKiCOTlion

Quantizations
2 models