DaringLotus-10.7B 4bpw EXL2
Description
EXL2 quant of BlueNipples/DaringLotus-10.7B
- 6bpw should be comfortable on 12 gb with 8k context
- 4bpw might just fit on 8gb of vram at 4k context
- if you have more ram get the 8bpw
Other quants:
Prompt Format
Alpaca:
I am not entirely certain of this but i think alpaca is correct for this model
Below is an instruction that describes a task. Write a response that appropriately completes the request.
### Instruction:
{prompt}
### Input:
{input}
### Response:
Contact
Kooten on discord
- Downloads last month
- 4
Inference Providers
NEW
This model is not currently available via any of the supported third-party Inference Providers, and
the model is not deployed on the HF Inference API.