Aeala's picture
Update README.md
6f1d0ba
|
raw
history blame
635 Bytes
metadata
datasets:
  - Aeala/ShareGPT_Vicuna_unfiltered

LoRA Info:

Please note that this is a highly experimental LoRA model. It may do some good stuff, it might do some undesirable stuff. Training is paused for now. Feel free to try it!~

Important Note: While this is trained on a cleaned ShareGPT dataset like Vicuna used, this was trained in the Alpaca format, so prompting should be something like:

### Instruction:

<prompt> (without the <>)

### Response:

Current upload: checkpoint of step 1200 in training.

Benchmarks

wikitext2: Coming soon...

ptb-new: Coming soon...

c4-new: Coming soon...