File size: 635 Bytes
96bd8fc
 
6f1d0ba
96bd8fc
 
 
7a62ac3
96bd8fc
 
 
 
 
 
 
 
 
 
 
081c115
96bd8fc
 
 
 
 
 
 
 
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
---
datasets:
- Aeala/ShareGPT_Vicuna_unfiltered
---

## LoRA Info:
Please note that this is a highly experimental LoRA model. It may do some good stuff, it might do some undesirable stuff. Training is paused for now. Feel free to try it!~

**Important Note**: While this is trained on a cleaned ShareGPT dataset like Vicuna used, this was trained in the *Alpaca* format, so prompting should be something like:

```
### Instruction:

<prompt> (without the <>)

### Response:
```

Current upload: checkpoint of step 1200 in training.


## Benchmarks
**wikitext2:** Coming soon...

**ptb-new:** Coming soon...

**c4-new:** Coming soon...