R136a1 commited on
Commit
5c0c11c
·
verified ·
1 Parent(s): 20fb157

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +2 -8
README.md CHANGED
@@ -7,15 +7,9 @@ tags:
7
  - mixtral
8
  ---
9
 
10
- I prefer this one instead of v1 since it's a bit more creative and _smart_, understand the story better. This use some different models from the v1 but perform very close to it (I guess since I used the same model for the base?). And yeah, I'm using this as daily model for RP (llama 3 is smart but not creative enough for RP, I think because lack of RP and _ERP_ data in their training set? idk.)
11
 
12
 
13
- Testing done.
14
-
15
- It performs really well in complex scenario and follows the character card quite well. The char card and previous message can affect a lot to the next reply style.
16
-
17
- The main idea is instead of _merging_ models to create new model, I try to put these best model into mixtral so it can work together. And the result is good, every model has its uniqueness and strength.
18
-
19
- Downside? it only support 8k (8192) context length...
20
 
21
  Alpaca prompting format.
 
7
  - mixtral
8
  ---
9
 
10
+ Exl2 of an old model. Found it still _really_ good and I'm still using it sometimes so 4.6bpw exl2 here is.
11
 
12
 
13
+ Only support 8k (8192) context length...
 
 
 
 
 
 
14
 
15
  Alpaca prompting format.