kubernetes-bad commited on
Commit
eda3e8d
1 Parent(s): 807b26c

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +46 -7
README.md CHANGED
@@ -1,16 +1,29 @@
1
  ---
2
- library_name: peft
 
 
 
 
 
3
  ---
4
 
5
- # Character Making Lora: test 2
6
 
7
- Changes: trained on more uniform, curated subset of test-1 cards, with scenario included into instruction.
8
 
9
- For now, it only supports plaintext cards. Try plist/w++/etc at your own risk.
10
 
11
- Address {{user}} as `User`. Character descriptions work best if they begin with `CharacterName is a ...` - for example "Martha is a middle-aged woman who is ..."
12
 
13
- Use alpaca template with the following instruction:
 
 
 
 
 
 
 
 
14
 
15
  ```
16
  Below is an instruction that describes a task, paired with an input that provides further context. Write a response that appropriately completes the request.
@@ -38,4 +51,30 @@ Fuckana is a friendly and talkative catgirl that has enormous breasts. Her voice
38
 
39
  ### Response:
40
 
41
- ```
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
  ---
2
+ license: llama2
3
+ language:
4
+ - en
5
+ tags:
6
+ - roleplay
7
+ - characters
8
  ---
9
 
10
+ # CharGen v1
11
 
12
+ > A model for creating characters for role play.
13
 
14
+ Trained on *lots* of character cards both from chub and janitor, with some post-processing.
15
 
16
+ For now, it only supports plaintext cards. Any other variation like plist/w++/etc is entirely untested.
17
 
18
+ Address {{user}} as `User`. Character descriptions work best if they begin with `CharacterName is a ...` - for example *"Martha is a middle-aged woman who is ..."*
19
+
20
+ NB: This model is **NOT** for roleplay directly. It creates characters that can then be used in roleplay with some other model like [MythoMax](https://huggingface.co/Gryphe/MythoMax-L2-13b).
21
+
22
+ It was trained on dynamic prompt template, so it should be able to accommodate your changes to the prompt.
23
+
24
+ Trained as a LoRA, the released model is a merge with [Airoboros 2.2](https://huggingface.co/jondurbin/airoboros-l2-13b-2.2) for extra-good instruction following.
25
+
26
+ Prompt template:
27
 
28
  ```
29
  Below is an instruction that describes a task, paired with an input that provides further context. Write a response that appropriately completes the request.
 
51
 
52
  ### Response:
53
 
54
+ ```
55
+
56
+ ### Dataset
57
+
58
+ ~34,000 cards from CharacterHub and another ~80,000 cards from Janitor were used as initial dataset, as captured in period between August and September 2023.
59
+
60
+ Dataset will not be released, unless explicit permission to do so would be granted from both Chub and Janitor.
61
+
62
+
63
+ ## Training procedure
64
+
65
+
66
+ The following `bitsandbytes` quantization config was used during training:
67
+ - quant_method: bitsandbytes
68
+ - load_in_8bit: True
69
+ - load_in_4bit: False
70
+ - llm_int8_threshold: 6.0
71
+ - llm_int8_skip_modules: None
72
+ - llm_int8_enable_fp32_cpu_offload: False
73
+ - llm_int8_has_fp16_weight: False
74
+ - bnb_4bit_quant_type: fp4
75
+ - bnb_4bit_use_double_quant: False
76
+ - bnb_4bit_compute_dtype: float32
77
+ ### Framework versions
78
+
79
+
80
+ - PEFT 0.6.0.dev0