zarakiquemparte commited on
Commit
41d0385
1 Parent(s): fa17c8c

Create README.md

Browse files
Files changed (1) hide show
  1. README.md +58 -0
README.md ADDED
@@ -0,0 +1,58 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ ---
2
+ license: other
3
+ library_name: peft
4
+ tags:
5
+ - llama2
6
+ ---
7
+
8
+
9
+ # Model Card for PIPPA ShareGPT Subset Lora 7b
10
+
11
+ It is an experimental Lora focused on Roleplay that uses a subset of (PIPPA ShareGPT)[https://huggingface.co/datasets/kingbri/PIPPA-shareGPT]
12
+
13
+ ## Usage
14
+
15
+ ShareGPT(Vicuna 1.1)
16
+ ```
17
+ A chat.
18
+ USER: {prompt}
19
+ ASSISTANT:
20
+ ```
21
+
22
+ ## Bias, Risks, and Limitations
23
+
24
+ This Lora is not intended for supplying factual information or advice in any form
25
+
26
+ [More Information Needed]
27
+
28
+ ## Training Details
29
+
30
+ ### Training Data
31
+
32
+ 1k of conversation from (PIPPA ShareGPT)[https://huggingface.co/datasets/kingbri/PIPPA-shareGPT]
33
+
34
+ ### Training Procedure
35
+
36
+ The version of this Lora uploaded on this repository was trained using a 8x RTX A6000 cluster in 8-bit with regular LoRA adapters and 32-bit AdamW optimizer.
37
+
38
+ #### Training Hyperparameters
39
+
40
+ Training using a fork of axolot with two paths [Patch 1](https://github.com/bdashore3/axolotl/commit/995557bdf3c6c8b3e839b224ef9513fc2b097f30) [Patch 2](https://github.com/bdashore3/axolotl/commit/8970280de2ea01e41c044406051922715f4086cb)
41
+
42
+ - lora_r: 16
43
+ - lora_alpha: 16
44
+ - lora_dropout: 0.01
45
+ - gradient_accumulation_steps: 6
46
+ - micro_batch_size: 4
47
+ - num_epochs: 3
48
+ - learning_rate: 0.000065
49
+
50
+ ## Model Examination [optional]
51
+
52
+ <!-- Relevant interpretability work for the model goes here -->
53
+
54
+ [More Information Needed]
55
+
56
+ ## Environmental Impact
57
+
58
+ Finetuning this model on 8xNVIDIA A6000 48GB in parallel takes about 30 minutes (7B)