PocketDoc commited on
Commit
2ee1edc
1 Parent(s): 8bf020f

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +63 -1
README.md CHANGED
@@ -3,4 +3,66 @@ language:
3
  - en
4
  ---
5
 
6
- This model was created through the merging of this LoRA https://huggingface.co/PocketDoc/Dans-PileOfSets-Mk1-llama-13b-gptq-4bit-128g-LoRA and this model https://huggingface.co/huggyllama/llama-13b
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
3
  - en
4
  ---
5
 
6
+ ### Description:
7
+ This is a llama 13b model merge of the LoRA with the same name.
8
+
9
+ ### Objective for this project:
10
+ To create a model that upholds a logical thread, regardless of whether the output is verbose or concise. Training has been performed on a version of the pile of sets, reduced to 40% of its original size, to expedite training iterations. I personally utilize this model as an aid for storytelling and writing. While it serves this purpose adequately, I still perceive this version as a prototype.
11
+
12
+ ### Prompt format:
13
+ Stanford Alpaca
14
+
15
+ The prompt should start on a new line after "### Response:"
16
+ - For examples with a non-empty input field:
17
+
18
+ ```
19
+ Below is an instruction that describes a task, paired with an input that provides further context. Write a response that appropriately completes the request.
20
+
21
+ ### Instruction:
22
+ {instruction}
23
+
24
+ ### Input:
25
+ {input}
26
+
27
+ ### Response:
28
+ ```
29
+
30
+ - For examples with an empty input field:
31
+
32
+ ```
33
+ Below is an instruction that describes a task. Write a response that appropriately completes the request.
34
+
35
+ ### Instruction:
36
+ {instruction}
37
+
38
+ ### Response:
39
+ ```
40
+
41
+ ### Training information
42
+ - 2 Epochs
43
+ - 64 / 32 R / A
44
+ - 1024 Cutoff
45
+ - 19 hours on an A6000
46
+
47
+ ### Data used in training
48
+ All cleaned and scrubbed in various ways then culled to various degrees.
49
+ - Camel biology, physics, chemistry, math, and AI society
50
+ - Alpaca evol instruct
51
+ - GPTeacher Instruct
52
+ - Alpaca GPT4
53
+ - Dolly Databricks
54
+
55
+ ### Plans for the future, a brief overview:
56
+ - Pivot to a conversational format going forward
57
+ - Train another 13b LoRA against the entirety of my pile of sets rather than just a portion of it for Mk2
58
+ - Train 30b on the Mk2 pile of sets
59
+ - Expand the story generation capabilities and likely more for Mk3
60
+
61
+ ### Model used for training and other information:
62
+ https://huggingface.co/PocketDoc/llama-13b-gptq-4bit-128g
63
+
64
+ Merge model:
65
+ https://huggingface.co/huggyllama/llama-13b
66
+
67
+ ### Disclaimer:
68
+ It has not been aligned and no warranty is given for the quality or safety of its outputs.