Merry commited on
Commit
efc0a40
1 Parent(s): 3fff0b7

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +62 -3
README.md CHANGED
@@ -1,13 +1,72 @@
1
  ---
2
  license: agpl-3.0
 
 
 
 
 
 
 
3
  ---
4
 
5
- ### (Not to be confused with [Pygmalion 13B](https://huggingface.co/TehVenom/Pygmalion-13b-GGML).)
6
 
7
- This is converted and quantized from [Pygmalion 1.3B](https://huggingface.co/PygmalionAI/pygmalion-1.3b), based on [an earlier version of Pythia 1.4B Deduped](https://huggingface.co/EleutherAI/pythia-1.4b-deduped-v0).
8
 
9
  # RAM USAGE (on KoboldCpp w/ OpenBLAS)
10
  Model | Initial RAM
11
  :--:|:--:
12
  ggml-pygmalion-1.3b-q4_0.bin | 1.1 GiB
13
- ggml-pygmalion-1.3b-q5_1.bin | 1.3 GiB
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
  ---
2
  license: agpl-3.0
3
+ language:
4
+ - en
5
+ thumbnail:
6
+ tags:
7
+ - text generation
8
+ - conversational
9
+ inference: false
10
  ---
11
 
12
+ *(Not to be confused with [Pygmalion 13B](https://huggingface.co/TehVenom/Pygmalion-13b-GGML).)*
13
 
14
+ This is converted and quantized from [Pygmalion 1.3B](https://huggingface.co/PygmalionAI/pygmalion-1.3b), based on [an earlier version](https://huggingface.co/EleutherAI/pythia-1.4b-deduped-v0) of Pythia 1.4B Deduped.
15
 
16
  # RAM USAGE (on KoboldCpp w/ OpenBLAS)
17
  Model | Initial RAM
18
  :--:|:--:
19
  ggml-pygmalion-1.3b-q4_0.bin | 1.1 GiB
20
+ ggml-pygmalion-1.3b-q5_1.bin | 1.3 GiB
21
+
22
+ Below is the original model card.
23
+
24
+ * * *
25
+
26
+ # Pygmalion 1.3B
27
+
28
+ ## Model description
29
+
30
+ Pymalion 1.3B is a proof-of-concept dialogue model based on EleutherAI's [pythia-1.3b-deduped](https://huggingface.co/EleutherAI/pythia-1.3b-deduped).
31
+
32
+ **Warning:** This model is **NOT** suitable for use by minors. It **will** output X-rated content under certain circumstances.
33
+
34
+ ## Training data
35
+
36
+ The fine-tuning dataset consisted of 56MB of dialogue data gathered from multiple sources, which includes both real _and_ partially machine-generated conversations.
37
+
38
+ ## Training procedure
39
+
40
+ Fine-tuning was done using [ColossalAI](https://github.com/hpcaitech/ColossalAI) (specifically, with a slightly modified version of their [OPT fine-tune example](https://github.com/hpcaitech/ColossalAI/blob/78509124d32b63b7fc36f6508e0576a326d51422/examples/language/opt/run_clm.py)) for around 11.4 million tokens over 5440 steps on a single 24GB GPU. The run took just under 21 hours.
41
+
42
+ ## Intended use
43
+
44
+ ### The easy way
45
+
46
+ We provide a notebook with a Gradio UI for playing around with the model without having to manually format inputs. This notebook can be found [here](https://github.com/PygmalionAI/gradio-ui/blob/master/notebooks/GPU.ipynb).
47
+
48
+ ### The manual way
49
+
50
+ The model can be used as a regular text generation model, but it'll perform best if the input prompt adheres to the following format:
51
+
52
+ ```
53
+ [CHARACTER]'s Persona: [A few sentences about the character you want the model to play]
54
+
55
+ [DIALOGUE HISTORY]
56
+ You: [Your input message here]
57
+ [CHARACTER]:
58
+ ```
59
+
60
+ Where `[CHARACTER] `is, as you can probably guess, the name of the character you want the model to portray, and `[DIALOGUE HISTORY]` is chat history so the model can have some conversational context to draw from. Ideally it'll be pairs of messages like:
61
+
62
+ ```
63
+ [CHARACTER]: [some dialogue here]
64
+ You: [your response to the dialogue above]
65
+ ```
66
+
67
+ Apart from chat history, you can also just add example conversations in `[DIALOGUE HISTORY]` to show how the character should speak - ideally at the beginning, so it doesn't get confused as to what's conversation history vs. character definition.
68
+
69
+ ## Known issues
70
+
71
+ - The model can get stuck repeating certain phrases, or sometimes even entire sentences.
72
+ - We believe this is due to that behavior being present in the training data itself, and plan to investigate and adjust accordingly for future versions.