Merry commited on
Commit
10b86c7
1 Parent(s): f2a6e85

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +5 -4
README.md CHANGED
@@ -11,12 +11,13 @@ inference: false
11
 
12
  *(Not to be confused with [Pygmalion 13B](https://huggingface.co/TehVenom/Pygmalion-13b-GGML).)*
13
 
14
- # Converted with ggerganov/ggml's gpt-neox conversion script, and tested with KoboldCpp.
15
- ## *(I can't promise that this will work with other frontends, if at all; I haven't had the most success myself. Use at your own risk!)*
16
-
17
  This is converted and quantized from [Pygmalion 1.3B](https://huggingface.co/PygmalionAI/pygmalion-1.3b), based on [an earlier version](https://huggingface.co/EleutherAI/pythia-1.4b-deduped-v0) of Pythia 1.4B Deduped.
18
 
19
- # RAM USAGE (on KoboldCpp w/ OpenBLAS)
 
 
 
 
20
  Model | Initial RAM
21
  :--:|:--:
22
  ggml-pygmalion-1.3b-q4_0.bin | 1.1 GiB
 
11
 
12
  *(Not to be confused with [Pygmalion 13B](https://huggingface.co/TehVenom/Pygmalion-13b-GGML).)*
13
 
 
 
 
14
  This is converted and quantized from [Pygmalion 1.3B](https://huggingface.co/PygmalionAI/pygmalion-1.3b), based on [an earlier version](https://huggingface.co/EleutherAI/pythia-1.4b-deduped-v0) of Pythia 1.4B Deduped.
15
 
16
+ Notes:
17
+ - Converted with ggerganov/ggml's gpt-neox conversion script, and tested with KoboldCpp.
18
+ - I can't promise that this will work with other frontends, if at all. I've had problems with the tokenizer. Could be related to the ggml implementation of GPT-NeoX [(source)](https://github.com/ggerganov/ggml/tree/master/examples/gpt-neox#notes).
19
+
20
+ ### RAM USAGE (on KoboldCpp w/ OpenBLAS)
21
  Model | Initial RAM
22
  :--:|:--:
23
  ggml-pygmalion-1.3b-q4_0.bin | 1.1 GiB