RichardErkhov commited on
Commit
1030548
·
verified ·
1 Parent(s): 078ec1a

uploaded readme

Browse files
Files changed (1) hide show
  1. README.md +60 -0
README.md ADDED
@@ -0,0 +1,60 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ Quantization made by Richard Erkhov.
2
+
3
+ [Github](https://github.com/RichardErkhov)
4
+
5
+ [Discord](https://discord.gg/pvy7H8DZMG)
6
+
7
+ [Request more models](https://github.com/RichardErkhov/quant_request)
8
+
9
+
10
+ GPT-Neo-2.7B-Picard - bnb 4bits
11
+ - Model creator: https://huggingface.co/KoboldAI/
12
+ - Original model: https://huggingface.co/KoboldAI/GPT-Neo-2.7B-Picard/
13
+
14
+
15
+
16
+
17
+ Original model description:
18
+ ---
19
+ language: en
20
+ license: mit
21
+ ---
22
+ # GPT-Neo 2.7B - Picard
23
+ ## Model Description
24
+ GPT-Neo 2.7B-Picard is a finetune created using EleutherAI's GPT-Neo 2.7B model.
25
+ ## Training data
26
+ The training data contains around 1800 ebooks, mostly in the sci-fi and fantasy genres.
27
+ ### How to use
28
+ You can use this model directly with a pipeline for text generation. This example generates a different sequence each time it's run:
29
+ ```py
30
+ >>> from transformers import pipeline
31
+ >>> generator = pipeline('text-generation', model='mrseeker87/GPT-Neo-2.7B-Picard')
32
+ >>> generator("Jean-Luc Picard", do_sample=True, min_length=50)
33
+ [{'generated_text': 'Jean-Luc Picard, the captain of a Federation starship in command of one of Starfleet's few fulltime scientists.'}]
34
+ ```
35
+ ### Limitations and Biases
36
+ GPT-Neo was trained as an autoregressive language model. This means that its core functionality is taking a string of text and predicting the next token. While language models are widely used for tasks other than this, there are a lot of unknowns with this work.
37
+ GPT-Neo was trained on the Pile, a dataset known to contain profanity, lewd, and otherwise abrasive language. Depending on your usecase GPT-Neo may produce socially unacceptable text. See Sections 5 and 6 of the Pile paper for a more detailed analysis of the biases in the Pile.
38
+ As with all language models, it is hard to predict in advance how GPT-Neo will respond to particular prompts and offensive content may occur without warning. We recommend having a human curate or filter the outputs before releasing them, both to censor undesirable content and to improve the quality of the results.
39
+ ### BibTeX entry and citation info
40
+ The model is made using the following software:
41
+ ```bibtex
42
+ @software{gpt-neo,
43
+ author = {Black, Sid and
44
+ Leo, Gao and
45
+ Wang, Phil and
46
+ Leahy, Connor and
47
+ Biderman, Stella},
48
+ title = {{GPT-Neo: Large Scale Autoregressive Language
49
+ Modeling with Mesh-Tensorflow}},
50
+ month = mar,
51
+ year = 2021,
52
+ note = {{If you use this software, please cite it using
53
+ these metadata.}},
54
+ publisher = {Zenodo},
55
+ version = {1.0},
56
+ doi = {10.5281/zenodo.5297715},
57
+ url = {https://doi.org/10.5281/zenodo.5297715}
58
+ }
59
+ ```
60
+