Text Generation
Transformers
PyTorch
English
gptj
mike-conover-db commited on
Commit
679c999
1 Parent(s): 0e76388

Updating Model Card

Browse files
Files changed (1) hide show
  1. README.md +2 -2
README.md CHANGED
@@ -11,7 +11,7 @@ library_name: transformers
11
 
12
  Databricks’ Dolly, a large language model trained on the [Databricks Machine Learning Platform](https://www.databricks.com/product/machine-learning), demonstrates that a
13
  two-years-old [open source model](https://huggingface.co/EleutherAI/gpt-j-6B) can, when subjected to just 30 minutes of fine tuning on a focused corpus of 50k records
14
- ([Stanford Alpaca](https://crfm.stanford.edu/2023/03/13/alpaca.html)), exhibit surprisingly high quality instruction following behavior not characteristic of the foundation
15
  model on which it is based. We believe this finding is important because it demonstrates that the ability to create powerful
16
  artificial intelligence technologies is vastly more accessible than previously realized.
17
 
@@ -39,7 +39,7 @@ competitively with more modern model architectures or models subject to larger p
39
 
40
  The Dolly model family is under active development, and so any list of shortcomings is unlikely to be exhaustive, but we include known limitations and misfires here as a means to document and share our preliminary findings with the community.
41
  In particular, `dolly-v1-6b` struggles with: syntactically complex prompts, programming problems, mathematical operations, factual errors,
42
- dates and times, open-ended question answering, hallucination, enumerating lists of specific length, stylistic mimicry, etc.
43
 
44
  ## Training Data, Bias & Objectionable Content
45
  Like all language models, `dolly-v1-6b` reflects the content and limitations of its training corpuses.
 
11
 
12
  Databricks’ Dolly, a large language model trained on the [Databricks Machine Learning Platform](https://www.databricks.com/product/machine-learning), demonstrates that a
13
  two-years-old [open source model](https://huggingface.co/EleutherAI/gpt-j-6B) can, when subjected to just 30 minutes of fine tuning on a focused corpus of 50k records
14
+ ([Stanford Alpaca](https://crfm.stanford.edu/2023/03/13/alpaca.html)), exhibits surprisingly high quality instruction following behavior not characteristic of the foundation
15
  model on which it is based. We believe this finding is important because it demonstrates that the ability to create powerful
16
  artificial intelligence technologies is vastly more accessible than previously realized.
17
 
 
39
 
40
  The Dolly model family is under active development, and so any list of shortcomings is unlikely to be exhaustive, but we include known limitations and misfires here as a means to document and share our preliminary findings with the community.
41
  In particular, `dolly-v1-6b` struggles with: syntactically complex prompts, programming problems, mathematical operations, factual errors,
42
+ dates and times, open-ended question answering, hallucination, enumerating lists of specific length, stylistic mimicry, having a sense of humor, etc.
43
 
44
  ## Training Data, Bias & Objectionable Content
45
  Like all language models, `dolly-v1-6b` reflects the content and limitations of its training corpuses.