matthayes commited on
Commit
79f6639
1 Parent(s): bb5cd0b

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +3 -3
README.md CHANGED
@@ -9,8 +9,8 @@ inference: false
9
  ## Summary
10
 
11
  Databricks’ `dolly-v2-12b`, an instruction-following large language model trained on the Databricks machine learning platform
12
- that is licensed for commercial use. based on `pythia-12b`, Dolly is trained on ~15k instruction/response fine tuning records
13
- [`databricks-dolly-15k`](https://huggingface.co/datasets/databricks/databricks-dolly-15k) generated
14
  by Databricks employees in capability domains from the InstructGPT paper, including brainstorming, classification, closed QA, generation,
15
  information extraction, open QA and summarization. `dolly-v2-12b` is not a state-of-the-art model, but does exhibit surprisingly
16
  high quality instruction following behavior not characteristic of the foundation model on which it is based.
@@ -20,7 +20,7 @@ high quality instruction following behavior not characteristic of the foundation
20
  ## Model Overview
21
  `dolly-v2-12b` is a 12 billion parameter causal language model created by [Databricks](https://databricks.com/) that is derived from
22
  [EleutherAI’s](https://www.eleuther.ai/) [Pythia-12b](https://huggingface.co/EleutherAI/pythia-12b) and fine-tuned
23
- on a ~15K record instruction corpus generated by Databricks employees and released under a permissive license (CC-BY-SA)
24
 
25
  ## Usage
26
 
 
9
  ## Summary
10
 
11
  Databricks’ `dolly-v2-12b`, an instruction-following large language model trained on the Databricks machine learning platform
12
+ that is licensed for commercial use. Based on `pythia-12b`, Dolly is trained on ~15k instruction/response fine tuning records
13
+ [`databricks-dolly-15k`](https://github.com/databrickslabs/dolly/tree/master/data) generated
14
  by Databricks employees in capability domains from the InstructGPT paper, including brainstorming, classification, closed QA, generation,
15
  information extraction, open QA and summarization. `dolly-v2-12b` is not a state-of-the-art model, but does exhibit surprisingly
16
  high quality instruction following behavior not characteristic of the foundation model on which it is based.
 
20
  ## Model Overview
21
  `dolly-v2-12b` is a 12 billion parameter causal language model created by [Databricks](https://databricks.com/) that is derived from
22
  [EleutherAI’s](https://www.eleuther.ai/) [Pythia-12b](https://huggingface.co/EleutherAI/pythia-12b) and fine-tuned
23
+ on a [~15K record instruction corpus](https://github.com/databrickslabs/dolly/tree/master/data) generated by Databricks employees and released under a permissive license (CC-BY-SA)
24
 
25
  ## Usage
26