rooa commited on
Commit
cf90c71
1 Parent(s): 9aaa0ce

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +4 -4
README.md CHANGED
@@ -8,7 +8,7 @@ license: apache-2.0
8
 
9
  [CodeGen2](https://github.com/salesforce/CodeGen2) is a family of autoregressive language models for **program synthesis**, introduced in the paper:
10
 
11
- [CodeGen2: Lessons for Training LLMs on Programming and Natural Languages]() by Erik Nijkamp\*, Hiroaki Hayashi\*, Caiming Xiong, Silvio Savarese, Yingbo Zhou.
12
 
13
  Unlike the original CodeGen model family (i.e., CodeGen1), CodeGen2 is capable of infilling, and supports more programming languages.
14
 
@@ -76,7 +76,7 @@ You might want to truncate the model output with `<eom>`.
76
 
77
  ## Training data
78
 
79
- This checkpoint is trained on the stricter permissive subset of [the deduplicated version of the Stack dataset (v1.1)](). Supported languages (and frameworks) are as follows:
80
  `c`, `c++`, `c-sharp`, `dart`, `go`, `java`, `javascript`, `kotlin`, `lua`, `php`, `python`, `ruby`, `rust`, `scala`, `shell`, `sql`, `swift`, `typescript`, `vue`.
81
 
82
  ## Training procedure
@@ -87,7 +87,7 @@ Please refer to the paper for more details.
87
 
88
  ## Evaluation results
89
 
90
- We evaluate our models on HumanEval and HumanEval-Infill. Please refer to the [paper]() for more details.
91
 
92
  ## Intended use and limitations
93
 
@@ -102,6 +102,6 @@ However, the model is intended for and best at **program synthesis**, that is, g
102
  title={CodeGen2: Lessons for Training LLMs on Programming and Natural Languages},
103
  author={Nijkamp, Erik and Hayashi, Hiroaki and Xiong, Caiming and Savarese, Silvio and Zhou, Yingbo},
104
  journal={arXiv preprint},
105
- year={2022}
106
  }
107
  ```
 
8
 
9
  [CodeGen2](https://github.com/salesforce/CodeGen2) is a family of autoregressive language models for **program synthesis**, introduced in the paper:
10
 
11
+ [CodeGen2: Lessons for Training LLMs on Programming and Natural Languages](https://arxiv.org/abs/2305.02309) by Erik Nijkamp\*, Hiroaki Hayashi\*, Caiming Xiong, Silvio Savarese, Yingbo Zhou.
12
 
13
  Unlike the original CodeGen model family (i.e., CodeGen1), CodeGen2 is capable of infilling, and supports more programming languages.
14
 
 
76
 
77
  ## Training data
78
 
79
+ This checkpoint is trained on the stricter permissive subset of [the deduplicated version of the Stack dataset (v1.1)](https://huggingface.co/datasets/bigcode/the-stack-dedup). Supported languages (and frameworks) are as follows:
80
  `c`, `c++`, `c-sharp`, `dart`, `go`, `java`, `javascript`, `kotlin`, `lua`, `php`, `python`, `ruby`, `rust`, `scala`, `shell`, `sql`, `swift`, `typescript`, `vue`.
81
 
82
  ## Training procedure
 
87
 
88
  ## Evaluation results
89
 
90
+ We evaluate our models on HumanEval and HumanEval-Infill. Please refer to the [paper](https://arxiv.org/abs/2305.02309) for more details.
91
 
92
  ## Intended use and limitations
93
 
 
102
  title={CodeGen2: Lessons for Training LLMs on Programming and Natural Languages},
103
  author={Nijkamp, Erik and Hayashi, Hiroaki and Xiong, Caiming and Savarese, Silvio and Zhou, Yingbo},
104
  journal={arXiv preprint},
105
+ year={2023}
106
  }
107
  ```