nickrosh commited on
Commit
0ed8ed3
1 Parent(s): 19aa0ae

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +1 -1
README.md CHANGED
@@ -6,7 +6,7 @@ datasets:
6
 
7
  This model uses the Evol-Instruct-Code-80k-v1 dataset generated using the [Evol-Teacher](https://github.com/nickrosh/evol-teacher) repo. Currently, WizardCoder is one the most performant Code Generation models, being beaten only by ChatGPT. This takes the Code Alpaca 20k dataset and evolves each instruction through a randomly chosen evolution prompt to increase instruction complexity. These prompts range from increase time/space complexity, to increasing requirements, to adding erroneus code to improve robustness, etc. This is done three times with pruning and post processing to remove unwanted instructions and responses. The iterative addition of more complexity gives higher quality and more in-depth instructions than what is ususally generated in Alpaca methods. This, like in the case of WizardCoder and WizardLM, can lead to strong performance that gets very close to RLHF model performance.
8
 
9
- This model was fine tuned with this generated dataset, using the following parameters:
10
  ```bash
11
  --model_name_or_path replit/replit-code-v1-3b \
12
  --data_path ./data/EvolInstruct-Code-80k/EvolInstruct-Code-80k.json \
 
6
 
7
  This model uses the Evol-Instruct-Code-80k-v1 dataset generated using the [Evol-Teacher](https://github.com/nickrosh/evol-teacher) repo. Currently, WizardCoder is one the most performant Code Generation models, being beaten only by ChatGPT. This takes the Code Alpaca 20k dataset and evolves each instruction through a randomly chosen evolution prompt to increase instruction complexity. These prompts range from increase time/space complexity, to increasing requirements, to adding erroneus code to improve robustness, etc. This is done three times with pruning and post processing to remove unwanted instructions and responses. The iterative addition of more complexity gives higher quality and more in-depth instructions than what is ususally generated in Alpaca methods. This, like in the case of WizardCoder and WizardLM, can lead to strong performance that gets very close to RLHF model performance.
8
 
9
+ This model uses [ReplitLM](https://huggingface.co/replit/replit-code-v1-3b) fine tuned with the following parameters:
10
  ```bash
11
  --model_name_or_path replit/replit-code-v1-3b \
12
  --data_path ./data/EvolInstruct-Code-80k/EvolInstruct-Code-80k.json \