Model Card for Self-instruct-starcoder
This model is an instruction-tuned version of βοΈ StarCoder. The instruction dataset involved is Self-instruct-starcoder which was built by boostrapping on StarCoder's generations.
Uses
The model was fine-tuned with the following template
Question: <instruction>
Answer: <output>
If you have your model and tokenizer loaded, you can use the following code to make the model generate the right output to a given instruction
instruction = "Write a function to compute the GCD between two integers a and b"
prompt = f"Question:{instruction}\n\nAnswer:"
input_ids = tokenizer(prompt, return_tensors="pt")["input_ids"]
completion = model.generate(input_ids, max_length=200)
print(tokenizer.batch_decode(completion[:,input_ids.shape[1]:])[0])
More information
For additional information, check
- Downloads last month
- 21
This model does not have enough activity to be deployed to Inference API (serverless) yet. Increase its social
visibility and check back later, or deploy to Inference Endpoints (dedicated)
instead.
Dataset used to train codeparrot/starcoder-self-instruct
Spaces using codeparrot/starcoder-self-instruct 2
Evaluation results
- pass@1 on InstructHumanEvalself-reported0.391
- pass@1 on HumanEvalself-reported0.346