Model Card for Self-instruct-starcoder

This model is an instruction-tuned version of ⭐️ StarCoder. The instruction dataset involved is Self-instruct-starcoder which was built by boostrapping on StarCoder's generations.

Uses

The model was fine-tuned with the following template

Question: <instruction>

Answer: <output>

If you have your model and tokenizer loaded, you can use the following code to make the model generate the right output to a given instruction

instruction = "Write a function to compute the GCD between two integers a and b"
prompt = f"Question:{instruction}\n\nAnswer:"
input_ids = tokenizer(prompt, return_tensors="pt")["input_ids"]
completion = model.generate(input_ids, max_length=200)
print(tokenizer.batch_decode(completion[:,input_ids.shape[1]:])[0])

More information

For additional information, check

Downloads last month
21
Inference Examples
This model does not have enough activity to be deployed to Inference API (serverless) yet. Increase its social visibility and check back later, or deploy to Inference Endpoints (dedicated) instead.

Dataset used to train codeparrot/starcoder-self-instruct

Spaces using codeparrot/starcoder-self-instruct 2

Evaluation results