File size: 1,034 Bytes
da804e0
23dc3e1
42c2d33
23dc3e1
e6c12e5
4f611f8
e6c12e5
1
2
3
4
5
6
7
A natural way to evaluate code programs is to see if they pass unit tests, it is the idea behind the [pass@k](https://huggingface.co/metrics/code_eval) metric, a popular evaluation framework for code generation models, on [HumanEval](https://huggingface.co/datasets/openai_humaneval) dataset, which was introduced in [Codex paper](https://arxiv.org/pdf/2107.03374v2.pdf). The dataset includes 164 handwritten programming problems. In the pass@k metric, k code samples are generated per problem, and a problem is considered solved if any sample passes the unit tests and the total fraction of problems solved is reported.
In most papers, 200 candidate program completions are sampled, and pass@1, pass@10, and pass@100 are computed using an unbiased sampling estimator. 

This plot shows the pass@100 by model size, for CodeParrot, InCoder, PolyCoder, CodeGen and Codex (not open-source):
<p align="center">
    <img src="https://huggingface.co/datasets/loubnabnl/repo-images/resolve/main/pass@100.png" alt="drawing" width="550"/>
</p>