Edit model card

How Do Humans Write Code? Large Models Do It the Same Way Too

Paper: https://arxiv.org/pdf/2402.15729

Code: https://github.com/seamoke/Human-Think-Language

Introduction

For this model, please sure your transformers>=4.39.2.

We introduce HTL, a model which utilizes the complete reasoning process of CoT to enhance PoT. This model was secondarily fine-tuned based on MAmmoTH-Coder-7B

Evaluation

The models are evaluated using open-ended and multiple-choice math problems from several datasets. Here are the results:

Model GSM GSM-Hard NumGLUE MATH Sim SVAMP MAWPS ASDiV
MAmmoTH-Coder-7B 59.4 56.3 66.4 33.4 45.9 70.7 91.9 69.3
TORA 72.6 56.0 46.2 44.6 48.5 70.4 91.3 78.7
MAmmoTH-Coder-7B 65.7 58.3 75.1 34.9 50.8 74.4 94.2 73.1

Prompt Format

If you want to do HTL:

Below is an instruction that describes a task. Write a response that appropriately completes the request.
I'd like you to solve this problem in 3 steps:
1.Answer the question in plain language without writing any code.\n
2.Output one line of *\n.
3.Write program code based on the solution process in step 1 to solve the problem.\n
### Instruction:
{query}
Let's write a program.
### Response:"

Citation

If you use the models, data, or code from this project, please cite the original paper:

@article{li2024humans,
  title={How Do Humans Write Code? Large Models Do It the Same Way Too},
  author={Li, Long},
  journal={arXiv preprint arXiv:2402.15729},
  year={2024}
}
Downloads last month
1
Safetensors
Model size
6.74B params
Tensor type
BF16
·
Inference API
Model is too large to load in Inference API (serverless). To try the model, launch it on Inference Endpoints (dedicated) instead.

Dataset used to train seamoke111/HTL-CodeLlama-7B