Edit model card

How Do Humans Write Code? Large Models Do It the Same Way Too

Paper: https://arxiv.org/pdf/2402.15729

Code: https://github.com/seamoke/Human-Think-Language

Introduction

For this model, please sure your transformers>=4.39.2.

We introduce HTL, a model which utilizes the complete reasoning process of CoT to enhance PoT. This model was secondarily fine-tuned based on MAmmoTH-Coder-7B

Evaluation

The models are evaluated using open-ended and multiple-choice math problems from several datasets. Here are the results:

Model GSM GSM-Hard NumGLUE MATH Sim SVAMP MAWPS ASDiV
MAmmoTH-Coder-7B 59.4 56.3 66.4 33.4 45.9 70.7 91.9 69.3
TORA 72.6 56.0 46.2 44.6 48.5 70.4 91.3 78.7
MAmmoTH-Coder-7B 65.7 58.3 75.1 34.9 50.8 74.4 94.2 73.1

Prompt Format

If you want to do HTL:

Below is an instruction that describes a task. Write a response that appropriately completes the request.
I'd like you to solve this problem in 3 steps:
1.Answer the question in plain language without writing any code.\n
2.Output one line of *\n.
3.Write program code based on the solution process in step 1 to solve the problem.\n
### Instruction:
{query}
Let's write a program.
### Response:"

Citation

If you use the models, data, or code from this project, please cite the original paper:

@article{li2024humans,
  title={How Do Humans Write Code? Large Models Do It the Same Way Too},
  author={Li, Long},
  journal={arXiv preprint arXiv:2402.15729},
  year={2024}
}
Downloads last month
6
Safetensors
Model size
6.74B params
Tensor type
BF16
·
Inference Examples
This model does not have enough activity to be deployed to Inference API (serverless) yet. Increase its social visibility and check back later, or deploy to Inference Endpoints (dedicated) instead.

Dataset used to train seamoke111/HTL-CodeLlama-7B