The score of this model on spider and WikiSQL

#7
by GZL11 - opened

I would like to know the score of this model on spider and WikiSQL. I'm not sure if you have submitted any results.
Also, do you have any scores on the test set!
Looking forward to a reply!

NumbersStation org

Here are some benchmark scores we have: https://www.numbersstation.ai/post/nsql-llama-2-7b and all numbers are on dev set.

Do you mind to share the evaluation script or the prompt template for us to duplicate the numbers? Many thanks

In the page https://www.numbersstation.ai/post/nsql-llama-2-7b, I found "Open Source Ours - Pretrain + Instruct". Could anyone explain what exactly is this "Instruct"? Only after this "Instruct", the accuracy can be close to chatgpt.

In the page https://www.numbersstation.ai/post/nsql-llama-2-7b, I found "Open Source Ours - Pretrain + Instruct". Could anyone explain what exactly is this "Instruct"? Only after this "Instruct", the accuracy can be close to chatgpt.

What they mean by pre-training is training it for Causal Language Modeling with SQL queries only. Instruct refers to the fine-tuning of the pretrained LLAMA model with text-to-SQL datasets, also known as instruction tuning.

NumbersStation org

Do you mind to share the evaluation script or the prompt template for us to duplicate the numbers? Many thanks

We've followed https://github.com/taoyds/spider for spider and geoquery evaluation and use Rajkumar prompt format. You can find examples here: https://github.com/NumbersStationAI/NSQL/tree/main/examples

Sign up or log in to comment