richardr1126 commited on
Commit
9469cf6
1 Parent(s): a160f26

Create README.md

Browse files
Files changed (1) hide show
  1. README.md +59 -0
README.md ADDED
@@ -0,0 +1,59 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ ---
2
+ tags:
3
+ - sql
4
+ - spider
5
+ - text-to-sql
6
+ - sql finetune
7
+ - 8bit
8
+ datasets:
9
+ - spider
10
+ - richardr1126/spider-skeleton-context-instruct
11
+ library_name: transformers
12
+ license: bigcode-openrail-m
13
+ ---
14
+ ### Spider Skeleton Wizard Coder 8bit Summary
15
+
16
+ - This model was created by finetuning [WizardLM/WizardCoder-15B-V1.0](https://huggingface.co/WizardLM/WizardCoder-15B-V1.0) on an enhanced Spider context training dataset: [richardr1126/spider-skeleton-context-instruct](https://huggingface.co/datasets/richardr1126/spider-skeleton-context-instruct).
17
+ - Finetuning was performed using QLoRa on 3x RTX6000 48GB.
18
+ - This is the bitsandbytes 8-bit version of the model. It needs to be loaded onto a GPU for it to work.
19
+ - If you want just the QLoRa/LoRA adapter: [richardr1126/spider-skeleton-wizard-coder-qlora](https://huggingface.co/richardr1126/spider-skeleton-wizard-coder-qlora)
20
+
21
+ ### Running the GGML model
22
+ - The best way to run this model is to use the [4-bit GGML version](https://huggingface.co/richardr1126/spider-skeleton-wizard-coder-ggml) on [koboldcpp](https://github.com/LostRuins/koboldcpp), with CuBlas support.
23
+
24
+ ### Spider Dataset
25
+
26
+ [Spider](https://arxiv.org/abs/1809.08887) is a large-scale complex and cross-domain semantic parsing and text-to-SQL dataset annotated by 11 Yale students
27
+ The goal of the Spider challenge is to develop natural language interfaces to cross-domain databases.
28
+
29
+ This dataset was used to finetune this model.
30
+
31
+ ## Citations
32
+
33
+ ```
34
+ @misc{luo2023wizardcoder,
35
+ title={WizardCoder: Empowering Code Large Language Models with Evol-Instruct},
36
+ author={Ziyang Luo and Can Xu and Pu Zhao and Qingfeng Sun and Xiubo Geng and Wenxiang Hu and Chongyang Tao and Jing Ma and Qingwei Lin and Daxin Jiang},
37
+ year={2023},
38
+ }
39
+ ```
40
+ ```
41
+ @article{yu2018spider,
42
+ title={Spider: A large-scale human-labeled dataset for complex and cross-domain semantic parsing and text-to-sql task},
43
+ author={Yu, Tao and Zhang, Rui and Yang, Kai and Yasunaga, Michihiro and Wang, Dongxu and Li, Zifan and Ma, James and Li, Irene and Yao, Qingning and Roman, Shanelle and others},
44
+ journal={arXiv preprint arXiv:1809.08887},
45
+ year={2018}
46
+ }
47
+ ```
48
+ ```
49
+ @article{dettmers2023qlora,
50
+ title={QLoRA: Efficient Finetuning of Quantized LLMs},
51
+ author={Dettmers, Tim and Pagnoni, Artidoro and Holtzman, Ari and Zettlemoyer, Luke},
52
+ journal={arXiv preprint arXiv:2305.14314},
53
+ year={2023}
54
+ }
55
+ ```
56
+
57
+ ## Disclaimer
58
+
59
+ The resources, including code, data, and model weights, associated with this project are restricted for academic research purposes only and cannot be used for commercial purposes. The content produced by any version of WizardCoder is influenced by uncontrollable variables such as randomness, and therefore, the accuracy of the output cannot be guaranteed by this project. This project does not accept any legal liability for the content of the model output, nor does it assume responsibility for any losses incurred due to the use of associated resources and output results.