onebitquantized commited on
Commit
8ad40ea
1 Parent(s): 682cb92

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +15 -0
README.md CHANGED
@@ -63,5 +63,20 @@ model = AutoGPTQForCausalLM.from_quantized(
63
  outputs = model.generate(**inputs, do_sample=True, max_new_tokens=1024)
64
  print(tokenizer.batch_decode(outputs, skip_special_tokens=True))
65
  ```
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
66
  # Contact Us
67
  For additional xMADified models, access to fine-tuning, and general questions, please contact us at support@xmad.ai and join our waiting list.
 
63
  outputs = model.generate(**inputs, do_sample=True, max_new_tokens=1024)
64
  print(tokenizer.batch_decode(outputs, skip_special_tokens=True))
65
  ```
66
+
67
+ # Citation
68
+
69
+ If you found this model useful, please cite our research paper.
70
+
71
+ ```
72
+ @article{zhang2024leanquant,
73
+ title={LeanQuant: Accurate and Scalable Large Language Model Quantization with Loss-error-aware Grid},
74
+ author={Zhang, Tianyi and Shrivastava, Anshumali},
75
+ journal={arXiv preprint arXiv:2407.10032},
76
+ year={2024},
77
+ url={https://arxiv.org/abs/2407.10032},
78
+ }
79
+ ```
80
+
81
  # Contact Us
82
  For additional xMADified models, access to fine-tuning, and general questions, please contact us at support@xmad.ai and join our waiting list.