jyhong836 commited on
Commit
83f93fd
1 Parent(s): 02a238d

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +20 -4
README.md CHANGED
@@ -5,13 +5,14 @@ colorFrom: pink
5
  colorTo: indigo
6
  sdk: static
7
  pinned: false
 
 
8
  ---
9
- # Compressed LLM Model Zone
10
-
11
 
 
12
  The models are prepared by [Visual Informatics Group @ University of Texas at Austin (VITA-group)](https://vita-group.github.io/) and
13
  [Center for Applied Scientific Computing](https://computing.llnl.gov/casc) at [LLNL](https://www.llnl.gov/).
14
- Credits to Ajay Jaiswal, Jinhao Duan, Zhenyu Zhang, Zhangheng Li, Lu Yin, Shiwei Liu and Junyuan Hong.
15
 
16
  License: [MIT License](https://opensource.org/license/mit/)
17
 
@@ -96,6 +97,18 @@ tokenizer.decode(outputs[0])
96
 
97
  If you are using models in this hub, please consider citing our papers.
98
  ```bibtex
 
 
 
 
 
 
 
 
 
 
 
 
99
  @article{jaiswal2023emergence,
100
  title={The Emergence of Essential Sparsity in Large Pre-trained Models: The Weights that Matter},
101
  author={Jaiswal, Ajay and Liu, Shiwei and Chen, Tianlong and Wang, Zhangyang},
@@ -110,6 +123,9 @@ If you are using models in this hub, please consider citing our papers.
110
  }
111
  ```
112
 
 
 
 
113
 
114
- For any question, please contact [Junyuan Hong](mailto:jyhong@utexas.edu).
115
 
 
 
5
  colorTo: indigo
6
  sdk: static
7
  pinned: false
8
+ license: mit
9
+ short_description: Compressed Large Language Models
10
  ---
11
+ # Compressed Large Language Models
 
12
 
13
+ This repo contains compressed LLMs used in the [Decoding Compressed Trust](https://decoding-comp-trust.github.io/) project.
14
  The models are prepared by [Visual Informatics Group @ University of Texas at Austin (VITA-group)](https://vita-group.github.io/) and
15
  [Center for Applied Scientific Computing](https://computing.llnl.gov/casc) at [LLNL](https://www.llnl.gov/).
 
16
 
17
  License: [MIT License](https://opensource.org/license/mit/)
18
 
 
97
 
98
  If you are using models in this hub, please consider citing our papers.
99
  ```bibtex
100
+ @article{hong2024comptrust,
101
+ title={Decoding Compressed Trust: Scrutinizing the Trustworthiness of Efficient LLMs Under Compression},
102
+ author={Hong, Junyuan and Duan, Jinhao and Zhang, Chenhui and Li, Zhangheng
103
+ and Xie, Chulin and Lieberman, Kelsey and Diffenderfer, James
104
+ and Bartoldson, Brian and Jaiswal, Ajay and Xu, Kaidi and Kailkhura, Bhavya
105
+ and Hendrycks, Dan and Song, Dawn and Wang, Zhangyang and Bo Li},
106
+ journal={arXiv},
107
+ year={2024}
108
+ }
109
+ ```
110
+ Some of the models were used in previous publications.
111
+ ```bibtex
112
  @article{jaiswal2023emergence,
113
  title={The Emergence of Essential Sparsity in Large Pre-trained Models: The Weights that Matter},
114
  author={Jaiswal, Ajay and Liu, Shiwei and Chen, Tianlong and Wang, Zhangyang},
 
123
  }
124
  ```
125
 
126
+ ## Acknowlegement
127
+
128
+ Main credits to Ajay Jaiswal, Jinhao Duan, Zhangheng Li and Junyuan Hong. We also appreciate Zhenyu Zhang, Lu Yin, and Shiwei Liu in some preparations.
129
 
 
130
 
131
+ For any question, please contact [Junyuan Hong](mailto:jyhong@utexas.edu).