levmckinney norabelrose commited on
Commit
6a5215d
1 Parent(s): dcf9cd2

- vicuna original (1f8e57581c432d4aa2096f33503f85c1f52648fe)
- remove vicuna-13b (4c124f188d46710922e36f48b1f5f5e1d4f8b7ae)
- reverted config from pointer (dc58cfe4667b747d404848ce8c2ec052be6fad73)


Co-authored-by: Nora Belrose <norabelrose@users.noreply.huggingface.co>

lens/facebook/llama-30b/config.json CHANGED
@@ -1 +1 @@
1
- {"base_model_name_or_path": "huggyllama/llama-30b", "d_model": 6656, "num_hidden_layers": 60, "bias": true, "base_model_revision": null, "unemebd_hash": "512b41bb8fc0e2b3647a877b5e114ca9503d7c800185b14a8e7ad9e921424367", "lens_type": "linear_tuned_lens"}
 
1
+ {"base_model_name_or_path": "huggyllama/llama-30b", "d_model": 6656, "num_hidden_layers": 60, "bias": true, "base_model_revision": null, "unemebd_hash": "512b41bb8fc0e2b3647a877b5e114ca9503d7c800185b14a8e7ad9e921424367", "lens_type": "linear_tuned_lens"}
lens/lymsys/vicuna-13b-v1.1/config.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"base_model_name_or_path": "vicuna-original-13b", "d_model": 5120, "num_hidden_layers": 40, "bias": true, "base_model_revision": null, "unemebd_hash": "86cefdfd94bb3da225b405dd1328136786f6177b03d82f90e5c734d23c47e8ca", "lens_type": "linear_tuned_lens"}
lens/lymsys/vicuna-13b-v1.1/params.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:0281f46b65b728cb390e41bd5848d8ab635b2847b48c080b2de2b78627baa40d
3
+ size 2097581027