IvoHoese commited on
Commit
f12f7aa
·
verified ·
1 Parent(s): c5d9d45

Upload folder using huggingface_hub

Browse files
models/pythia-prefix/adapter_config.json ADDED
@@ -0,0 +1,22 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "auto_mapping": null,
3
+ "base_model_name_or_path": "EleutherAI/pythia-410m",
4
+ "encoder_hidden_size": 1024,
5
+ "inference_mode": true,
6
+ "modules_to_save": [
7
+ "classifier",
8
+ "score",
9
+ "classifier",
10
+ "score"
11
+ ],
12
+ "num_attention_heads": 16,
13
+ "num_layers": 24,
14
+ "num_transformer_submodules": 1,
15
+ "num_virtual_tokens": 10,
16
+ "peft_type": "PREFIX_TUNING",
17
+ "peft_version": "0.18.1",
18
+ "prefix_projection": false,
19
+ "revision": null,
20
+ "task_type": "SEQ_CLS",
21
+ "token_dim": 1024
22
+ }
models/pythia-prefix/adapter_model.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:3fc13b2a1f0b83df9e2e58606f7f79ba3c0e25a13835f2ad1a2e1bc81e7ee38d
3
+ size 1974480
models/pythia-prefix/tokenizer.json ADDED
The diff for this file is too large to render. See raw diff
 
models/pythia-prefix/tokenizer_config.json ADDED
@@ -0,0 +1,13 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "add_prefix_space": false,
3
+ "backend": "tokenizers",
4
+ "bos_token": "<|endoftext|>",
5
+ "eos_token": "<|endoftext|>",
6
+ "errors": "replace",
7
+ "is_local": false,
8
+ "model_max_length": 1000000000000000019884624838656,
9
+ "pad_token": "<|endoftext|>",
10
+ "tokenizer_class": "GPTNeoXTokenizer",
11
+ "trim_offsets": true,
12
+ "unk_token": "<|endoftext|>"
13
+ }