Menouar commited on
Commit
168fe9a
1 Parent(s): e95a12b

Training in progress, epoch 1

Browse files
adapter_config.json CHANGED
@@ -19,8 +19,8 @@
19
  "rank_pattern": {},
20
  "revision": null,
21
  "target_modules": [
22
- "query_key_value",
23
  "dense_4h_to_h",
 
24
  "dense",
25
  "dense_h_to_4h"
26
  ],
 
19
  "rank_pattern": {},
20
  "revision": null,
21
  "target_modules": [
 
22
  "dense_4h_to_h",
23
+ "query_key_value",
24
  "dense",
25
  "dense_h_to_4h"
26
  ],
adapter_model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:04f5c57ac4a62e671f8d752a6632adce6f04dd0253b35faa4069f637948b0fe6
3
- size 2226331968
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:d044aeca9390c2caf2257e39fe8fbdeff305b80cd0b90b47cc3eb54c80642427
3
+ size 1044418808
runs/Jan25_21-15-01_f75ef5543b30/events.out.tfevents.1706217309.f75ef5543b30.1559.0 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:fa76445fe18fa7b4eefd1465076d497c9ea88edd68a1df5fec976b8fee90fc36
3
+ size 5624
special_tokens_map.json CHANGED
@@ -1,21 +1,23 @@
1
  {
2
  "additional_special_tokens": [
3
- {
4
- "content": "<|im_start|>",
5
- "lstrip": false,
6
- "normalized": false,
7
- "rstrip": false,
8
- "single_word": false
9
- },
10
- {
11
- "content": "<|im_end|>",
12
- "lstrip": false,
13
- "normalized": false,
14
- "rstrip": false,
15
- "single_word": false
16
- }
17
  ],
18
- "bos_token": "<|im_start|>",
19
- "eos_token": "<|im_end|>",
20
- "pad_token": "<|im_end|>"
 
 
 
 
 
21
  }
 
1
  {
2
  "additional_special_tokens": [
3
+ ">>TITLE<<",
4
+ ">>ABSTRACT<<",
5
+ ">>INTRODUCTION<<",
6
+ ">>SUMMARY<<",
7
+ ">>COMMENT<<",
8
+ ">>ANSWER<<",
9
+ ">>QUESTION<<",
10
+ ">>DOMAIN<<",
11
+ ">>PREFIX<<",
12
+ ">>SUFFIX<<",
13
+ ">>MIDDLE<<"
 
 
 
14
  ],
15
+ "eos_token": {
16
+ "content": "<|endoftext|>",
17
+ "lstrip": false,
18
+ "normalized": false,
19
+ "rstrip": false,
20
+ "single_word": false
21
+ },
22
+ "pad_token": "<|endoftext|>"
23
  }
tokenizer.json CHANGED
@@ -110,24 +110,6 @@
110
  "rstrip": false,
111
  "normalized": false,
112
  "special": true
113
- },
114
- {
115
- "id": 65024,
116
- "content": "<|im_start|>",
117
- "single_word": false,
118
- "lstrip": false,
119
- "rstrip": false,
120
- "normalized": false,
121
- "special": true
122
- },
123
- {
124
- "id": 65025,
125
- "content": "<|im_end|>",
126
- "single_word": false,
127
- "lstrip": false,
128
- "rstrip": false,
129
- "normalized": false,
130
- "special": true
131
  }
132
  ],
133
  "normalizer": null,
 
110
  "rstrip": false,
111
  "normalized": false,
112
  "special": true
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
113
  }
114
  ],
115
  "normalizer": null,
tokenizer_config.json CHANGED
@@ -96,37 +96,28 @@
96
  "rstrip": false,
97
  "single_word": false,
98
  "special": true
99
- },
100
- "65024": {
101
- "content": "<|im_start|>",
102
- "lstrip": false,
103
- "normalized": false,
104
- "rstrip": false,
105
- "single_word": false,
106
- "special": true
107
- },
108
- "65025": {
109
- "content": "<|im_end|>",
110
- "lstrip": false,
111
- "normalized": false,
112
- "rstrip": false,
113
- "single_word": false,
114
- "special": true
115
  }
116
  },
117
  "additional_special_tokens": [
118
- "<|im_start|>",
119
- "<|im_end|>"
 
 
 
 
 
 
 
 
 
120
  ],
121
- "bos_token": "<|im_start|>",
122
- "chat_template": "{% for message in messages %}{{'<|im_start|>' + message['role'] + '\n' + message['content'] + '<|im_end|>' + '\n'}}{% endfor %}{% if add_generation_prompt %}{{ '<|im_start|>assistant\n' }}{% endif %}",
123
  "clean_up_tokenization_spaces": true,
124
- "eos_token": "<|im_end|>",
125
  "model_input_names": [
126
  "input_ids",
127
  "attention_mask"
128
  ],
129
  "model_max_length": 2048,
130
- "pad_token": "<|im_end|>",
131
  "tokenizer_class": "PreTrainedTokenizerFast"
132
  }
 
96
  "rstrip": false,
97
  "single_word": false,
98
  "special": true
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
99
  }
100
  },
101
  "additional_special_tokens": [
102
+ ">>TITLE<<",
103
+ ">>ABSTRACT<<",
104
+ ">>INTRODUCTION<<",
105
+ ">>SUMMARY<<",
106
+ ">>COMMENT<<",
107
+ ">>ANSWER<<",
108
+ ">>QUESTION<<",
109
+ ">>DOMAIN<<",
110
+ ">>PREFIX<<",
111
+ ">>SUFFIX<<",
112
+ ">>MIDDLE<<"
113
  ],
 
 
114
  "clean_up_tokenization_spaces": true,
115
+ "eos_token": "<|endoftext|>",
116
  "model_input_names": [
117
  "input_ids",
118
  "attention_mask"
119
  ],
120
  "model_max_length": 2048,
121
+ "pad_token": "<|endoftext|>",
122
  "tokenizer_class": "PreTrainedTokenizerFast"
123
  }
training_args.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:555ef360b85bef941bcf53a7de4604001e2b05f0d87204a6465b8b895e897547
3
  size 4792
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:5c245a9dc58c7484be3dee58f048a567636a81d806c4f68d158ac8d0a1ff8ce8
3
  size 4792