terry69 commited on
Commit
9c15355
1 Parent(s): 34a1d07

Training in progress, step 300

Browse files
adapter_config.json CHANGED
@@ -19,11 +19,11 @@
19
  "rank_pattern": {},
20
  "revision": null,
21
  "target_modules": [
22
- "gate_proj",
23
- "k_proj",
24
  "up_proj",
 
25
  "o_proj",
26
  "down_proj",
 
27
  "v_proj",
28
  "q_proj"
29
  ],
 
19
  "rank_pattern": {},
20
  "revision": null,
21
  "target_modules": [
 
 
22
  "up_proj",
23
+ "k_proj",
24
  "o_proj",
25
  "down_proj",
26
+ "gate_proj",
27
  "v_proj",
28
  "q_proj"
29
  ],
adapter_model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:55b6d63204976919dde9f61d2e6e029bcef4ac019ca65c4c995936e92a6a0db1
3
  size 30042168
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:6bb858c266b17c71c2ceab7e42fea15dacc9d8a28a09b82c3a5d8e5655ac2a1b
3
  size 30042168
runs/Apr30_19-43-07_ip-172-31-69-60.ec2.internal/events.out.tfevents.1714506314.ip-172-31-69-60.ec2.internal.22715.0 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:9c01256c332f327e2ed1b887f3feda1e7a44be48cde49c8f11e3a379b3bbf62d
3
+ size 17492
special_tokens_map.json CHANGED
@@ -1,23 +1,19 @@
1
  {
2
- "additional_special_tokens": [
3
- {
4
- "content": "<|im_start|>",
5
- "lstrip": false,
6
- "normalized": false,
7
- "rstrip": false,
8
- "single_word": false
9
- },
10
- {
11
- "content": "<|im_end|>",
12
- "lstrip": false,
13
- "normalized": false,
14
- "rstrip": false,
15
- "single_word": false
16
- }
17
- ],
18
- "bos_token": "<|im_start|>",
19
- "eos_token": "<|im_end|>",
20
- "pad_token": "<|im_end|>",
21
  "unk_token": {
22
  "content": "<unk>",
23
  "lstrip": false,
 
1
  {
2
+ "bos_token": {
3
+ "content": "<s>",
4
+ "lstrip": false,
5
+ "normalized": false,
6
+ "rstrip": false,
7
+ "single_word": false
8
+ },
9
+ "eos_token": {
10
+ "content": "</s>",
11
+ "lstrip": false,
12
+ "normalized": false,
13
+ "rstrip": false,
14
+ "single_word": false
15
+ },
16
+ "pad_token": "</s>",
 
 
 
 
17
  "unk_token": {
18
  "content": "<unk>",
19
  "lstrip": false,
tokenizer.json CHANGED
@@ -34,24 +34,6 @@
34
  "rstrip": false,
35
  "normalized": false,
36
  "special": true
37
- },
38
- {
39
- "id": 32000,
40
- "content": "<|im_start|>",
41
- "single_word": false,
42
- "lstrip": false,
43
- "rstrip": false,
44
- "normalized": false,
45
- "special": true
46
- },
47
- {
48
- "id": 32001,
49
- "content": "<|im_end|>",
50
- "single_word": false,
51
- "lstrip": false,
52
- "rstrip": false,
53
- "normalized": false,
54
- "special": true
55
  }
56
  ],
57
  "normalizer": {
 
34
  "rstrip": false,
35
  "normalized": false,
36
  "special": true
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
37
  }
38
  ],
39
  "normalizer": {
tokenizer_config.json CHANGED
@@ -25,35 +25,15 @@
25
  "rstrip": false,
26
  "single_word": false,
27
  "special": true
28
- },
29
- "32000": {
30
- "content": "<|im_start|>",
31
- "lstrip": false,
32
- "normalized": false,
33
- "rstrip": false,
34
- "single_word": false,
35
- "special": true
36
- },
37
- "32001": {
38
- "content": "<|im_end|>",
39
- "lstrip": false,
40
- "normalized": false,
41
- "rstrip": false,
42
- "single_word": false,
43
- "special": true
44
  }
45
  },
46
- "additional_special_tokens": [
47
- "<|im_start|>",
48
- "<|im_end|>"
49
- ],
50
- "bos_token": "<|im_start|>",
51
- "chat_template": "{% for message in messages %}{{'<|im_start|>' + message['role'] + '\n' + message['content'] + '<|im_end|>' + '\n'}}{% endfor %}{% if add_generation_prompt %}{{ '<|im_start|>assistant\n' }}{% endif %}",
52
  "clean_up_tokenization_spaces": false,
53
- "eos_token": "<|im_end|>",
54
  "legacy": false,
55
  "model_max_length": 2048,
56
- "pad_token": "<|im_end|>",
57
  "padding_side": "right",
58
  "sp_model_kwargs": {},
59
  "tokenizer_class": "LlamaTokenizer",
 
25
  "rstrip": false,
26
  "single_word": false,
27
  "special": true
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
28
  }
29
  },
30
+ "bos_token": "<s>",
31
+ "chat_template": "{% for message in messages %}\n{% if message['role'] == 'user' %}\n{{ '<|user|>\n' + message['content'] + eos_token }}\n{% elif message['role'] == 'system' %}\n{{ '<|system|>\n' + message['content'] + eos_token }}\n{% elif message['role'] == 'assistant' %}\n{{ '<|assistant|>\n' + message['content'] + eos_token }}\n{% endif %}\n{% if loop.last and add_generation_prompt %}\n{{ '<|assistant|>' }}\n{% endif %}\n{% endfor %}",
 
 
 
 
32
  "clean_up_tokenization_spaces": false,
33
+ "eos_token": "</s>",
34
  "legacy": false,
35
  "model_max_length": 2048,
36
+ "pad_token": "</s>",
37
  "padding_side": "right",
38
  "sp_model_kwargs": {},
39
  "tokenizer_class": "LlamaTokenizer",
training_args.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:417b859fbc6e5d035471956059ac8a1372a207b186b72a448ab076d28df4c179
3
  size 6072
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:09c3c5667603993f97fa1984b34612d1a0e5aa7365b49e0b2f248ff97a9fef5b
3
  size 6072