smangrul commited on
Commit
2d2e326
1 Parent(s): 93827b6

Training in progress, epoch 0

Browse files
adapter_config.json CHANGED
@@ -19,13 +19,13 @@
19
  "rank_pattern": {},
20
  "revision": null,
21
  "target_modules": [
 
 
22
  "lm_head",
 
23
  "k_proj",
24
- "gate_proj",
25
- "down_proj",
26
  "embed_tokens",
27
- "o_proj",
28
- "q_proj",
29
  "up_proj",
30
  "v_proj"
31
  ],
 
19
  "rank_pattern": {},
20
  "revision": null,
21
  "target_modules": [
22
+ "gate_proj",
23
+ "q_proj",
24
  "lm_head",
25
+ "o_proj",
26
  "k_proj",
 
 
27
  "embed_tokens",
28
+ "down_proj",
 
29
  "up_proj",
30
  "v_proj"
31
  ],
adapter_model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:4bde66cc63d4facaec61009347f5f7e3d7675a398c8a39661d69aa5c5702159c
3
  size 1658577808
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:7dc74c7bef328e78eda345e27167817561eb07da606e6e03141318fc499a2cdc
3
  size 1658577808
runs/Dec22_05-07-57_38ec33681dde/events.out.tfevents.1703221756.38ec33681dde.4322.0 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:22b60fc4100c3097271ce9b612585f20157f3b04a762b3ec2ed1ff3d2a569aef
3
+ size 9935
special_tokens_map.json CHANGED
@@ -1,4 +1,12 @@
1
  {
 
 
 
 
 
 
 
 
2
  "bos_token": {
3
  "content": "<s>",
4
  "lstrip": false,
@@ -6,14 +14,8 @@
6
  "rstrip": false,
7
  "single_word": false
8
  },
9
- "eos_token": {
10
- "content": "</s>",
11
- "lstrip": false,
12
- "normalized": false,
13
- "rstrip": false,
14
- "single_word": false
15
- },
16
- "pad_token": "[PAD]",
17
  "unk_token": {
18
  "content": "<unk>",
19
  "lstrip": false,
 
1
  {
2
+ "additional_special_tokens": [
3
+ "<|im_start|>user",
4
+ "<|im_start|>assistant",
5
+ "<|im_start|>system",
6
+ "<|im_end|>",
7
+ "<s>",
8
+ "<pad>"
9
+ ],
10
  "bos_token": {
11
  "content": "<s>",
12
  "lstrip": false,
 
14
  "rstrip": false,
15
  "single_word": false
16
  },
17
+ "eos_token": "<|im_end|>",
18
+ "pad_token": "<pad>",
 
 
 
 
 
 
19
  "unk_token": {
20
  "content": "<unk>",
21
  "lstrip": false,
tokenizer.json CHANGED
@@ -31,8 +31,44 @@
31
  "special": true
32
  },
33
  {
34
- "id": 32000,
35
- "content": "[PAD]",
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
36
  "single_word": false,
37
  "lstrip": false,
38
  "rstrip": false,
 
31
  "special": true
32
  },
33
  {
34
+ "id": 48064,
35
+ "content": "<|im_end|>",
36
+ "single_word": false,
37
+ "lstrip": false,
38
+ "rstrip": false,
39
+ "normalized": false,
40
+ "special": true
41
+ },
42
+ {
43
+ "id": 48065,
44
+ "content": "<pad>",
45
+ "single_word": false,
46
+ "lstrip": false,
47
+ "rstrip": false,
48
+ "normalized": false,
49
+ "special": true
50
+ },
51
+ {
52
+ "id": 48066,
53
+ "content": "<|im_start|>user",
54
+ "single_word": false,
55
+ "lstrip": false,
56
+ "rstrip": false,
57
+ "normalized": false,
58
+ "special": true
59
+ },
60
+ {
61
+ "id": 48067,
62
+ "content": "<|im_start|>assistant",
63
+ "single_word": false,
64
+ "lstrip": false,
65
+ "rstrip": false,
66
+ "normalized": false,
67
+ "special": true
68
+ },
69
+ {
70
+ "id": 48068,
71
+ "content": "<|im_start|>system",
72
  "single_word": false,
73
  "lstrip": false,
74
  "rstrip": false,
tokenizer_config.json CHANGED
@@ -26,8 +26,40 @@
26
  "single_word": false,
27
  "special": true
28
  },
29
- "32000": {
30
- "content": "[PAD]",
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
31
  "lstrip": false,
32
  "normalized": false,
33
  "rstrip": false,
@@ -35,12 +67,21 @@
35
  "special": true
36
  }
37
  },
 
 
 
 
 
 
 
 
38
  "bos_token": "<s>",
 
39
  "clean_up_tokenization_spaces": false,
40
- "eos_token": "</s>",
41
  "legacy": false,
42
  "model_max_length": 1000000000000000019884624838656,
43
- "pad_token": "[PAD]",
44
  "sp_model_kwargs": {},
45
  "spaces_between_special_tokens": false,
46
  "tokenizer_class": "LlamaTokenizer",
 
26
  "single_word": false,
27
  "special": true
28
  },
29
+ "48064": {
30
+ "content": "<|im_end|>",
31
+ "lstrip": false,
32
+ "normalized": false,
33
+ "rstrip": false,
34
+ "single_word": false,
35
+ "special": true
36
+ },
37
+ "48065": {
38
+ "content": "<pad>",
39
+ "lstrip": false,
40
+ "normalized": false,
41
+ "rstrip": false,
42
+ "single_word": false,
43
+ "special": true
44
+ },
45
+ "48066": {
46
+ "content": "<|im_start|>user",
47
+ "lstrip": false,
48
+ "normalized": false,
49
+ "rstrip": false,
50
+ "single_word": false,
51
+ "special": true
52
+ },
53
+ "48067": {
54
+ "content": "<|im_start|>assistant",
55
+ "lstrip": false,
56
+ "normalized": false,
57
+ "rstrip": false,
58
+ "single_word": false,
59
+ "special": true
60
+ },
61
+ "48068": {
62
+ "content": "<|im_start|>system",
63
  "lstrip": false,
64
  "normalized": false,
65
  "rstrip": false,
 
67
  "special": true
68
  }
69
  },
70
+ "additional_special_tokens": [
71
+ "<|im_start|>user",
72
+ "<|im_start|>assistant",
73
+ "<|im_start|>system",
74
+ "<|im_end|>",
75
+ "<s>",
76
+ "<pad>"
77
+ ],
78
  "bos_token": "<s>",
79
+ "chat_template": "{% for message in messages %}\n{{'<|im_start|>' + message['role'] + '\n' + message['content'] + '<|im_end|>' + '\n'}}{% if loop.last and add_generation_prompt %}{{'<|im_start|>assistant\n' }}{% endif %}{% endfor %}",
80
  "clean_up_tokenization_spaces": false,
81
+ "eos_token": "<|im_end|>",
82
  "legacy": false,
83
  "model_max_length": 1000000000000000019884624838656,
84
+ "pad_token": "<pad>",
85
  "sp_model_kwargs": {},
86
  "spaces_between_special_tokens": false,
87
  "tokenizer_class": "LlamaTokenizer",
training_args.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:6b83ed20e35726c09cba0b5807be094d003fa7033758dee750ce3cb4d823b008
3
  size 4792
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:911acb062982bd0f8d737bfcb4ff9ee94c5d1736ce19eba2644b6092af3af10a
3
  size 4792