IntellectusAI commited on
Commit
b20033d
1 Parent(s): 9f99449

Training in progress, epoch 1

Browse files
adapter_config.json CHANGED
@@ -21,7 +21,10 @@
21
  "revision": null,
22
  "target_modules": [
23
  "q_proj",
24
- "v_proj"
 
 
 
25
  ],
26
  "task_type": "CAUSAL_LM",
27
  "use_dora": false,
 
21
  "revision": null,
22
  "target_modules": [
23
  "q_proj",
24
+ "v_proj",
25
+ "gate_proj",
26
+ "o_proj",
27
+ "k_proj"
28
  ],
29
  "task_type": "CAUSAL_LM",
30
  "use_dora": false,
adapter_model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:094a9f72ea61eb1f2062873c22d53bfa2426f10fcdad6f9dc3234ac3ebd76a18
3
- size 27280152
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:d13cb9f4a0ec2a88a34f262c30263706a54a2d3da18e4aea32fbd511cdcc2ff3
3
+ size 92317600
runs/May28_11-21-19_e6addf98c5a3/events.out.tfevents.1716895393.e6addf98c5a3.572.0 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:134ae7e104129c5ad9a1f454b75c7e765e067b782ccdbfe4944f53e6f78ada28
3
+ size 4184
special_tokens_map.json CHANGED
@@ -1,9 +1,4 @@
1
  {
2
- "additional_special_tokens": [
3
- "<unk>",
4
- "<s>",
5
- "</s>"
6
- ],
7
  "bos_token": {
8
  "content": "<s>",
9
  "lstrip": false,
 
1
  {
 
 
 
 
 
2
  "bos_token": {
3
  "content": "<s>",
4
  "lstrip": false,
tokenizer.json CHANGED
@@ -1,7 +1,7 @@
1
  {
2
  "version": "1.0",
3
  "truncation": {
4
- "direction": "Left",
5
  "max_length": 1024,
6
  "strategy": "LongestFirst",
7
  "stride": 0
@@ -67,6 +67,12 @@
67
  "id": "A",
68
  "type_id": 0
69
  }
 
 
 
 
 
 
70
  }
71
  ],
72
  "pair": [
@@ -82,6 +88,12 @@
82
  "type_id": 0
83
  }
84
  },
 
 
 
 
 
 
85
  {
86
  "SpecialToken": {
87
  "id": "<s>",
@@ -93,9 +105,24 @@
93
  "id": "B",
94
  "type_id": 1
95
  }
 
 
 
 
 
 
96
  }
97
  ],
98
  "special_tokens": {
 
 
 
 
 
 
 
 
 
99
  "<s>": {
100
  "id": "<s>",
101
  "ids": [
 
1
  {
2
  "version": "1.0",
3
  "truncation": {
4
+ "direction": "Right",
5
  "max_length": 1024,
6
  "strategy": "LongestFirst",
7
  "stride": 0
 
67
  "id": "A",
68
  "type_id": 0
69
  }
70
+ },
71
+ {
72
+ "SpecialToken": {
73
+ "id": "</s>",
74
+ "type_id": 0
75
+ }
76
  }
77
  ],
78
  "pair": [
 
88
  "type_id": 0
89
  }
90
  },
91
+ {
92
+ "SpecialToken": {
93
+ "id": "</s>",
94
+ "type_id": 0
95
+ }
96
+ },
97
  {
98
  "SpecialToken": {
99
  "id": "<s>",
 
105
  "id": "B",
106
  "type_id": 1
107
  }
108
+ },
109
+ {
110
+ "SpecialToken": {
111
+ "id": "</s>",
112
+ "type_id": 1
113
+ }
114
  }
115
  ],
116
  "special_tokens": {
117
+ "</s>": {
118
+ "id": "</s>",
119
+ "ids": [
120
+ 2
121
+ ],
122
+ "tokens": [
123
+ "</s>"
124
+ ]
125
+ },
126
  "<s>": {
127
  "id": "<s>",
128
  "ids": [
tokenizer_config.json CHANGED
@@ -1,6 +1,7 @@
1
  {
2
  "add_bos_token": true,
3
- "add_eos_token": false,
 
4
  "added_tokens_decoder": {
5
  "0": {
6
  "content": "<unk>",
@@ -27,21 +28,18 @@
27
  "special": true
28
  }
29
  },
30
- "additional_special_tokens": [
31
- "<unk>",
32
- "<s>",
33
- "</s>"
34
- ],
35
  "bos_token": "<s>",
 
36
  "clean_up_tokenization_spaces": false,
37
  "eos_token": "</s>",
38
  "legacy": true,
39
- "model_max_length": 1000000000000000019884624838656,
40
  "pad_token": "</s>",
 
41
  "sp_model_kwargs": {},
42
  "spaces_between_special_tokens": false,
43
  "tokenizer_class": "LlamaTokenizer",
44
- "truncation_side": "left",
45
  "unk_token": "<unk>",
46
  "use_default_system_prompt": true
47
  }
 
1
  {
2
  "add_bos_token": true,
3
+ "add_eos_token": true,
4
+ "add_prefix_space": null,
5
  "added_tokens_decoder": {
6
  "0": {
7
  "content": "<unk>",
 
28
  "special": true
29
  }
30
  },
31
+ "additional_special_tokens": [],
 
 
 
 
32
  "bos_token": "<s>",
33
+ "chat_template": "{{ bos_token }}{% for message in messages %}{% if (message['role'] == 'user') != (loop.index0 % 2 == 0) %}{{ raise_exception('Conversation roles must alternate user/assistant/user/assistant/...') }}{% endif %}{% if message['role'] == 'user' %}{{ '[INST] ' + message['content'] + ' [/INST]' }}{% elif message['role'] == 'assistant' %}{{ message['content'] + eos_token + ' ' }}{% else %}{{ raise_exception('Only user and assistant roles are supported!') }}{% endif %}{% endfor %}",
34
  "clean_up_tokenization_spaces": false,
35
  "eos_token": "</s>",
36
  "legacy": true,
37
+ "model_max_length": 512,
38
  "pad_token": "</s>",
39
+ "padding_side": "right",
40
  "sp_model_kwargs": {},
41
  "spaces_between_special_tokens": false,
42
  "tokenizer_class": "LlamaTokenizer",
 
43
  "unk_token": "<unk>",
44
  "use_default_system_prompt": true
45
  }
training_args.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:8f20da8ec14c267bfbcd89d50c443481e66dd3e009b8785cfc1eb43ea62f931c
3
- size 4984
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:81d2e914bbda72e796fb4e27d0c5854686f9898d29b24d616c5b1d6c9e08aeeb
3
+ size 5112