kangdawei commited on
Commit
1756695
·
verified ·
1 Parent(s): 9bd5f70

Training in progress, step 110

Browse files
adapter_config.json CHANGED
@@ -29,13 +29,13 @@
29
  "rank_pattern": {},
30
  "revision": null,
31
  "target_modules": [
 
 
 
32
  "q_proj",
33
- "down_proj",
34
  "up_proj",
35
- "o_proj",
36
  "gate_proj",
37
- "k_proj",
38
- "v_proj"
39
  ],
40
  "target_parameters": null,
41
  "task_type": "CAUSAL_LM",
 
29
  "rank_pattern": {},
30
  "revision": null,
31
  "target_modules": [
32
+ "o_proj",
33
+ "v_proj",
34
+ "k_proj",
35
  "q_proj",
 
36
  "up_proj",
 
37
  "gate_proj",
38
+ "down_proj"
 
39
  ],
40
  "target_parameters": null,
41
  "task_type": "CAUSAL_LM",
adapter_model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:60d95b10b6e140a9626a7058d5038528f2ff80148dc4569b881db56052046509
3
- size 40
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:c4d4268fbced7e166240d2fd404e3dca36a791adfefec718d3067887406b39a8
3
+ size 335605144
config.json CHANGED
@@ -18,6 +18,7 @@
18
  "num_attention_heads": 32,
19
  "num_hidden_layers": 32,
20
  "num_key_value_heads": 8,
 
21
  "pretraining_tp": 1,
22
  "rms_norm_eps": 1e-05,
23
  "rope_scaling": {
 
18
  "num_attention_heads": 32,
19
  "num_hidden_layers": 32,
20
  "num_key_value_heads": 8,
21
+ "pad_token_id": 128001,
22
  "pretraining_tp": 1,
23
  "rms_norm_eps": 1e-05,
24
  "rope_scaling": {
generation_config.json CHANGED
@@ -2,7 +2,10 @@
2
  "_from_model_config": true,
3
  "bos_token_id": 128000,
4
  "do_sample": true,
5
- "eos_token_id": 128001,
 
 
 
6
  "temperature": 0.6,
7
  "top_p": 0.95,
8
  "transformers_version": "4.57.1"
 
2
  "_from_model_config": true,
3
  "bos_token_id": 128000,
4
  "do_sample": true,
5
+ "eos_token_id": [
6
+ 128001
7
+ ],
8
+ "pad_token_id": 128001,
9
  "temperature": 0.6,
10
  "top_p": 0.95,
11
  "transformers_version": "4.57.1"
model-00001-of-00004.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:00a90f92cc5605b61a0073398c7225385e2d035612aca045b9d71240bfb77849
3
  size 4976698672
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:3e2f1874281ef5e5f9c9fc1ec19710d47fe40bf7058ea969134658f8d1686463
3
  size 4976698672
model-00002-of-00004.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:68438e6709775ab0e8e322a60ae80ebd1f131a378bc2aa7f1e831ef147422ef2
3
  size 4999802720
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:3e35564ec1b452ecd3eb081a514ccd022ec8af23a9231644fed0d1456fa71fa8
3
  size 4999802720
model-00003-of-00004.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:40666efae41cbd67e18ba3620ac5caf3a61e7b666862ae014c99ff94c38d5154
3
  size 4915916176
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:7bb86752df561fbf75077a016cb99761d06176ac60dcdd0f82fc1c47979503a2
3
  size 4915916176
model-00004-of-00004.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:bc6192f16067485b8e90111208f423ba9a2c323053d4792dc50e20f7a99b816e
3
  size 1168138808
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:2e089bac37bd76643d34de725e54ba059c6e05312d087d3d68798fa7d49fb6af
3
  size 1168138808
tokenizer.json CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:d91915040cfac999d8c55f4b5bc6e67367c065e3a7a4e4b9438ce1f256addd86
3
- size 17209530
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:26c881aaf4ef935b1516ec79ad6405dd2a459f2b5d431a8a4a1399c92f3ba022
3
+ size 17209711
training_args.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:0de2821aab93454e4a327821e4e141f31c7342c176f9e11dfa2b4d0afd1130e8
3
  size 8760
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:87f504b40016b12c079540469b81ed1e8baa9535c9cd14bb840a8ed722825f4a
3
  size 8760