RuoxiL commited on
Commit
5f93312
1 Parent(s): 80fc1e5

Training in progress, epoch 0

Browse files
Files changed (38) hide show
  1. adapter_config.json +33 -0
  2. adapter_model.safetensors +0 -0
  3. added_tokens.json +4 -0
  4. merges.txt +0 -0
  5. runs/Apr15_00-02-03_c20/events.out.tfevents.1713153729.c20 +3 -0
  6. runs/Apr15_00-02-04_c14/events.out.tfevents.1713153729.c14 +3 -0
  7. runs/Apr15_00-02-04_c16/events.out.tfevents.1713153729.c16 +3 -0
  8. runs/Apr15_00-02-05_c21/events.out.tfevents.1713153729.c21 +3 -0
  9. runs/Apr15_00-02-05_c22/events.out.tfevents.1713153729.c22 +3 -0
  10. runs/Apr15_00-55-05_c20/events.out.tfevents.1713156909.c20 +3 -0
  11. runs/Apr15_00-55-06_c14/events.out.tfevents.1713156910.c14 +3 -0
  12. runs/Apr15_00-55-06_c16/events.out.tfevents.1713156909.c16 +3 -0
  13. runs/Apr15_00-55-10_c21/events.out.tfevents.1713156914.c21 +3 -0
  14. runs/Apr15_00-56-14_c20/events.out.tfevents.1713156978.c20 +3 -0
  15. runs/Apr15_00-56-15_c14/events.out.tfevents.1713156979.c14 +3 -0
  16. runs/Apr15_00-56-15_c16/events.out.tfevents.1713156979.c16 +3 -0
  17. runs/Apr15_00-56-15_c21/events.out.tfevents.1713156978.c21 +3 -0
  18. runs/Apr15_00-56-20_c22/events.out.tfevents.1713156984.c22 +3 -0
  19. runs/Apr15_01-09-44_c14/events.out.tfevents.1713157793.c14 +3 -0
  20. runs/Apr15_01-09-44_c16/events.out.tfevents.1713157794.c16 +3 -0
  21. runs/Apr15_01-09-44_c20/events.out.tfevents.1713157793.c20 +3 -0
  22. runs/Apr15_01-09-58_c21/events.out.tfevents.1713157807.c21 +3 -0
  23. runs/Apr15_01-09-58_c22/events.out.tfevents.1713157807.c22 +3 -0
  24. runs/Apr15_01-22-06_c20/events.out.tfevents.1713158535.c20 +3 -0
  25. runs/Apr15_01-22-06_c21/events.out.tfevents.1713158535.c21 +3 -0
  26. runs/Apr15_01-22-07_c14/events.out.tfevents.1713158536.c14 +3 -0
  27. runs/Apr15_01-22-07_c16/events.out.tfevents.1713158536.c16 +3 -0
  28. runs/Apr15_01-22-07_c22/events.out.tfevents.1713158536.c22 +3 -0
  29. runs/Apr15_02-23-35_c20/events.out.tfevents.1713162230.c20 +3 -0
  30. runs/Apr15_02-23-35_c21/events.out.tfevents.1713162230.c21 +3 -0
  31. runs/Apr15_02-23-35_c22/events.out.tfevents.1713162230.c22 +3 -0
  32. runs/Apr15_02-23-36_c14/events.out.tfevents.1713162230.c14 +3 -0
  33. runs/Apr15_02-23-36_c16/events.out.tfevents.1713162230.c16 +3 -0
  34. special_tokens_map.json +28 -0
  35. tokenizer.json +0 -0
  36. tokenizer_config.json +51 -0
  37. training_args.bin +3 -0
  38. vocab.json +0 -0
adapter_config.json ADDED
@@ -0,0 +1,33 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "alpha_pattern": {},
3
+ "auto_mapping": null,
4
+ "base_model_name_or_path": "facebook/opt-2.7b",
5
+ "bias": "none",
6
+ "fan_in_fan_out": false,
7
+ "inference_mode": true,
8
+ "init_lora_weights": true,
9
+ "layer_replication": null,
10
+ "layers_pattern": null,
11
+ "layers_to_transform": null,
12
+ "loftq_config": {},
13
+ "lora_alpha": 128,
14
+ "lora_dropout": 0.05,
15
+ "megatron_config": null,
16
+ "megatron_core": "megatron.core",
17
+ "modules_to_save": null,
18
+ "peft_type": "LORA",
19
+ "r": 256,
20
+ "rank_pattern": {},
21
+ "revision": null,
22
+ "target_modules": [
23
+ "k_proj",
24
+ "v_proj",
25
+ "q_proj",
26
+ "fc1",
27
+ "out_proj",
28
+ "fc2"
29
+ ],
30
+ "task_type": "CAUSAL_LM",
31
+ "use_dora": false,
32
+ "use_rslora": false
33
+ }
adapter_model.safetensors ADDED
File without changes
added_tokens.json ADDED
@@ -0,0 +1,4 @@
 
 
 
 
 
1
+ {
2
+ "<|im_end|>": 50266,
3
+ "<|im_start|>": 50265
4
+ }
merges.txt ADDED
The diff for this file is too large to render. See raw diff
 
runs/Apr15_00-02-03_c20/events.out.tfevents.1713153729.c20 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:70f661dfa0a43c0987d39671ec54014553466a0e842ff511d0ba52024dae0642
3
+ size 5244
runs/Apr15_00-02-04_c14/events.out.tfevents.1713153729.c14 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:d90d95a457f987930053aa828df53dc0a56fb5dbbc173c02506caa5ccd5f7fbb
3
+ size 5244
runs/Apr15_00-02-04_c16/events.out.tfevents.1713153729.c16 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:b810bc30aea062cff4ac30d3bae62e53d36e984696dcbfadfbab301d45674526
3
+ size 5244
runs/Apr15_00-02-05_c21/events.out.tfevents.1713153729.c21 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:420462afe429cbd197c12e07bc64fe05ead80fe0b09ae53be597638e7a5231e0
3
+ size 5244
runs/Apr15_00-02-05_c22/events.out.tfevents.1713153729.c22 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:533fa6ea74a9136f7d0c1b53a947ce077e9453ff38f5a14368956a4435523dcf
3
+ size 5244
runs/Apr15_00-55-05_c20/events.out.tfevents.1713156909.c20 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:3451fdfa7736b8ac61703eba8d0a898d474be71525440a57713507ad706e7ac8
3
+ size 5244
runs/Apr15_00-55-06_c14/events.out.tfevents.1713156910.c14 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:b7cc6f0c02f95883fafca360539fab2beb43468951ba256475aa651d77852c97
3
+ size 5244
runs/Apr15_00-55-06_c16/events.out.tfevents.1713156909.c16 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:f56be4af1b8f56305da3650e986772b503ce2e6056dfb4d19273b7731262c806
3
+ size 5244
runs/Apr15_00-55-10_c21/events.out.tfevents.1713156914.c21 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:7b950e257f2c6a171f9a297381f229cede7ed5c0d8d6a2aa6e4f55ce5a4feaf4
3
+ size 5244
runs/Apr15_00-56-14_c20/events.out.tfevents.1713156978.c20 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:2e4f991ec35cf75a100c71e0d4c2bd478ca743d5d1431d4b870652401b84a132
3
+ size 5451
runs/Apr15_00-56-15_c14/events.out.tfevents.1713156979.c14 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:9e9178d47e0e65432d9e83b1a28ab7a85a3decdd9ae7c7bed908c9cdfc031287
3
+ size 5451
runs/Apr15_00-56-15_c16/events.out.tfevents.1713156979.c16 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:884b64c1202d68081cfc5b5bead77df8e1fb6201202bd882d4594c2c9b5a272b
3
+ size 5451
runs/Apr15_00-56-15_c21/events.out.tfevents.1713156978.c21 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:228e32e3bc047ba8a0256fc12118c31b2676b132f3b5d02101a60e702d0b1122
3
+ size 5451
runs/Apr15_00-56-20_c22/events.out.tfevents.1713156984.c22 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:bbe00b05469bec43ced66e480a2c6a2a4245284d91854b1b270d0fd7ac022523
3
+ size 5451
runs/Apr15_01-09-44_c14/events.out.tfevents.1713157793.c14 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:e987736e18f4508b70e243812ace62eea2d2739e3fc9c703819a7fca8a4924f9
3
+ size 5449
runs/Apr15_01-09-44_c16/events.out.tfevents.1713157794.c16 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:a24acc4b359b0c88e3123316a551b3e0e1f3949dad84e29c455b900216b94f18
3
+ size 5449
runs/Apr15_01-09-44_c20/events.out.tfevents.1713157793.c20 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:7f57a948a7589a505e81988a2e27acca17ee759a7ef3e16032686d12592dad4c
3
+ size 5449
runs/Apr15_01-09-58_c21/events.out.tfevents.1713157807.c21 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:266bc99ed0bb80c97a2bbb4379e614217088887e9a150e50ecd7cb628b61645c
3
+ size 5449
runs/Apr15_01-09-58_c22/events.out.tfevents.1713157807.c22 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:9f8e0fe2d61ba88f7aa62b73d4f5a4b0d189a470050fb154cc8ad1b63d0a56f0
3
+ size 5449
runs/Apr15_01-22-06_c20/events.out.tfevents.1713158535.c20 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:b667b80a31fc227e3a44affb92bdbff13c3918eb08d8aa26615cd2f1350c393e
3
+ size 7105
runs/Apr15_01-22-06_c21/events.out.tfevents.1713158535.c21 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:8077ced039127a56ac41f7598716369cc4ff94228e8df493dc713f3a4f9385ac
3
+ size 7105
runs/Apr15_01-22-07_c14/events.out.tfevents.1713158536.c14 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:43c1271fbfe642bce2e909d8e74c4986eb8a61bc9bf0ba4216dd31fa7e454ef3
3
+ size 7105
runs/Apr15_01-22-07_c16/events.out.tfevents.1713158536.c16 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:3fff2700ee5b54164eaf1a58d3a076bbcdb99ac73d8e369bf80ccc5a86281328
3
+ size 7105
runs/Apr15_01-22-07_c22/events.out.tfevents.1713158536.c22 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:f679608a89040781bf416c0c35789777edfa243db80dd3056d5d89fc36cd722f
3
+ size 7105
runs/Apr15_02-23-35_c20/events.out.tfevents.1713162230.c20 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:e6ca48c86ec3e1eb0eb39ff8fb5c61d85018f084c8ebe9aa4e98c3204f32c387
3
+ size 9625
runs/Apr15_02-23-35_c21/events.out.tfevents.1713162230.c21 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:d83b5db31fa6d2cc0160478641ee226f67a162928e283c1792350e479f9b4eb7
3
+ size 9625
runs/Apr15_02-23-35_c22/events.out.tfevents.1713162230.c22 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:d9edaf906d70cb20715f9c69f34fcd3689e367eef447b47d71ffc54fac154ae3
3
+ size 9625
runs/Apr15_02-23-36_c14/events.out.tfevents.1713162230.c14 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:fcbcdacd80685a59edc23de8ada2d5432edd1d1a4a50b42e8c672220785c8440
3
+ size 9625
runs/Apr15_02-23-36_c16/events.out.tfevents.1713162230.c16 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:85a00100e141a7c85abfb98620e2cfdc0fd9bf06c54c33da4c305f0af3b1f8e2
3
+ size 9625
special_tokens_map.json ADDED
@@ -0,0 +1,28 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "additional_special_tokens": [
3
+ {
4
+ "content": "<|im_start|>",
5
+ "lstrip": false,
6
+ "normalized": false,
7
+ "rstrip": false,
8
+ "single_word": false
9
+ },
10
+ {
11
+ "content": "<|im_end|>",
12
+ "lstrip": false,
13
+ "normalized": false,
14
+ "rstrip": false,
15
+ "single_word": false
16
+ }
17
+ ],
18
+ "bos_token": "<|im_start|>",
19
+ "eos_token": "<|im_end|>",
20
+ "pad_token": "<|im_end|>",
21
+ "unk_token": {
22
+ "content": "</s>",
23
+ "lstrip": false,
24
+ "normalized": true,
25
+ "rstrip": false,
26
+ "single_word": false
27
+ }
28
+ }
tokenizer.json ADDED
The diff for this file is too large to render. See raw diff
 
tokenizer_config.json ADDED
@@ -0,0 +1,51 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "add_bos_token": true,
3
+ "add_prefix_space": false,
4
+ "added_tokens_decoder": {
5
+ "1": {
6
+ "content": "<pad>",
7
+ "lstrip": false,
8
+ "normalized": true,
9
+ "rstrip": false,
10
+ "single_word": false,
11
+ "special": true
12
+ },
13
+ "2": {
14
+ "content": "</s>",
15
+ "lstrip": false,
16
+ "normalized": true,
17
+ "rstrip": false,
18
+ "single_word": false,
19
+ "special": true
20
+ },
21
+ "50265": {
22
+ "content": "<|im_start|>",
23
+ "lstrip": false,
24
+ "normalized": false,
25
+ "rstrip": false,
26
+ "single_word": false,
27
+ "special": true
28
+ },
29
+ "50266": {
30
+ "content": "<|im_end|>",
31
+ "lstrip": false,
32
+ "normalized": false,
33
+ "rstrip": false,
34
+ "single_word": false,
35
+ "special": true
36
+ }
37
+ },
38
+ "additional_special_tokens": [
39
+ "<|im_start|>",
40
+ "<|im_end|>"
41
+ ],
42
+ "bos_token": "<|im_start|>",
43
+ "chat_template": "{% for message in messages %}{{'<|im_start|>' + message['role'] + '\n' + message['content'] + '<|im_end|>' + '\n'}}{% endfor %}{% if add_generation_prompt %}{{ '<|im_start|>assistant\n' }}{% endif %}",
44
+ "clean_up_tokenization_spaces": true,
45
+ "eos_token": "<|im_end|>",
46
+ "errors": "replace",
47
+ "model_max_length": 1000000000000000019884624838656,
48
+ "pad_token": "<|im_end|>",
49
+ "tokenizer_class": "GPT2Tokenizer",
50
+ "unk_token": "</s>"
51
+ }
training_args.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:d1a3e01ef2009f4507553f80d65f1d6bcfc0285392be86b52d201470e87c3203
3
+ size 4984
vocab.json ADDED
The diff for this file is too large to render. See raw diff