aaabiao commited on
Commit
1382f2b
1 Parent(s): a275af4

Add files using large-upload tool

Browse files
all_results.json ADDED
@@ -0,0 +1,8 @@
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "epoch": 2.999650716032134,
3
+ "total_flos": 8645827988029440.0,
4
+ "train_loss": 0.35461137654308095,
5
+ "train_runtime": 104325.7524,
6
+ "train_samples_per_second": 63.226,
7
+ "train_steps_per_second": 0.123
8
+ }
checkpoint-2000/global_step2000/zero_pp_rank_20_mp_rank_00_model_states.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:74222b34b48fd4459bb0fc32e6b2d5f94969bd373d27483222b1de2b1409e691
3
+ size 151116
checkpoint-2000/rng_state_10.pth ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:4d4c67c1d0ec6b889370f634ff29f584f829acaeaf69196a69304fdba936f9d7
3
+ size 15997
checkpoint-2000/rng_state_12.pth ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:a9d1c7f561c5512193e8053ecf371e4fef647a8503481f2dad76332ddbe164fd
3
+ size 15997
checkpoint-2000/rng_state_14.pth ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:4d984186a3bc06445d6c077fcbf64da96d47d3ed704e68f0212045efb4b91dbb
3
+ size 15997
checkpoint-2000/rng_state_15.pth ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:1b3945ba3535361fa13a8460a70ebbd44ece2514a1c9dcb76abfe6f54a775c60
3
+ size 15997
checkpoint-2000/rng_state_18.pth ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:e167524b369e20a3a17ee7f6124322da680ccf2ae553b4adc0e7462d2dfba82b
3
+ size 15997
checkpoint-2000/rng_state_19.pth ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:a0915a78eba7c1d2768ecd03818e3c01afd7f3168a3318b1212413bd2c62f7f1
3
+ size 15997
checkpoint-2000/rng_state_26.pth ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:6b29094f6dcc92be7d58da39df551c573728836c5aec67bef90790810c0d3882
3
+ size 15997
checkpoint-2000/rng_state_27.pth ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:b64ff98dfb30bbfa08760a0a79d9be7407917ff0e32f22b03d6528bf7e292e44
3
+ size 15997
checkpoint-2000/rng_state_3.pth ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:ac55dba0b79d5fa4699d239da2f966d52040d576d31234ac8d4632e6956481bc
3
+ size 15984
checkpoint-2000/rng_state_38.pth ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:eae0a9bd1151c39106890b996fd767e0569650a669126b6513dbbf0513900d10
3
+ size 15997
checkpoint-2000/rng_state_4.pth ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:af2d0c015100768ffa23faf3b6c2d54ea89eb045603e30e55cd211e06ff34972
3
+ size 15984
checkpoint-2000/rng_state_46.pth ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:1d075e9c27d02dbb9e2a6dddb02dd39b943708e5e4bf03d03996614e5265fb38
3
+ size 15997
checkpoint-2000/rng_state_48.pth ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:1f0ab6144c506e09c95b742f41deb186593d88524f1c50e69d2856c4ba23697f
3
+ size 15997
checkpoint-2000/rng_state_51.pth ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:6abf2565abbb33088dd3547734e27c1d8b840d31677119e167880f3f38cf78e9
3
+ size 15997
checkpoint-2000/rng_state_54.pth ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:326ed1e97b4c78dc80c69932991f5732cb347f8847a4b9ff237df98dd72764f3
3
+ size 15997
checkpoint-2000/rng_state_55.pth ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:8b723bb3b70aee1abb8abb975a19dcdaab9ecd3443d374918a5789711f3045db
3
+ size 15997
checkpoint-2000/rng_state_59.pth ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:80f5d05d5236f8c5cd58c93f8e304f862f6c665e75f39a0f6e24734fc561ac6c
3
+ size 15997
checkpoint-2000/rng_state_60.pth ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:9cbfdc077a0836fc06ab3fd73bbabbe35c4db41f4318e20835410444fe1702b2
3
+ size 15997
checkpoint-2000/rng_state_62.pth ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:df375bc22957348c6eaebfa0b8cd9dc556deb98c8a760ba820df320268db26a1
3
+ size 15997
checkpoint-2000/rng_state_7.pth ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:38bc23a138cc800b22881742c0f3f9a71731a9a7111c6058a0077e6274d21773
3
+ size 15984
special_tokens_map.json ADDED
@@ -0,0 +1,30 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "bos_token": {
3
+ "content": "<s>",
4
+ "lstrip": false,
5
+ "normalized": false,
6
+ "rstrip": false,
7
+ "single_word": false
8
+ },
9
+ "eos_token": {
10
+ "content": "</s>",
11
+ "lstrip": false,
12
+ "normalized": false,
13
+ "rstrip": false,
14
+ "single_word": false
15
+ },
16
+ "pad_token": {
17
+ "content": "</s>",
18
+ "lstrip": false,
19
+ "normalized": false,
20
+ "rstrip": false,
21
+ "single_word": false
22
+ },
23
+ "unk_token": {
24
+ "content": "<unk>",
25
+ "lstrip": false,
26
+ "normalized": false,
27
+ "rstrip": false,
28
+ "single_word": false
29
+ }
30
+ }
tokenizer_config.json ADDED
@@ -0,0 +1,43 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "add_bos_token": true,
3
+ "add_eos_token": false,
4
+ "added_tokens_decoder": {
5
+ "0": {
6
+ "content": "<unk>",
7
+ "lstrip": false,
8
+ "normalized": false,
9
+ "rstrip": false,
10
+ "single_word": false,
11
+ "special": true
12
+ },
13
+ "1": {
14
+ "content": "<s>",
15
+ "lstrip": false,
16
+ "normalized": false,
17
+ "rstrip": false,
18
+ "single_word": false,
19
+ "special": true
20
+ },
21
+ "2": {
22
+ "content": "</s>",
23
+ "lstrip": false,
24
+ "normalized": false,
25
+ "rstrip": false,
26
+ "single_word": false,
27
+ "special": true
28
+ }
29
+ },
30
+ "additional_special_tokens": [],
31
+ "bos_token": "<s>",
32
+ "chat_template": "{{ bos_token }}{% for message in messages %}{% if (message['role'] == 'user') != (loop.index0 % 2 == 0) %}{{ raise_exception('Conversation roles must alternate user/assistant/user/assistant/...') }}{% endif %}{% if message['role'] == 'user' %}{{ '[INST] ' + message['content'] + ' [/INST]' }}{% elif message['role'] == 'assistant' %}{{ message['content'] + eos_token + ' ' }}{% else %}{{ raise_exception('Only user and assistant roles are supported!') }}{% endif %}{% endfor %}",
33
+ "clean_up_tokenization_spaces": false,
34
+ "eos_token": "</s>",
35
+ "legacy": true,
36
+ "model_max_length": 1000000000000000019884624838656,
37
+ "pad_token": null,
38
+ "sp_model_kwargs": {},
39
+ "spaces_between_special_tokens": false,
40
+ "tokenizer_class": "LlamaTokenizer",
41
+ "unk_token": "<unk>",
42
+ "use_default_system_prompt": false
43
+ }
train_results.json ADDED
@@ -0,0 +1,8 @@
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "epoch": 2.999650716032134,
3
+ "total_flos": 8645827988029440.0,
4
+ "train_loss": 0.35461137654308095,
5
+ "train_runtime": 104325.7524,
6
+ "train_samples_per_second": 63.226,
7
+ "train_steps_per_second": 0.123
8
+ }