mkshing commited on
Commit
b6cc5b9
·
verified ·
1 Parent(s): 678dcfc

Delete step-210k

Browse files
step-210k/added_tokens.json DELETED
@@ -1,5 +0,0 @@
1
- {
2
- "<|endoftext|>": 151643,
3
- "<|im_end|>": 151645,
4
- "<|im_start|>": 151644
5
- }
 
 
 
 
 
 
step-210k/config.json DELETED
@@ -1,29 +0,0 @@
1
- {
2
- "_name_or_path": "Qwen/Qwen2.5-1.5B",
3
- "architectures": [
4
- "Qwen2ForCausalLM"
5
- ],
6
- "attention_dropout": 0.0,
7
- "bos_token_id": 151643,
8
- "eos_token_id": 151643,
9
- "hidden_act": "silu",
10
- "hidden_size": 1536,
11
- "initializer_range": 0.02,
12
- "intermediate_size": 8960,
13
- "max_position_embeddings": 131072,
14
- "max_window_layers": 28,
15
- "model_type": "qwen2",
16
- "num_attention_heads": 12,
17
- "num_hidden_layers": 28,
18
- "num_key_value_heads": 2,
19
- "rms_norm_eps": 1e-06,
20
- "rope_theta": 1000000.0,
21
- "sliding_window": 131072,
22
- "tie_word_embeddings": true,
23
- "torch_dtype": "bfloat16",
24
- "transformers_version": "4.41.2",
25
- "use_cache": true,
26
- "use_mrope": false,
27
- "use_sliding_window": false,
28
- "vocab_size": 151936
29
- }
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
step-210k/generation_config.json DELETED
@@ -1,6 +0,0 @@
1
- {
2
- "bos_token_id": 151643,
3
- "eos_token_id": 151643,
4
- "max_new_tokens": 2048,
5
- "transformers_version": "4.41.2"
6
- }
 
 
 
 
 
 
 
step-210k/merges.txt DELETED
The diff for this file is too large to render. See raw diff
 
step-210k/model.safetensors DELETED
@@ -1,3 +0,0 @@
1
- version https://git-lfs.github.com/spec/v1
2
- oid sha256:cf1ff606a9c0281f824cca9d301e4c3571200575e38f2a6ca3329d8c992dc41d
3
- size 3087467144
 
 
 
 
step-210k/special_tokens_map.json DELETED
@@ -1,20 +0,0 @@
1
- {
2
- "additional_special_tokens": [
3
- "<|im_start|>",
4
- "<|im_end|>"
5
- ],
6
- "eos_token": {
7
- "content": "<|im_end|>",
8
- "lstrip": false,
9
- "normalized": false,
10
- "rstrip": false,
11
- "single_word": false
12
- },
13
- "pad_token": {
14
- "content": "<|endoftext|>",
15
- "lstrip": false,
16
- "normalized": false,
17
- "rstrip": false,
18
- "single_word": false
19
- }
20
- }
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
step-210k/tokenizer.json DELETED
The diff for this file is too large to render. See raw diff
 
step-210k/tokenizer_config.json DELETED
@@ -1,43 +0,0 @@
1
- {
2
- "add_prefix_space": false,
3
- "added_tokens_decoder": {
4
- "151643": {
5
- "content": "<|endoftext|>",
6
- "lstrip": false,
7
- "normalized": false,
8
- "rstrip": false,
9
- "single_word": false,
10
- "special": true
11
- },
12
- "151644": {
13
- "content": "<|im_start|>",
14
- "lstrip": false,
15
- "normalized": false,
16
- "rstrip": false,
17
- "single_word": false,
18
- "special": true
19
- },
20
- "151645": {
21
- "content": "<|im_end|>",
22
- "lstrip": false,
23
- "normalized": false,
24
- "rstrip": false,
25
- "single_word": false,
26
- "special": true
27
- }
28
- },
29
- "additional_special_tokens": [
30
- "<|im_start|>",
31
- "<|im_end|>"
32
- ],
33
- "bos_token": null,
34
- "chat_template": "{% for message in messages %}{% if loop.first and messages[0]['role'] != 'system' %}{{ '<|im_start|>system\nYou are a helpful assistant.<|im_end|>\n' }}{% endif %}{{'<|im_start|>' + message['role'] + '\n' + message['content'] + '<|im_end|>' + '\n'}}{% endfor %}{% if add_generation_prompt %}{{ '<|im_start|>assistant\n' }}{% endif %}",
35
- "clean_up_tokenization_spaces": false,
36
- "eos_token": "<|im_end|>",
37
- "errors": "replace",
38
- "model_max_length": 32768,
39
- "pad_token": "<|endoftext|>",
40
- "split_special_tokens": false,
41
- "tokenizer_class": "Qwen2Tokenizer",
42
- "unk_token": null
43
- }
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
step-210k/vocab.json DELETED
The diff for this file is too large to render. See raw diff