Pclanglais commited on
Commit
df17903
1 Parent(s): aa4897a

Upload folder using huggingface_hub

Browse files
config.json CHANGED
@@ -1,8 +1,9 @@
1
  {
2
- "_name_or_path": "mistral-hermes",
3
  "architectures": [
4
  "MistralForCausalLM"
5
  ],
 
6
  "bos_token_id": 1,
7
  "eos_token_id": 32000,
8
  "hidden_act": "silu",
@@ -19,7 +20,7 @@
19
  "sliding_window": 4096,
20
  "tie_word_embeddings": false,
21
  "torch_dtype": "bfloat16",
22
- "transformers_version": "4.35.1",
23
  "use_cache": false,
24
  "vocab_size": 32002
25
  }
 
1
  {
2
+ "_name_or_path": "mistral-hermes-2.5",
3
  "architectures": [
4
  "MistralForCausalLM"
5
  ],
6
+ "attention_dropout": 0.0,
7
  "bos_token_id": 1,
8
  "eos_token_id": 32000,
9
  "hidden_act": "silu",
 
20
  "sliding_window": 4096,
21
  "tie_word_embeddings": false,
22
  "torch_dtype": "bfloat16",
23
+ "transformers_version": "4.37.0",
24
  "use_cache": false,
25
  "vocab_size": 32002
26
  }
generation_config.json CHANGED
@@ -2,6 +2,5 @@
2
  "_from_model_config": true,
3
  "bos_token_id": 1,
4
  "eos_token_id": 32000,
5
- "transformers_version": "4.35.1",
6
- "use_cache": false
7
  }
 
2
  "_from_model_config": true,
3
  "bos_token_id": 1,
4
  "eos_token_id": 32000,
5
+ "transformers_version": "4.37.0"
 
6
  }
model-00001-of-00003.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:c4720cc70b4c1600cc14d35d5514fbec34ec7d6672aa91519ee862740a257021
3
  size 4943178720
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:6f183b901f05af17b043657e7f1d4b570b903c9386400e31b53e31320a4dfef4
3
  size 4943178720
model-00002-of-00003.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:699affccba1d2f8d626a6bd6f5958ef503e5c81726052ab63df00e0d30168398
3
  size 4999819336
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:43d9be88a4817499bb516292d8806461debd9462c70fe3a9e20ca301625190db
3
  size 4999819336
model-00003-of-00003.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:f00bba375036f763c987ef484d7499ac99bd0847a98c3e0231064dd179899997
3
  size 4540532728
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:198d2f5141fb26479dd63cae58926e12f3576b0876096a3be5be95f8bb5e19ea
3
  size 4540532728
special_tokens_map.json CHANGED
@@ -1,9 +1,4 @@
1
  {
2
- "additional_special_tokens": [
3
- "<unk>",
4
- "<s>",
5
- "</s>"
6
- ],
7
  "bos_token": {
8
  "content": "<s>",
9
  "lstrip": false,
@@ -18,13 +13,7 @@
18
  "rstrip": false,
19
  "single_word": false
20
  },
21
- "pad_token": {
22
- "content": "</s>",
23
- "lstrip": false,
24
- "normalized": false,
25
- "rstrip": false,
26
- "single_word": false
27
- },
28
  "unk_token": {
29
  "content": "<unk>",
30
  "lstrip": false,
 
1
  {
 
 
 
 
 
2
  "bos_token": {
3
  "content": "<s>",
4
  "lstrip": false,
 
13
  "rstrip": false,
14
  "single_word": false
15
  },
16
+ "pad_token": "<|im_end|>",
 
 
 
 
 
 
17
  "unk_token": {
18
  "content": "<unk>",
19
  "lstrip": false,
tokenizer.json CHANGED
@@ -51,7 +51,7 @@
51
  "lstrip": false,
52
  "rstrip": false,
53
  "normalized": false,
54
- "special": false
55
  }
56
  ],
57
  "normalizer": {
 
51
  "lstrip": false,
52
  "rstrip": false,
53
  "normalized": false,
54
+ "special": true
55
  }
56
  ],
57
  "normalizer": {
tokenizer_config.json CHANGED
@@ -40,21 +40,17 @@
40
  "normalized": false,
41
  "rstrip": false,
42
  "single_word": false,
43
- "special": false
44
  }
45
  },
46
- "additional_special_tokens": [
47
- "<unk>",
48
- "<s>",
49
- "</s>"
50
- ],
51
  "bos_token": "<s>",
52
  "chat_template": "{% for message in messages %}{{'<|im_start|>' + message['role'] + '\n' + message['content'] + '<|im_end|>' + '\n'}}{% endfor %}{% if add_generation_prompt %}{{ '<|im_start|>assistant\n' }}{% endif %}",
53
  "clean_up_tokenization_spaces": false,
54
  "eos_token": "<|im_end|>",
55
  "legacy": true,
56
  "model_max_length": 1000000000000000019884624838656,
57
- "pad_token": "</s>",
58
  "sp_model_kwargs": {},
59
  "spaces_between_special_tokens": false,
60
  "tokenizer_class": "LlamaTokenizer",
 
40
  "normalized": false,
41
  "rstrip": false,
42
  "single_word": false,
43
+ "special": true
44
  }
45
  },
46
+ "additional_special_tokens": [],
 
 
 
 
47
  "bos_token": "<s>",
48
  "chat_template": "{% for message in messages %}{{'<|im_start|>' + message['role'] + '\n' + message['content'] + '<|im_end|>' + '\n'}}{% endfor %}{% if add_generation_prompt %}{{ '<|im_start|>assistant\n' }}{% endif %}",
49
  "clean_up_tokenization_spaces": false,
50
  "eos_token": "<|im_end|>",
51
  "legacy": true,
52
  "model_max_length": 1000000000000000019884624838656,
53
+ "pad_token": "<|im_end|>",
54
  "sp_model_kwargs": {},
55
  "spaces_between_special_tokens": false,
56
  "tokenizer_class": "LlamaTokenizer",