satyamt commited on
Commit
e28ccaa
1 Parent(s): c1c15fc

Upload folder using huggingface_hub

Browse files
README.md CHANGED
@@ -8,7 +8,6 @@ language:
8
  - en
9
  tags:
10
  - medical
11
- license: mit
12
  ---
13
 
14
 
 
8
  - en
9
  tags:
10
  - medical
 
11
  ---
12
 
13
 
added_tokens.json CHANGED
@@ -4,8 +4,6 @@
4
  "<MASK>": 32003,
5
  "<PAD>": 32004,
6
  "<SEP>": 32001,
7
- "<|im_end|>": 32018,
8
- "<|im_start|>": 32017,
9
  "[/bib]": 32010,
10
  "[/bib_ref]": 32006,
11
  "[/fig]": 32012,
 
4
  "<MASK>": 32003,
5
  "<PAD>": 32004,
6
  "<SEP>": 32001,
 
 
7
  "[/bib]": 32010,
8
  "[/bib_ref]": 32006,
9
  "[/fig]": 32012,
config.json CHANGED
@@ -1,5 +1,5 @@
1
  {
2
- "_name_or_path": "Technoculture/MT7Bi-sft",
3
  "architectures": [
4
  "LlamaForCausalLM"
5
  ],
@@ -11,7 +11,7 @@
11
  "hidden_size": 4096,
12
  "initializer_range": 0.02,
13
  "intermediate_size": 11008,
14
- "max_position_embeddings": 4096,
15
  "model_type": "llama",
16
  "num_attention_heads": 32,
17
  "num_hidden_layers": 32,
@@ -22,7 +22,7 @@
22
  "rope_theta": 10000.0,
23
  "tie_word_embeddings": false,
24
  "torch_dtype": "bfloat16",
25
- "transformers_version": "4.37.0",
26
  "use_cache": false,
27
- "vocab_size": 32019
28
  }
 
1
  {
2
+ "_name_or_path": "epfl-llm/meditron-7b",
3
  "architectures": [
4
  "LlamaForCausalLM"
5
  ],
 
11
  "hidden_size": 4096,
12
  "initializer_range": 0.02,
13
  "intermediate_size": 11008,
14
+ "max_position_embeddings": 2048,
15
  "model_type": "llama",
16
  "num_attention_heads": 32,
17
  "num_hidden_layers": 32,
 
22
  "rope_theta": 10000.0,
23
  "tie_word_embeddings": false,
24
  "torch_dtype": "bfloat16",
25
+ "transformers_version": "4.37.0.dev0",
26
  "use_cache": false,
27
+ "vocab_size": 32017
28
  }
generation_config.json CHANGED
@@ -1,7 +1,6 @@
1
  {
2
  "_from_model_config": true,
3
  "bos_token_id": 1,
4
- "do_sample": true,
5
  "eos_token_id": 2,
6
- "transformers_version": "4.37.0"
7
  }
 
1
  {
2
  "_from_model_config": true,
3
  "bos_token_id": 1,
 
4
  "eos_token_id": 2,
5
+ "transformers_version": "4.37.0.dev0"
6
  }
pytorch_model-00001-of-00003.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:1f6bd28f1e053c6166ca1bdb7f2bf06e2508174ed4accd0de88751cef663dc12
3
- size 4939165746
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:fc9b70dc0a4162670b04bafde40ef83b97d1038c17c6d2095ac7d97a1d48d43a
3
+ size 4939149362
pytorch_model-00002-of-00003.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:eaa8a0f33296fa5b3d8b8b30873e80e5fd2f1f6982f7671c1cc11bb6b18afe6a
3
  size 4947416538
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:5a83c5ed744d67e2ceb0623ab02cb78ef8324863812bec76ddd0fd31a87fe0aa
3
  size 4947416538
pytorch_model-00003-of-00003.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:ccc4c78c8c212d0534d07eb11d9c9fe16049ae1a99d03a42c6f8367c7ead95db
3
- size 3590662393
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:7ccfd4cd5d3e53b8bb0a4a9cafd0e6bbe9bbab53406b1369cafd6d0bb130f86d
3
+ size 3590646009
pytorch_model.bin.index.json CHANGED
@@ -1,6 +1,6 @@
1
  {
2
  "metadata": {
3
- "total_size": 13477142528
4
  },
5
  "weight_map": {
6
  "lm_head.weight": "pytorch_model-00003-of-00003.bin",
 
1
  {
2
  "metadata": {
3
+ "total_size": 13477109760
4
  },
5
  "weight_map": {
6
  "lm_head.weight": "pytorch_model-00003-of-00003.bin",
tokenizer_config.json CHANGED
@@ -161,22 +161,6 @@
161
  "rstrip": false,
162
  "single_word": false,
163
  "special": true
164
- },
165
- "32017": {
166
- "content": "<|im_start|>",
167
- "lstrip": false,
168
- "normalized": false,
169
- "rstrip": false,
170
- "single_word": false,
171
- "special": false
172
- },
173
- "32018": {
174
- "content": "<|im_end|>",
175
- "lstrip": false,
176
- "normalized": false,
177
- "rstrip": false,
178
- "single_word": false,
179
- "special": false
180
  }
181
  },
182
  "additional_special_tokens": [
@@ -194,7 +178,6 @@
194
  "[/formula]"
195
  ],
196
  "bos_token": "<s>",
197
- "chat_template": "{% if messages[0]['role'] == 'system' %}{% set loop_messages = messages[1:] %}{% set system_message = messages[0]['content'] %}{% else %}{% set loop_messages = messages %}{% set system_message = 'You are a helpful assistant.' %}{% endif %}{% if not add_generation_prompt is defined %}{% set add_generation_prompt = false %}{% endif %}{% for message in loop_messages %}{% if loop.index0 == 0 %}{{'<|im_start|>system\n' + system_message + '<|im_end|>\n'}}{% endif %}{{'<|im_start|>' + message['role'] + '\n' + message['content'] + '<|im_end|>' + '\n'}}{% endfor %}{% if add_generation_prompt %}{{ '<|im_start|>assistant\n' }}{% endif %}",
198
  "clean_up_tokenization_spaces": false,
199
  "cls_token": "<CLS>",
200
  "eos_token": "</s>",
 
161
  "rstrip": false,
162
  "single_word": false,
163
  "special": true
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
164
  }
165
  },
166
  "additional_special_tokens": [
 
178
  "[/formula]"
179
  ],
180
  "bos_token": "<s>",
 
181
  "clean_up_tokenization_spaces": false,
182
  "cls_token": "<CLS>",
183
  "eos_token": "</s>",