LeroyDyer commited on
Commit
ce9e871
1 Parent(s): a9dfc1d

(Trained with Unsloth)

Browse files
config.json CHANGED
@@ -1,5 +1,5 @@
1
  {
2
- "_name_or_path": "Mixtral_AI_CyberBrain_3_0",
3
  "architectures": [
4
  "MistralForCausalLM"
5
  ],
@@ -7,7 +7,7 @@
7
  "auto_map": {
8
  "AutoConfig": "LeroyDyer/Mixtral_AI_CyberBrain_3_0--configuration_mistral.MistralConfig",
9
  "AutoModelForCausalLM": "LeroyDyer/Mixtral_AI_CyberBrain_3_0--modeling_yarn.MistralForCausalLM"
10
- },
11
  "bos_token_id": 1,
12
  "eos_token_id": 2,
13
  "hidden_act": "silu",
@@ -16,7 +16,7 @@
16
  "intermediate_size": 14336,
17
  "max_position_embeddings": 32768,
18
  "max_sequence_length": 131072,
19
- "max_thoughts": 10,
20
  "merged_lm_and_talk_heads": false,
21
  "merged_lm_and_think_heads": true,
22
  "merged_talk_heads": true,
@@ -35,7 +35,7 @@
35
  "sliding_window": 131072,
36
  "tie_word_embeddings": false,
37
  "torch_dtype": "float16",
38
- "transformers_version": "4.38.2",
39
  "unsloth_version": "2024.3",
40
  "use_cache": true,
41
  "use_complex_talk_head": true,
 
1
  {
2
+ "_name_or_path": "LeroyDyer/Mixtral_AI_CyberBrain_3_0",
3
  "architectures": [
4
  "MistralForCausalLM"
5
  ],
 
7
  "auto_map": {
8
  "AutoConfig": "LeroyDyer/Mixtral_AI_CyberBrain_3_0--configuration_mistral.MistralConfig",
9
  "AutoModelForCausalLM": "LeroyDyer/Mixtral_AI_CyberBrain_3_0--modeling_yarn.MistralForCausalLM"
10
+ },
11
  "bos_token_id": 1,
12
  "eos_token_id": 2,
13
  "hidden_act": "silu",
 
16
  "intermediate_size": 14336,
17
  "max_position_embeddings": 32768,
18
  "max_sequence_length": 131072,
19
+ "max_thoughts": 3,
20
  "merged_lm_and_talk_heads": false,
21
  "merged_lm_and_think_heads": true,
22
  "merged_talk_heads": true,
 
35
  "sliding_window": 131072,
36
  "tie_word_embeddings": false,
37
  "torch_dtype": "float16",
38
+ "transformers_version": "4.40.0.dev0",
39
  "unsloth_version": "2024.3",
40
  "use_cache": true,
41
  "use_complex_talk_head": true,
generation_config.json CHANGED
@@ -2,5 +2,5 @@
2
  "_from_model_config": true,
3
  "bos_token_id": 1,
4
  "eos_token_id": 2,
5
- "transformers_version": "4.38.2"
6
  }
 
2
  "_from_model_config": true,
3
  "bos_token_id": 1,
4
  "eos_token_id": 2,
5
+ "transformers_version": "4.40.0.dev0"
6
  }
model-00001-of-00008.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:0cdf38792df4d5ee04361c485ce862de093555818b01badae5d5b10fd6b68116
3
  size 1889603392
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:43a7b89698db41c12893cdacf27f6b75620c70a5f2ad5167ad8c8e2b1b9f8b1f
3
  size 1889603392
model-00002-of-00008.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:032324e8b0b3dc13986ef61cae1fd95a4131d2d05b1193fc5ccd9be80e306cf4
3
  size 1946243896
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:13c367ff186b247da401fb2ac65dd27cfe8a3a384888f97e3c7626c9aff885a1
3
  size 1946243896
model-00003-of-00008.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:48304e4348902bc9eacf7a01e4a4d435db6705681e1c0b975721331c46bcf6d2
3
  size 1979781392
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:f54586c6e7e055fe2e2b7ae604bba9c44c16c3b17497b3970a0a2cb02b1ca957
3
  size 1979781392
model-00004-of-00008.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:a3536440cd845f80e6fc52a2ddc0570b252a783e495d53d80d062e6dd0eb045d
3
  size 1946243936
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:4309ab251636ecb7dff8f87cbc3c415e63f1a44b11cef0818b6dcdc325735d81
3
  size 1946243936
model-00005-of-00008.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:13b377d239c604b2cb40a6275a233f16038269a41f070ae228b70e8776513f7c
3
  size 1979781416
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:a32d7068f3c9ef207142f92cb8fe66a99e58b6070a6ebc20d507655697c0b21c
3
  size 1979781416
model-00006-of-00008.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:6a87fefad4957ac37e2dffb0181ac52d117cfed7e90ac22aa77f4e2f86e4065c
3
  size 1946243936
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:de67bc9d39baa7d2cc3983e494a7ea83e9be19659ab8915b5afdec8d9987ead9
3
  size 1946243936
model-00007-of-00008.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:874f5891cec7fe53e49b4ca940902965f7f3aa348ac9b0559ad56add49a90c2e
3
  size 1979781416
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:cd5c86883b18b92a8394ebef60d1cd9a36945c18c4707485da9a2914963cb8c5
3
  size 1979781416
model-00008-of-00008.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:e753d8616330e0954d88edda5aa56ab03fcd416626922c4424ed5ca75c9b1ef1
3
  size 815851048
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:77bca604fa9c3a207f86e0cbb4a4a5fe3d63498d8507f0b90491567f96925313
3
  size 815851048
special_tokens_map.json CHANGED
@@ -1,19 +1,7 @@
1
  {
2
  "additional_special_tokens": [
3
- {
4
- "content": "<|startthought|>",
5
- "lstrip": false,
6
- "normalized": false,
7
- "rstrip": false,
8
- "single_word": false
9
- },
10
- {
11
- "content": "<|endthought|>",
12
- "lstrip": false,
13
- "normalized": false,
14
- "rstrip": false,
15
- "single_word": false
16
- }
17
  ],
18
  "bos_token": {
19
  "content": "<s>",
@@ -29,7 +17,13 @@
29
  "rstrip": false,
30
  "single_word": false
31
  },
32
- "pad_token": "</s>",
 
 
 
 
 
 
33
  "unk_token": {
34
  "content": "<unk>",
35
  "lstrip": false,
 
1
  {
2
  "additional_special_tokens": [
3
+ "<|startthought|>",
4
+ "<|endthought|>"
 
 
 
 
 
 
 
 
 
 
 
 
5
  ],
6
  "bos_token": {
7
  "content": "<s>",
 
17
  "rstrip": false,
18
  "single_word": false
19
  },
20
+ "pad_token": {
21
+ "content": "</s>",
22
+ "lstrip": false,
23
+ "normalized": false,
24
+ "rstrip": false,
25
+ "single_word": false
26
+ },
27
  "unk_token": {
28
  "content": "<unk>",
29
  "lstrip": false,