Upload folder using huggingface_hub (#1)
Browse files- 6ecb23591a48b6867ebfc8b828ca5c5b2db56f36cbb440887355db94b8785885 (7fc134bdee6fd5c077f033c47ac700d6b42a11dd)
- d70194bd13bc26b0a6ce9d16e101473d9701f0e0e78629076209a960d7ad1a58 (0c247cc2aeb5d8df921c729e6cff1ea3247b913c)
- 4684530db0fc07d5f647fe42060a152945a92e622832321133de7d02698ea3bc (073c734e2f93db930c4596a47af6b5edff801e0d)
- dbde3ec763a36a4e0e76e0e645fd52e0049d5967dc373c4ae726aa8db2775604 (9b727baf7e3b272d8136624782f56a9918de9355)
- f246ba5e5badf170e1cec4afe9c17d65903c82fdb80d8b6fb255f680d4c832e4 (8c3c3edc84a816b958f612ba17299b93761570f4)
- 3fbf90629794ee806458feb06feb635156129cec51997e61034c5adc454ccb12 (5d0a9c5e433020e0f93bcd6b36f843bd47280743)
- 6548b06326cc41a1184599a448cdf7b98b873ef30aab3c025b45c931f4f49872 (7c409b491ed14be32b1763f5782324a8ef983c32)
- 1bbe04e8e77ddeda22ef180038a73a7914867c4a5364df8dd3c1fd1c69860940 (80b9ec5fd2dbb7ea9ef32462556ded2dabf7a186)
- 103cbb32264220012ef129cea62213d3c369affdfb13db999e0bd74a897f640a (72da5283ccd74679349e12b0772a341b5ff63748)
- 6afaf69382c7e0263219f1f8475cd80bc9a399843d4b89a8171a606fde76be98 (4ecb7d30b56571899d1b03800c70a2fc0317b876)
- 243118b3a560eb61ffe537c63d6e0775a363e99c693300ea1c388a2a7e148194 (40a1458f6efd3d067892f6756d587669e7f6b2d8)
- a8f734997758ef99816736d5ec1967d82ce8a15479075b88b25d328c69d47348 (2fb6e48c13fe78ae21020574ff88e0d79be3e86b)
- 0c7fad0c7b366059590dd3e2ba5a7ec741899b8567d70577874119e5f29de023 (eaf820965e261f20d9786d859765fa503ca21a0e)
- 9bd2115bd6c3ee081171cbcddbf4c71f204319fa13155bc10875acfc46f95c8a (d98e96f928c784705c3b0480dde25dce3b1ac1a8)
- cdaa10a48297b297e09c4302ba2e18eda011271fef6f9141971c63283ee220c8 (dd2fa4e07635f6c077b74fad9b6c9de30393a3df)
- aeb865b243028c638d287c8bf6074e906ca59226b53a3873860a67e219f9436c (660f4a7d57fd46c8570f4be5d203008b069a2bc9)
- f4fbc5352335cc478b6e1ac03b0a55e4688df9da7d38ab22d070ed7fb94071e0 (0d369323c0280741a045e0985b48bcb6a840df69)
- config.json +46 -0
- generation_config.json +6 -0
- model-00001-of-00017.safetensors +3 -0
- model-00002-of-00017.safetensors +3 -0
- model-00003-of-00017.safetensors +3 -0
- model-00004-of-00017.safetensors +3 -0
- model-00005-of-00017.safetensors +3 -0
- model-00006-of-00017.safetensors +3 -0
- model-00007-of-00017.safetensors +3 -0
- model-00008-of-00017.safetensors +3 -0
- model-00009-of-00017.safetensors +3 -0
- model-00010-of-00017.safetensors +3 -0
- model-00011-of-00017.safetensors +3 -0
- model-00012-of-00017.safetensors +3 -0
- model-00013-of-00017.safetensors +3 -0
- model-00014-of-00017.safetensors +3 -0
- model-00015-of-00017.safetensors +3 -0
- model-00016-of-00017.safetensors +3 -0
- model-00017-of-00017.safetensors +3 -0
- model.safetensors.index.json +0 -0
- special_tokens_map.json +23 -0
- tokenizer.json +0 -0
- tokenizer.model +3 -0
- tokenizer_config.json +42 -0
@@ -0,0 +1,46 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
{
|
2 |
+
"_name_or_path": "v2ray/Mixtral-8x22B-v0.1",
|
3 |
+
"architectures": [
|
4 |
+
"MixtralForCausalLM"
|
5 |
+
],
|
6 |
+
"attention_dropout": 0.0,
|
7 |
+
"bos_token_id": 1,
|
8 |
+
"eos_token_id": 2,
|
9 |
+
"hidden_act": "silu",
|
10 |
+
"hidden_size": 6144,
|
11 |
+
"initializer_range": 0.02,
|
12 |
+
"intermediate_size": 16384,
|
13 |
+
"max_position_embeddings": 65536,
|
14 |
+
"model_type": "mixtral",
|
15 |
+
"num_attention_heads": 48,
|
16 |
+
"num_experts_per_tok": 2,
|
17 |
+
"num_hidden_layers": 56,
|
18 |
+
"num_key_value_heads": 8,
|
19 |
+
"num_local_experts": 8,
|
20 |
+
"output_router_logits": false,
|
21 |
+
"quantization_config": {
|
22 |
+
"_load_in_4bit": true,
|
23 |
+
"_load_in_8bit": false,
|
24 |
+
"bnb_4bit_compute_dtype": "float32",
|
25 |
+
"bnb_4bit_quant_storage": "uint8",
|
26 |
+
"bnb_4bit_quant_type": "fp4",
|
27 |
+
"bnb_4bit_use_double_quant": false,
|
28 |
+
"llm_int8_enable_fp32_cpu_offload": false,
|
29 |
+
"llm_int8_has_fp16_weight": false,
|
30 |
+
"llm_int8_skip_modules": null,
|
31 |
+
"llm_int8_threshold": 6.0,
|
32 |
+
"load_in_4bit": true,
|
33 |
+
"load_in_8bit": false,
|
34 |
+
"quant_method": "bitsandbytes"
|
35 |
+
},
|
36 |
+
"rms_norm_eps": 1e-05,
|
37 |
+
"rope_theta": 1000000,
|
38 |
+
"router_aux_loss_coef": 0.001,
|
39 |
+
"router_jitter_noise": 0.0,
|
40 |
+
"sliding_window": null,
|
41 |
+
"tie_word_embeddings": false,
|
42 |
+
"torch_dtype": "float16",
|
43 |
+
"transformers_version": "4.40.0.dev0",
|
44 |
+
"use_cache": true,
|
45 |
+
"vocab_size": 32000
|
46 |
+
}
|
@@ -0,0 +1,6 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
{
|
2 |
+
"_from_model_config": true,
|
3 |
+
"bos_token_id": 1,
|
4 |
+
"eos_token_id": 2,
|
5 |
+
"transformers_version": "4.40.0.dev0"
|
6 |
+
}
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:af5378c0a4bed3863f26e6b65558f37e4d824eb095eb389b7627f926308d7767
|
3 |
+
size 4951626618
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:b4e3ac4e6ba722998b646a20055b28bc5d96b458457e2d84f071fb53d540b3a0
|
3 |
+
size 4961824784
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:473456eedd4c4cdd8f2f228d180bef29e524fa3401db4c2aeeebf5418119fe7b
|
3 |
+
size 4954801840
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:887263a956a21493f7a5e7b8f5260d89a83ccd65986cc16763f0259685c2ff3d
|
3 |
+
size 4961825192
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:efa1e6d8e4ad38636e4ba8f11c38a50161cb85ed82d17fd8239d740e9547dfa7
|
3 |
+
size 4954802256
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:f8e70fa3c2bde8f5924224ba85f3256146c7ec585f3771634af108e4f32a4785
|
3 |
+
size 4961825192
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:0445dc8f86c2fcfb207a8077ab6218a7cef652682e3b32bbca91b0da2ecc8165
|
3 |
+
size 4954802256
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:fcc62ae79d78295adb99e570e788542d067b9caa1ac1bafb7877f7f4eb8be80b
|
3 |
+
size 4961825192
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:e9cd84e01652c02750189919683913d1821a57a44e1d1a481216c418df31e710
|
3 |
+
size 4954802256
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:42db7e99ec200b68d74e5fbefcbfb739bc9328475eeb9e9bacf3e1ed2da162b8
|
3 |
+
size 4961825192
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:84b63e7417ebd8a84ed6d47e8bdd748e3c3a3f8cc8d3e1719f0097db9c82222b
|
3 |
+
size 4954802256
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:f6974f9361f892ccb15a0869372aaf032525b0a77d117597a01c1c3696558773
|
3 |
+
size 4961825192
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:a54b97b419c07fe7612aa3b0ea0ff68d9fd7e84d9a87879b3c5b623f7d8c4df0
|
3 |
+
size 4954802256
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:e9cc813539c5a1e3593b13f8b3906576e4f08aea087260846242b445a914ea76
|
3 |
+
size 4990163459
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:fe8dd6215f686b5d83af6e5386ed07570db21d65ad747dac7f858b65357259a8
|
3 |
+
size 4983087791
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:e1e1553ebf287a0c6f2e481070887e0063346a474b71858e3d2b439335d67918
|
3 |
+
size 4848614772
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:5a9b656fa319bc8871482483a1345f0a9c4cf80cd53554668b81946e9b74223e
|
3 |
+
size 393216128
|
The diff for this file is too large to render.
See raw diff
|
|
@@ -0,0 +1,23 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
{
|
2 |
+
"bos_token": {
|
3 |
+
"content": "<s>",
|
4 |
+
"lstrip": false,
|
5 |
+
"normalized": false,
|
6 |
+
"rstrip": false,
|
7 |
+
"single_word": false
|
8 |
+
},
|
9 |
+
"eos_token": {
|
10 |
+
"content": "</s>",
|
11 |
+
"lstrip": false,
|
12 |
+
"normalized": false,
|
13 |
+
"rstrip": false,
|
14 |
+
"single_word": false
|
15 |
+
},
|
16 |
+
"unk_token": {
|
17 |
+
"content": "<unk>",
|
18 |
+
"lstrip": false,
|
19 |
+
"normalized": false,
|
20 |
+
"rstrip": false,
|
21 |
+
"single_word": false
|
22 |
+
}
|
23 |
+
}
|
The diff for this file is too large to render.
See raw diff
|
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:dadfd56d766715c61d2ef780a525ab43b8e6da4de6865bda3d95fdef5e134055
|
3 |
+
size 493443
|
@@ -0,0 +1,42 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
{
|
2 |
+
"add_bos_token": true,
|
3 |
+
"add_eos_token": false,
|
4 |
+
"added_tokens_decoder": {
|
5 |
+
"0": {
|
6 |
+
"content": "<unk>",
|
7 |
+
"lstrip": false,
|
8 |
+
"normalized": false,
|
9 |
+
"rstrip": false,
|
10 |
+
"single_word": false,
|
11 |
+
"special": true
|
12 |
+
},
|
13 |
+
"1": {
|
14 |
+
"content": "<s>",
|
15 |
+
"lstrip": false,
|
16 |
+
"normalized": false,
|
17 |
+
"rstrip": false,
|
18 |
+
"single_word": false,
|
19 |
+
"special": true
|
20 |
+
},
|
21 |
+
"2": {
|
22 |
+
"content": "</s>",
|
23 |
+
"lstrip": false,
|
24 |
+
"normalized": false,
|
25 |
+
"rstrip": false,
|
26 |
+
"single_word": false,
|
27 |
+
"special": true
|
28 |
+
}
|
29 |
+
},
|
30 |
+
"additional_special_tokens": [],
|
31 |
+
"bos_token": "<s>",
|
32 |
+
"clean_up_tokenization_spaces": false,
|
33 |
+
"eos_token": "</s>",
|
34 |
+
"legacy": true,
|
35 |
+
"model_max_length": 1000000000000000019884624838656,
|
36 |
+
"pad_token": null,
|
37 |
+
"sp_model_kwargs": {},
|
38 |
+
"spaces_between_special_tokens": false,
|
39 |
+
"tokenizer_class": "LlamaTokenizer",
|
40 |
+
"unk_token": "<unk>",
|
41 |
+
"use_default_system_prompt": false
|
42 |
+
}
|