MaziyarPanahi commited on
Commit
1e379c2
1 Parent(s): 20c9270

Upload folder using huggingface_hub (#3)

Browse files

- 265500d51aa1684c0156b02045fdcf0f9c4f4f44a530a7924ce95e9e872ea959 (74fdc4a3853d02715686da30d39c67d00f333f77)
- 6a15be82ea3b8d48956c832ba1bfabb8334702324e34eea710262de8bdf62799 (85bda76a624d6ff835532f2ffada220537980b0c)
- 10dbb411e9cae7c3ef9a5b759871085a6b7ae45e77ace452f48de203470685e5 (e5d3063eb558fd74efb55335bdf5dee02bb03520)
- 7d19c84fab639bcd4fd4c6149ecfc1e30c828218cee7b3fdff56d0570a7fcf63 (0639f90d4ac9b00427737fa4e062332c868a07c9)
- b419387ec633e778d2d5878cf2e687badaaa439be5727e133499e170db3c1a2d (668085c55b23e5e32d762db574b8af1457d4046a)
- 4be4ce8caecf449bcd1cd5adc35b33d96d0e4d1e2f26826b037b6f5f1ba2bb68 (eeca26939627d9c43aa636501b96584b38634921)
- 515616968a067ea7c18c5257fdd9cda1a24931fe7193cc2788f8615c19ebb918 (9930aebcb64b6dd7ecd2100de6a33b1e3973fcc7)
- e84c84a94fc18aaf8c98d6cc00ecf62a5bc0269e58e65911c258051b9a6a48f9 (2c65c24fa3dabc1b3fbc0510fe6257e4bdaa7feb)
- ea0280efe8c8a83cea7681ff746d5b88c832c51739d3ad6dcc7bbc1e988b4939 (edf63aad4822bddcbc208d9cc607f654d3ef8a60)
- 54dbdef946978f5584f19d8e87b9c4bc7e28954d04574c22c244e4d6f07ccb8e (e9780e39cb7716a83108cd12353fb9b881cf0f0e)
- 5284a5ff5a881cc90b0c050a944cc6154205edac995250fdbe9a498722c5ab4f (7822b3fe6c61838219e1a200a175d3993c2d5c99)
- d28fa0cc6abcaf641af643a6afc24a47a058e6e0076d1c43d5ae228fd9d7aaaf (f8d31f3df227fdab30a5080ff6668ddb828d098f)
- b71e0ef15a9358cdc7348dcbd4b30d8736472111cd9b3e835bd6d17e6b971bf3 (b150ca832b31fa31ecc62032c6e251f7f5df1442)
- 35b2589768aef0cfc55ee15b3d8ce3128484a337b2d8260a06dfad0d7e3fdf3c (7f238f858437223fb2ea7afa5023b035fa076224)
- 84e4594798891bee222ce06470d259a9e7b087aeb8edf42c1db61d9553761635 (441ad9546e9c7c9cb27476ac9b5125b20a3f2bf7)
- 936b2bce08bccaef57af291c42e79f24c500e68840dc2bc189ccc1224fa2632a (1c5c49644f0336522a1a80209be2ab05229a3c43)
- 180b3eb5c3b03fb8b40f56cd5f2f490d48e4ac67664b115d22a26b955bdd3216 (150bd73d16607e532d3947077409cb40aa14a991)
- 8538dbe6d6c4ac7e576ea9ce8542eaeb4fb09eacd459d1209ad17f0105ff11ae (2ca0aec77a2b7e6d2d78890dffee2de729204aca)
- 1950ddba285e8f0522abd8800c308d7837fec43273f8f80549f07146260e7dd7 (a418ddc670fb55d4ab80c6f4eb82b2997329dc14)
- e705428137ad96b26693de6e290ebefffdaa7f77d1531c25edbce5eded575c45 (dfe4700d54b02357580fc7c3090717ffdaf8bea0)
- 3aa0e0e5ba25387b6d1b6c44f41049dc6c6086dbda55b8c6f459dd0ffc1109d4 (873f137f0d6d9525fa55793f77b346c879a8ac5d)
- 977702fae6aff2b34bfc7ff0de677f21576e1c080a142a71d17be0e814c80663 (28ed1366119ee637ba65f2e10265f89554212b49)
- ac42c7e1d4714065e468c38d472a16ae86868ca2c1fdd1320851800fbfcd0671 (c3a6009c817e714ad3d458f2d8c0f3ffe9f52ca2)
- 47dc9d77d3f4393291b27df6f8de4437226728c6848ac65dd48e7fed9013032e (fe08e2314fc5014472a5f3d4a69d8a02b87aab42)
- 0b402b5ed3e7085d5eb76108b0b5546860fffa5348f3ed46daa5c0d2a91c2829 (ac74addf1d2d5ca6b43f996ca27520d624ea939d)
- 40dcea8d5f8bac105f2aca70d82ba428dfb5635d3b36cadcbb7268ef56effca0 (1eb8766031d92ee1efb99feed3ea8b12c43316fe)
- 10f7f501a1561ffc9cb51bf7d7c471d36ccf3355862f6db5f14f4ad4f2c4d09c (6d8600afc5d22d45905b6e602a63e48af7986fbf)
- 8254accd300db6ded49452236a02f44f005801ea7958d9e501b5fa61bb73077e (ee2f95dbd1bfe264f23398b1ceff99f50c1ffe46)
- 1b36baf5f412350e692577906f7754e8db0bf5c3a87f324e8cf1bd1ace1dd962 (07815a9c86aa4bf62968b7e4a3fe72d6d06e23cc)
- a0b48f7e0f40fc8b2a5d56b322b16b9c9ef2e31d421e735998ccece88ad28dc2 (3e382381e20ececb3d2293fd213524f9bf1508c5)
- e4952920eb20d6eb99557dd7cc6616898d5c86af3e69d0631c5e3e45bf130514 (4d083de32f72db902fe2d1768e5cd50245a26b22)
- af0966cbd01f80cad678561cfa73d0712678ae819a774cd7a9a6113a2deb93fa (bafa5fab55c85a436b56c9835fb4dec7c543db21)
- d030214df47b23259fe7f291f273825667e43c16280c57bb43b266b340d554d1 (a8ea4c82ef1742d8ef017ccf2f750af6e226a445)
- e432682700a0eabd09e3cd0ac30b318c6d50966ba7f207344544379919ecea87 (6d487fb087b6f6aaed0e01fb6251189aef502f84)
- 03f58da15a1a1b175e420a4d0dfbaddd7b258d8ed358a2527dea7d5a9635e41e (c09eb9b7c43fb816c7ded68ef0b82596c747cf52)
- 261809da1ccb62ae6f71945125736b1730a2a51b1c2f97e8dcbd2023cadfab8e (a43092d5f335dfe83871b146ab8295fae2a74c4b)
- 4197cde6ba448d656ba0806f27ea31ffa1e895e26b12fb7cab7c060b411b70d8 (b42200d91747a5d1aaf0a0b81369ab335b6692a1)
- 13a5d5097ea9ae8279d49f90316fc9002ac3317764497883526d2c21c85b2291 (13d74b6ffbc7e806a178891cac784aa4f9cd6ee7)
- d4064a22e88a8bb6d5f1fa30895560e7c2b280dfc9e0af2562c22d3e803553e7 (cc21865cdf3cbff4d7ab1687d985ea26b36d2be3)
- 5d2d8d0018dc93c259d2cafa3d2159c6e27a760b06ba02f7959ab85aaa5ca2f4 (08b405bf046822326a9128e28b1fc399ac4ba4a7)
- e0a1c66adf8e78db96b6ab68a0828431219deade6eac026429e4b8892bed6cbc (912f092f72bcfa20a99b3c014fe54fa0d023c21d)
- 359e62a435e461fcf3e940a947de99efe9e80c4b33fbe97ac30faee4ad5ebb50 (99a0b2d4fda007f6b1eb898b666a726327c3d467)
- 97da9a01c7dd22171aff3bfd02a424f5996172b00c83f26fccb8dc16fd212455 (bbe9ab7682419e5c9787e4800a3a096ae9cb713c)
- 9bd937d962f79a1af3e7e8797eeccde806008639adc3f2b106e46920a7838837 (c557427678725c014afdd66506b2a6c83af8faa6)
- 99c0af5e496d74ee7848f7a0e9c5ad31bb809555b9e775309fdeec9b45ce173f (333f05d1ff0ebdc18f8e46bc5b8411d9a955d674)
- 5a6008ff1e206e514d2af805bfc0a867ab0365cd807eeb80af52dd47a12f4de0 (04cd330521bbebc865fe8ea0bb43b8ec4a8ec082)
- e21c0c3a96649e77a2c932210e8e51acc5c8136b2b0d36e273a5c3b3a7381c5c (46782fd95750cfb345b50ce60d61d8deff75596b)
- bd50aec17576ae11ff4d0e9578c095b0338e8f3061076a085b76fa6a0dd85868 (1a056fff3f252223f207d75d0ffcb8143f5a4665)
- 56782fd1fcc27ca5a86ee026ff74a514ece7d314bcd1304dfbe3a9e7a639ac0c (892306151e3d9303058ba5a0ccf0a6359c62f084)
- f1e114654f84914e6c9feecf06a669e2e42c933788aeeef40d393dcd33310e6c (c8c5f6f269e5c441d1820c5b836b60d706a56e8d)
- 4922e9627a854a34ba68e1747e554108eb793b306a20fd9763a182ccb7ff0170 (8bfc572ec728c0f9e6c58125c83ded94df4d1d4d)
- 477bc0850f151fe9ff4fc421b33294d95f277262123812288a74eb0d55e7ea59 (ee15308f93e54309151ca49a67a67a788b7cb0bf)
- 183760ade2caedfcc254333dbd25e0760b88a6491f9ae8e75fdc297c2b2734ff (3e3b6c1c4ddcc583e3bcc75e0ad2539df591e3f3)
- e53fdee8235a09e9493934bee757f78cb4b1e167cc04a2f7a30ace2852c7b964 (b46e5b9a4cb58db8020352d9c51b976ed390122f)
- 037c1183f3faf38590f07dce7023ef6dc0938cb90f5d1361d39519e4ed906089 (848805a9160a2182440b97ba09190dc5268f5deb)
- 9b5cecefabd192bde4342f467e5486ab6817e8439a0300f2ffdfe633e72bf30b (d5474f7e4ddbc44bfa5dde6267ca9de36814fc66)
- 542578164691a4c93a82365abb55e2c8065181fd5f3018f490a51e21df81b878 (aa4c7fa81bd1264ad05e531906c39e952679936a)
- d5d64f3d2969163c25f172dfbe6640f3c654a60444d349d0a5d41f6d140b2b7c (56b381b273a2df3216d823e3e34a05312111b894)
- 627a971a7fc5730f9ec6daeb95051eba776828a048fba3ea0648cad7247e8cb2 (355a78344bb7f9cc0180ee17d8dbfbf5e1f9a9d6)
- bcc788e13f079ac25b6457a4fc9abc69eb72b4971a193856bb4ac2129643c064 (1a11d78ff48e5d40e5cd95c12789313114da342c)
- 43cd993498ef406118513539aa61541ffe133c7bc7680dfd7336b541097086f5 (42eabf44528f7132b3a8e8f6f4dca671d5d41ed3)
- 1f886a92507482d9d8b9e16146c95a71416b65c8aea3034c5a359966391b1dee (c0008423a75dd56d02a96897b4aa3840551d251a)
- e7d62887cdbbe94086f3398cbcb988126a619404d595661b217a8a49b36844d5 (27f08525ba3ab1f5c7e6cb201bfbf1617d52c5d3)
- 961276326e4ca1ca7ef379c32708713845cf92a16621e93b8092d0c82bf92546 (d19e0a295d8f8634146529591813817178ff6029)
- c45e50f03181c47ee5c6f467cb48d9c8f4bf2ecce707725aed71c234a2cd4034 (bdaf5718db6ec54ea6895e041432d07833e1e6ce)
- 6added3eb66c8a0583f8398f3533c407dc1e0a238dd9691f7f7e268967aa3e2c (5256f40f2d8e5e28044338421570fd4dbd73e0ec)
- a637cd823977660efa04df9d1b2e6781f089805edafe0064c399c368da62bb73 (e51f1f331cc236a9aaa3ef03db24bfaac01d0afb)
- 3b24c08ed385177c0215b61e513ea0c0fe080657f1c2e67d757dddada48ac461 (ac2dd23b0b61013db7a48143dc30dbaa38cb2063)
- 84acfecf13b41ce15b68dac6df1f50980f59542a6569e164b65938c48fddb033 (533a261b7722a8fe9ade69b17638e2a6c5af8364)
- e558ab781199610a5ac37742379d7e7fba8db478716a2be8d334fcd0579ba288 (56852d9b81dcad20bc842b8ac1321e359f6942a4)
- 121b64dc68801dc781b8496fe69b97234e4eca8054f53129a7140984a7eef908 (bb869123cace296a4afed9ba3c4546cf6563015c)
- fe53749dd7e211c178c31191f816891b42554126ebd72751142f6d7539990092 (106c80beed0f41ac8015372cb47c5c712b41577a)
- 926ec18d07c76c835ebb9e6956ab73ac67cb07960894cf695cb897897cebb203 (a6231ac3a7e658cac1e7ec87669448ab55cde96b)
- 68459da06ec26022db799281b6e0baf53afa074f9dbc14d6dc98d287c949ce74 (4350274def37a48e48e0599147c24dc9ac79b58b)
- 1af04d58715bfa10cd3926fee300ebfa13c46614850095557915a6b4b4df595f (1046583596bbff97b8b2753a84d0c0514687b88b)
- 84476f4f99ef30734f0b9b55dc7bd45abe5a993d29a447850934f350bafef082 (e6b0adba028c07452d6cec8cdb674dea878f4573)
- 57e3b105b089e355a5706c86bad1da575086228cf5b48711a459528210416dd2 (299028ac960e48dbeea895f096dad2594087697d)
- d4e0f8707ad181cc04876caed1b0a3894916c2915291c69ec46119e8312bb1c4 (e337bb755aefae271b71ec6828d0c0bf4c0ca3ae)
- a0f40e63fc954d5dbe57401729aeeed62acd0c38fb18e78179bcd25e974873df (76a1e6435c99fd9bc01fe3e9472505a6020f5390)
- 34baea6cd6c9f0dab2e9021687c30b336141272e4991b02ed64e18192809eeca (26a8562ca30ec7b671b335d0b7e3ddf16353bddb)
- 9f92274ac959487f827a763009954b01606dd5db6cb248639b8a680bd9f5946d (2f5c46143a48dfd4029bea0269c244673e71c606)
- 136340078d1e6ac8895018b51dc4c0866e44fe5be6d5fcf8d818090bc51cc758 (ba69586a8c5e748e6002844016626445de9f7318)
- 1af1bb5d47390f861b698b5a4cb1442873269b18520d03ab4b85497ad0c5c162 (0398cbad6d9045daa205778e3266fbacd1fdd5ca)
- ce000b7ef1601646f2c46ec00ca2a72c2b89a046c98998b0b5d771ea3366f7f6 (0a7c13fb909d9770e7c9c4e9fa161393557d1a72)
- 2826e12be021bdfe8cc2d4c6551b1fa8cb623574fbc37986f6bb4e49904102e2 (d06d02498827eb94c6bd4cfc51ab6654a13e3cb1)
- b68cee6b6f60d28ae746d675c27fdea4635e0f2b7ef1165c3e34778a41bd7cd7 (d8a9118df58567d7284f39d8b231df8b9278aa54)
- fe97f1bed10b3ae1ea1a83dffdf288073a89a2ac807b4db7210c24b8f359b2b0 (a6d9ff33d1a166fc4d0c8f97c130e8dce638e5df)
- e24416708a2c23ce65f7229dca6cd68fd9494537c234ad1a128a67bb6d1a7e92 (c4f2599765375b9e3d7592c7833862256b7f6aec)
- c1ee6c0cc4007487d28c1e338756f92bec679c83156a402d57736609dd5cf6e8 (5918040edec0a4238125b1a91b33727bebc7334c)
- 3f4bc800280d9215737299219d3d693bcdd0fe8fe82a83501fb60ff9f6a52d44 (b79d885b8b3de394dfa3df99bf41b80fa859bfda)
- 693567ef0b67c1b354d48afbf65e00813899d05a0fd7fbd15f57416341de3b66 (4ad9e043ad271c39ed7900abc77a3a23ab67ad49)
- 6e216a6f21df82b512767db8befd799fc83e4e8b3159c25ab885ba9542fbab3b (02e6b6abeeb2900e676076bb885708e6c317ebe3)
- f035ab14d8255c8c38bfcc4b368715f012893e3c2e0bba41fb7a3298df3abba3 (eb56f52b3

This view is limited to 50 files because it contains too many changes.   See raw diff
Files changed (50) hide show
  1. .ipynb_checkpoints/config-checkpoint.json +31 -0
  2. .ipynb_checkpoints/tokenizer_config-checkpoint.json +61 -0
  3. adapter/.ipynb_checkpoints/config-checkpoint.json +46 -0
  4. adapter/.ipynb_checkpoints/tokenizer_config-checkpoint.json +61 -0
  5. adapter/adapter_config.json +38 -0
  6. adapter/adapter_model.safetensors +3 -0
  7. adapter/added_tokens.json +4 -0
  8. adapter/config.json +46 -0
  9. adapter/special_tokens_map.json +24 -0
  10. adapter/tokenizer.model +3 -0
  11. adapter/tokenizer_config.json +61 -0
  12. added_tokens.json +4 -0
  13. config.json +31 -0
  14. generation_config.json +7 -0
  15. model-00001-of-00117.safetensors +3 -0
  16. model-00002-of-00117.safetensors +3 -0
  17. model-00003-of-00117.safetensors +3 -0
  18. model-00004-of-00117.safetensors +3 -0
  19. model-00005-of-00117.safetensors +3 -0
  20. model-00006-of-00117.safetensors +3 -0
  21. model-00007-of-00117.safetensors +3 -0
  22. model-00008-of-00117.safetensors +3 -0
  23. model-00009-of-00117.safetensors +3 -0
  24. model-00010-of-00117.safetensors +3 -0
  25. model-00011-of-00117.safetensors +3 -0
  26. model-00012-of-00117.safetensors +3 -0
  27. model-00013-of-00117.safetensors +3 -0
  28. model-00014-of-00117.safetensors +3 -0
  29. model-00015-of-00117.safetensors +3 -0
  30. model-00016-of-00117.safetensors +3 -0
  31. model-00017-of-00117.safetensors +3 -0
  32. model-00018-of-00117.safetensors +3 -0
  33. model-00019-of-00117.safetensors +3 -0
  34. model-00020-of-00117.safetensors +3 -0
  35. model-00021-of-00117.safetensors +3 -0
  36. model-00022-of-00117.safetensors +3 -0
  37. model-00023-of-00117.safetensors +3 -0
  38. model-00024-of-00117.safetensors +3 -0
  39. model-00025-of-00117.safetensors +3 -0
  40. model-00026-of-00117.safetensors +3 -0
  41. model-00027-of-00117.safetensors +3 -0
  42. model-00028-of-00117.safetensors +3 -0
  43. model-00029-of-00117.safetensors +3 -0
  44. model-00030-of-00117.safetensors +3 -0
  45. model-00031-of-00117.safetensors +3 -0
  46. model-00032-of-00117.safetensors +3 -0
  47. model-00033-of-00117.safetensors +3 -0
  48. model-00034-of-00117.safetensors +3 -0
  49. model-00035-of-00117.safetensors +3 -0
  50. model-00036-of-00117.safetensors +3 -0
.ipynb_checkpoints/config-checkpoint.json ADDED
@@ -0,0 +1,31 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "_name_or_path": "v2ray/Mixtral-8x22B-v0.1",
3
+ "architectures": [
4
+ "MixtralForCausalLM"
5
+ ],
6
+ "attention_dropout": 0.0,
7
+ "bos_token_id": 1,
8
+ "eos_token_id": 32000,
9
+ "hidden_act": "silu",
10
+ "hidden_size": 6144,
11
+ "initializer_range": 0.02,
12
+ "intermediate_size": 16384,
13
+ "max_position_embeddings": 65536,
14
+ "model_type": "mixtral",
15
+ "num_attention_heads": 48,
16
+ "num_experts_per_tok": 2,
17
+ "num_hidden_layers": 56,
18
+ "num_key_value_heads": 8,
19
+ "num_local_experts": 8,
20
+ "output_router_logits": false,
21
+ "rms_norm_eps": 1e-05,
22
+ "rope_theta": 1000000,
23
+ "router_aux_loss_coef": 0.001,
24
+ "router_jitter_noise": 0.0,
25
+ "sliding_window": null,
26
+ "tie_word_embeddings": false,
27
+ "torch_dtype": "float32",
28
+ "transformers_version": "4.40.0.dev0",
29
+ "use_cache": false,
30
+ "vocab_size": 32002
31
+ }
.ipynb_checkpoints/tokenizer_config-checkpoint.json ADDED
@@ -0,0 +1,61 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "add_bos_token": true,
3
+ "add_eos_token": false,
4
+ "add_prefix_space": true,
5
+ "added_tokens_decoder": {
6
+ "0": {
7
+ "content": "<unk>",
8
+ "lstrip": false,
9
+ "normalized": false,
10
+ "rstrip": false,
11
+ "single_word": false,
12
+ "special": true
13
+ },
14
+ "1": {
15
+ "content": "<s>",
16
+ "lstrip": false,
17
+ "normalized": false,
18
+ "rstrip": false,
19
+ "single_word": false,
20
+ "special": true
21
+ },
22
+ "2": {
23
+ "content": "</s>",
24
+ "lstrip": false,
25
+ "normalized": false,
26
+ "rstrip": false,
27
+ "single_word": false,
28
+ "special": true
29
+ },
30
+ "32000": {
31
+ "content": "<|im_end|>",
32
+ "lstrip": false,
33
+ "normalized": false,
34
+ "rstrip": false,
35
+ "single_word": false,
36
+ "special": true
37
+ },
38
+ "32001": {
39
+ "content": "<|im_start|>",
40
+ "lstrip": false,
41
+ "normalized": false,
42
+ "rstrip": false,
43
+ "single_word": false,
44
+ "special": false
45
+ }
46
+ },
47
+ "additional_special_tokens": [],
48
+ "bos_token": "<s>",
49
+ "chat_template": "{% if not add_generation_prompt is defined %}{% set add_generation_prompt = false %}{% endif %}{% for message in messages %}{{'<|im_start|>' + message['role'] + '\n' + message['content'] + '<|im_end|>' + '\n'}}{% endfor %}{% if add_generation_prompt %}{{ '<|im_start|>assistant\n' }}{% endif %}",
50
+ "clean_up_tokenization_spaces": false,
51
+ "eos_token": "<|im_end|>",
52
+ "legacy": true,
53
+ "model_max_length": 1000000000000000019884624838656,
54
+ "pad_token": "</s>",
55
+ "sp_model_kwargs": {},
56
+ "spaces_between_special_tokens": false,
57
+ "tokenizer_class": "LlamaTokenizer",
58
+ "unk_token": "<unk>",
59
+ "use_default_system_prompt": false,
60
+ "use_fast": true
61
+ }
adapter/.ipynb_checkpoints/config-checkpoint.json ADDED
@@ -0,0 +1,46 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "_name_or_path": "v2ray/Mixtral-8x22B-v0.1",
3
+ "architectures": [
4
+ "MixtralForCausalLM"
5
+ ],
6
+ "attention_dropout": 0.0,
7
+ "bos_token_id": 1,
8
+ "eos_token_id": 32000,
9
+ "hidden_act": "silu",
10
+ "hidden_size": 6144,
11
+ "initializer_range": 0.02,
12
+ "intermediate_size": 16384,
13
+ "max_position_embeddings": 65536,
14
+ "model_type": "mixtral",
15
+ "num_attention_heads": 48,
16
+ "num_experts_per_tok": 2,
17
+ "num_hidden_layers": 56,
18
+ "num_key_value_heads": 8,
19
+ "num_local_experts": 8,
20
+ "output_router_logits": false,
21
+ "quantization_config": {
22
+ "_load_in_4bit": true,
23
+ "_load_in_8bit": false,
24
+ "bnb_4bit_compute_dtype": "bfloat16",
25
+ "bnb_4bit_quant_storage": "bfloat16",
26
+ "bnb_4bit_quant_type": "nf4",
27
+ "bnb_4bit_use_double_quant": true,
28
+ "llm_int8_enable_fp32_cpu_offload": false,
29
+ "llm_int8_has_fp16_weight": false,
30
+ "llm_int8_skip_modules": null,
31
+ "llm_int8_threshold": 6.0,
32
+ "load_in_4bit": true,
33
+ "load_in_8bit": false,
34
+ "quant_method": "bitsandbytes"
35
+ },
36
+ "rms_norm_eps": 1e-05,
37
+ "rope_theta": 1000000,
38
+ "router_aux_loss_coef": 0.001,
39
+ "router_jitter_noise": 0.0,
40
+ "sliding_window": null,
41
+ "tie_word_embeddings": false,
42
+ "torch_dtype": "bfloat16",
43
+ "transformers_version": "4.40.0.dev0",
44
+ "use_cache": false,
45
+ "vocab_size": 32002
46
+ }
adapter/.ipynb_checkpoints/tokenizer_config-checkpoint.json ADDED
@@ -0,0 +1,61 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "add_bos_token": true,
3
+ "add_eos_token": false,
4
+ "add_prefix_space": true,
5
+ "added_tokens_decoder": {
6
+ "0": {
7
+ "content": "<unk>",
8
+ "lstrip": false,
9
+ "normalized": false,
10
+ "rstrip": false,
11
+ "single_word": false,
12
+ "special": true
13
+ },
14
+ "1": {
15
+ "content": "<s>",
16
+ "lstrip": false,
17
+ "normalized": false,
18
+ "rstrip": false,
19
+ "single_word": false,
20
+ "special": true
21
+ },
22
+ "2": {
23
+ "content": "</s>",
24
+ "lstrip": false,
25
+ "normalized": false,
26
+ "rstrip": false,
27
+ "single_word": false,
28
+ "special": true
29
+ },
30
+ "32000": {
31
+ "content": "<|im_end|>",
32
+ "lstrip": false,
33
+ "normalized": false,
34
+ "rstrip": false,
35
+ "single_word": false,
36
+ "special": true
37
+ },
38
+ "32001": {
39
+ "content": "<|im_start|>",
40
+ "lstrip": false,
41
+ "normalized": false,
42
+ "rstrip": false,
43
+ "single_word": false,
44
+ "special": false
45
+ }
46
+ },
47
+ "additional_special_tokens": [],
48
+ "bos_token": "<s>",
49
+ "chat_template": "{% if not add_generation_prompt is defined %}{% set add_generation_prompt = false %}{% endif %}{% for message in messages %}{{'<|im_start|>' + message['role'] + '\n' + message['content'] + '<|im_end|>' + '\n'}}{% endfor %}{% if add_generation_prompt %}{{ '<|im_start|>assistant\n' }}{% endif %}",
50
+ "clean_up_tokenization_spaces": false,
51
+ "eos_token": "<|im_end|>",
52
+ "legacy": true,
53
+ "model_max_length": 1000000000000000019884624838656,
54
+ "pad_token": "</s>",
55
+ "sp_model_kwargs": {},
56
+ "spaces_between_special_tokens": false,
57
+ "tokenizer_class": "LlamaTokenizer",
58
+ "unk_token": "<unk>",
59
+ "use_default_system_prompt": false,
60
+ "use_fast": true
61
+ }
adapter/adapter_config.json ADDED
@@ -0,0 +1,38 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "alpha_pattern": {},
3
+ "auto_mapping": null,
4
+ "base_model_name_or_path": "v2ray/Mixtral-8x22B-v0.1",
5
+ "bias": "none",
6
+ "fan_in_fan_out": null,
7
+ "inference_mode": true,
8
+ "init_lora_weights": true,
9
+ "layer_replication": null,
10
+ "layers_pattern": null,
11
+ "layers_to_transform": null,
12
+ "loftq_config": {},
13
+ "lora_alpha": 8,
14
+ "lora_dropout": 0.05,
15
+ "megatron_config": null,
16
+ "megatron_core": "megatron.core",
17
+ "modules_to_save": [
18
+ "embed_tokens",
19
+ "lm_head"
20
+ ],
21
+ "peft_type": "LORA",
22
+ "r": 16,
23
+ "rank_pattern": {},
24
+ "revision": null,
25
+ "target_modules": [
26
+ "q_proj",
27
+ "w2",
28
+ "w3",
29
+ "o_proj",
30
+ "v_proj",
31
+ "k_proj",
32
+ "gate",
33
+ "w1"
34
+ ],
35
+ "task_type": "CAUSAL_LM",
36
+ "use_dora": false,
37
+ "use_rslora": false
38
+ }
adapter/adapter_model.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:e7c3e6eac571b2034dd287ea2c037b3187e9ea2a77b00a0166a08708fc006e87
3
+ size 3409573984
adapter/added_tokens.json ADDED
@@ -0,0 +1,4 @@
 
 
 
 
 
1
+ {
2
+ "<|im_end|>": 32000,
3
+ "<|im_start|>": 32001
4
+ }
adapter/config.json ADDED
@@ -0,0 +1,46 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "_name_or_path": "v2ray/Mixtral-8x22B-v0.1",
3
+ "architectures": [
4
+ "MixtralForCausalLM"
5
+ ],
6
+ "attention_dropout": 0.0,
7
+ "bos_token_id": 1,
8
+ "eos_token_id": 32000,
9
+ "hidden_act": "silu",
10
+ "hidden_size": 6144,
11
+ "initializer_range": 0.02,
12
+ "intermediate_size": 16384,
13
+ "max_position_embeddings": 65536,
14
+ "model_type": "mixtral",
15
+ "num_attention_heads": 48,
16
+ "num_experts_per_tok": 2,
17
+ "num_hidden_layers": 56,
18
+ "num_key_value_heads": 8,
19
+ "num_local_experts": 8,
20
+ "output_router_logits": false,
21
+ "quantization_config": {
22
+ "_load_in_4bit": true,
23
+ "_load_in_8bit": false,
24
+ "bnb_4bit_compute_dtype": "bfloat16",
25
+ "bnb_4bit_quant_storage": "bfloat16",
26
+ "bnb_4bit_quant_type": "nf4",
27
+ "bnb_4bit_use_double_quant": true,
28
+ "llm_int8_enable_fp32_cpu_offload": false,
29
+ "llm_int8_has_fp16_weight": false,
30
+ "llm_int8_skip_modules": null,
31
+ "llm_int8_threshold": 6.0,
32
+ "load_in_4bit": true,
33
+ "load_in_8bit": false,
34
+ "quant_method": "bitsandbytes"
35
+ },
36
+ "rms_norm_eps": 1e-05,
37
+ "rope_theta": 1000000,
38
+ "router_aux_loss_coef": 0.001,
39
+ "router_jitter_noise": 0.0,
40
+ "sliding_window": null,
41
+ "tie_word_embeddings": false,
42
+ "torch_dtype": "bfloat16",
43
+ "transformers_version": "4.40.0.dev0",
44
+ "use_cache": false,
45
+ "vocab_size": 32002
46
+ }
adapter/special_tokens_map.json ADDED
@@ -0,0 +1,24 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "bos_token": {
3
+ "content": "<s>",
4
+ "lstrip": false,
5
+ "normalized": false,
6
+ "rstrip": false,
7
+ "single_word": false
8
+ },
9
+ "eos_token": {
10
+ "content": "<|im_end|>",
11
+ "lstrip": false,
12
+ "normalized": false,
13
+ "rstrip": false,
14
+ "single_word": false
15
+ },
16
+ "pad_token": "</s>",
17
+ "unk_token": {
18
+ "content": "<unk>",
19
+ "lstrip": false,
20
+ "normalized": false,
21
+ "rstrip": false,
22
+ "single_word": false
23
+ }
24
+ }
adapter/tokenizer.model ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:dadfd56d766715c61d2ef780a525ab43b8e6da4de6865bda3d95fdef5e134055
3
+ size 493443
adapter/tokenizer_config.json ADDED
@@ -0,0 +1,61 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "add_bos_token": true,
3
+ "add_eos_token": false,
4
+ "add_prefix_space": true,
5
+ "added_tokens_decoder": {
6
+ "0": {
7
+ "content": "<unk>",
8
+ "lstrip": false,
9
+ "normalized": false,
10
+ "rstrip": false,
11
+ "single_word": false,
12
+ "special": true
13
+ },
14
+ "1": {
15
+ "content": "<s>",
16
+ "lstrip": false,
17
+ "normalized": false,
18
+ "rstrip": false,
19
+ "single_word": false,
20
+ "special": true
21
+ },
22
+ "2": {
23
+ "content": "</s>",
24
+ "lstrip": false,
25
+ "normalized": false,
26
+ "rstrip": false,
27
+ "single_word": false,
28
+ "special": true
29
+ },
30
+ "32000": {
31
+ "content": "<|im_end|>",
32
+ "lstrip": false,
33
+ "normalized": false,
34
+ "rstrip": false,
35
+ "single_word": false,
36
+ "special": true
37
+ },
38
+ "32001": {
39
+ "content": "<|im_start|>",
40
+ "lstrip": false,
41
+ "normalized": false,
42
+ "rstrip": false,
43
+ "single_word": false,
44
+ "special": false
45
+ }
46
+ },
47
+ "additional_special_tokens": [],
48
+ "bos_token": "<s>",
49
+ "chat_template": "{% if not add_generation_prompt is defined %}{% set add_generation_prompt = false %}{% endif %}{% for message in messages %}{{'<|im_start|>' + message['role'] + '\n' + message['content'] + '<|im_end|>' + '\n'}}{% endfor %}{% if add_generation_prompt %}{{ '<|im_start|>assistant\n' }}{% endif %}",
50
+ "clean_up_tokenization_spaces": false,
51
+ "eos_token": "<|im_end|>",
52
+ "legacy": true,
53
+ "model_max_length": 1000000000000000019884624838656,
54
+ "pad_token": "</s>",
55
+ "sp_model_kwargs": {},
56
+ "spaces_between_special_tokens": false,
57
+ "tokenizer_class": "LlamaTokenizer",
58
+ "unk_token": "<unk>",
59
+ "use_default_system_prompt": false,
60
+ "use_fast": true
61
+ }
added_tokens.json ADDED
@@ -0,0 +1,4 @@
 
 
 
 
 
1
+ {
2
+ "<|im_end|>": 32000,
3
+ "<|im_start|>": 32001
4
+ }
config.json ADDED
@@ -0,0 +1,31 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "_name_or_path": "v2ray/Mixtral-8x22B-v0.1",
3
+ "architectures": [
4
+ "MixtralForCausalLM"
5
+ ],
6
+ "attention_dropout": 0.0,
7
+ "bos_token_id": 1,
8
+ "eos_token_id": 32000,
9
+ "hidden_act": "silu",
10
+ "hidden_size": 6144,
11
+ "initializer_range": 0.02,
12
+ "intermediate_size": 16384,
13
+ "max_position_embeddings": 65536,
14
+ "model_type": "mixtral",
15
+ "num_attention_heads": 48,
16
+ "num_experts_per_tok": 2,
17
+ "num_hidden_layers": 56,
18
+ "num_key_value_heads": 8,
19
+ "num_local_experts": 8,
20
+ "output_router_logits": false,
21
+ "rms_norm_eps": 1e-05,
22
+ "rope_theta": 1000000,
23
+ "router_aux_loss_coef": 0.001,
24
+ "router_jitter_noise": 0.0,
25
+ "sliding_window": null,
26
+ "tie_word_embeddings": false,
27
+ "torch_dtype": "float32",
28
+ "transformers_version": "4.40.0.dev0",
29
+ "use_cache": false,
30
+ "vocab_size": 32002
31
+ }
generation_config.json ADDED
@@ -0,0 +1,7 @@
 
 
 
 
 
 
 
 
1
+ {
2
+ "_from_model_config": true,
3
+ "bos_token_id": 1,
4
+ "do_sample": true,
5
+ "eos_token_id": 2,
6
+ "transformers_version": "4.40.0.dev0"
7
+ }
model-00001-of-00117.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:ed9ccee4b063e10c87cc2dfceff015e0a3e678e2fb90f3abbfb1fc8e7915aa9c
3
+ size 4762879840
model-00002-of-00117.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:f22d549ed12f075f97104a759327e1f500f3d22693c2d1fde680317b123223ac
3
+ size 4831839800
model-00003-of-00117.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:69c6b76c9cce92c00ba519b615a8f06fbe59dafc4e3de4b79a06408c2a9c1514
3
+ size 4781754592
model-00004-of-00117.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:243be14b5542d668490df8848ff2b1589c6486165bf69c38cf8b48e47cb5d8f3
3
+ size 4831839800
model-00005-of-00117.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:570b13e32ac883b39d079ac4c34c836f13d681e67963d21f5b3f9c6404a39b22
3
+ size 4781754592
model-00006-of-00117.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:3a8a5c88af32c56a7ca1dc2082402effcae75e4ea341241cbc56009820614845
3
+ size 4831839800
model-00007-of-00117.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:1ca84f92b0c4bb27248d8e51a02cc5664fe3dc3f4121eed732b25a9ec53e3220
3
+ size 4781754592
model-00008-of-00117.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:c5045952f7053f129a938a4c1f92f271d8662bf3302dc2544a309d015a244717
3
+ size 4831839800
model-00009-of-00117.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:3d49115ba7f3eb498d8e90f46ee39fca11b8e8cd46e3b3650f192291a49e1e79
3
+ size 4781754592
model-00010-of-00117.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:9c24577203ef245bf9c9c4733263ff803955aa45198ac695ef058e5223ec50c8
3
+ size 4831839800
model-00011-of-00117.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:61eb41a21b93c2762da0381320e8a2815ef22c72da0204217d05c064105b606b
3
+ size 4781754592
model-00012-of-00117.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:cb9dbb091b3bff6126fd5cb482672d64df73f549e2f35e0ee0a0b05b6cef6114
3
+ size 4831839800
model-00013-of-00117.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:50fadd1d36aeaf2c6c97157c3668695113e86bc823a9510d29eff3d26685920d
3
+ size 4781754592
model-00014-of-00117.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:e63553794ae97e292ccabcf8c279aa58e96fa7ffc366888f6fd28b61bf728094
3
+ size 4831839800
model-00015-of-00117.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:9836e419e59e2997b020c82fabf89431def9fbeb30454fe472f1b245f9c280cf
3
+ size 4781754592
model-00016-of-00117.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:33bf7f116407b576afcaf16a7aef0a7daa43694ff49a90e6444ffb78d65819be
3
+ size 4831839800
model-00017-of-00117.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:29a9d2cf8345f044fba69fa5e3db916b922f23de90b33f1ca20af7c9a5888ed3
3
+ size 4781754592
model-00018-of-00117.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:1aadacd71c89329cbb00956f5184a44e2b00ec70d990186fee188942faae39e3
3
+ size 4831839800
model-00019-of-00117.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:fa2aa34ae8d24e7b01224c11e4d5d8fc68d824d56e948ac9329d670394442107
3
+ size 4781754592
model-00020-of-00117.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:2e6e48d5f00e00a9722ade9d3aca8b11f984314c724fd1ed6f46009af47dd687
3
+ size 4831839800
model-00021-of-00117.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:044656ef715db71f006c4e9cdd6a790f0b90017d3378dc0ef44a75aa7dfa62f8
3
+ size 4982884240
model-00022-of-00117.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:dc0bfa03e064a941ad41dfff8359414bce4ba670c065b42bd52bdcec3d18dfcd
3
+ size 4630710168
model-00023-of-00117.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:b43e20babd6871217657425e7690a64407f62740150529ca6c7e27774a531c19
3
+ size 4831839808
model-00024-of-00117.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:fcfff7ffae614189f50adff5519690df74290d122b241afd6de26319e051f011
3
+ size 4781754608
model-00025-of-00117.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:86ebf3ad56678d6860cac36cfcf9c42061129e91ebe8af6ac34c68ebefdece0b
3
+ size 4831839808
model-00026-of-00117.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:3743756ea336fec881893a1236f70fc75d635783e42a574bfac763a7495f7d01
3
+ size 4781754608
model-00027-of-00117.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:41a7d915254a7191916791c7d26a700a3c9043f87e82b7244bdb88d554aff7f9
3
+ size 4831839808
model-00028-of-00117.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:804c38ba54981e0cc73ab557ed2e61676dc66802865b70d4363001ad53898ccf
3
+ size 4781754608
model-00029-of-00117.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:8f9a91e9b92885c8f316f8a5bb966fdc7293bc82ebeb0f3e867f9d2ea8870623
3
+ size 4831839808
model-00030-of-00117.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:fd9a6501726d7f0616eef570b17a26c9ff30a1295b7b2ffb7ffebb635f54347f
3
+ size 4781754608
model-00031-of-00117.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:8f3ce74a0727d90631ffe4aa90a5c3c43caa32d3e8c7f0400f0bf18ed4ad6caf
3
+ size 4831839808
model-00032-of-00117.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:acd28605ac4b054904ed535da531cbd965e812b2b2d0a6a8fa5825a8973d4d11
3
+ size 4781754608
model-00033-of-00117.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:c8ec7fee4a70220c3dae069c26858b6a8fa0ea8da756199ea005c77274800452
3
+ size 4831839808
model-00034-of-00117.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:4cc54c2e044db500106bfff731c7d1ae0aa3c5cfa38e366f7b999c8fbf8f4cf5
3
+ size 4781754608
model-00035-of-00117.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:725e9151be7acebc2d1f40e10d8e11688254f540cbfa2d06effb0d19c43410e4
3
+ size 4831839808
model-00036-of-00117.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:cdbc7a37fa36bea7902e2af4f7846f1be91b13900a5f6646589b2d16ac376058
3
+ size 4781754608