diff --git a/config.json b/config.json new file mode 100644 index 0000000000000000000000000000000000000000..baec789f6cdfa169ff32b825679566f15d605dac --- /dev/null +++ b/config.json @@ -0,0 +1,39 @@ +{ + "_name_or_path": "/opt/llama3.1/Meta-Llama-3.1-405B-Instruct/", + "architectures": [ + "LlamaForCausalLM" + ], + "attention_bias": false, + "attention_dropout": 0.0, + "bos_token_id": 128000, + "eos_token_id": [ + 128001, + 128008, + 128009 + ], + "hidden_act": "silu", + "hidden_size": 16384, + "initializer_range": 0.02, + "intermediate_size": 53248, + "max_position_embeddings": 131072, + "mlp_bias": false, + "model_type": "llama", + "num_attention_heads": 128, + "num_hidden_layers": 126, + "num_key_value_heads": 8, + "pretraining_tp": 1, + "rms_norm_eps": 1e-05, + "rope_scaling": { + "factor": 8.0, + "high_freq_factor": 4.0, + "low_freq_factor": 1.0, + "original_max_position_embeddings": 8192, + "rope_type": "llama3" + }, + "rope_theta": 500000.0, + "tie_word_embeddings": false, + "torch_dtype": "bfloat16", + "transformers_version": "4.43.4", + "use_cache": true, + "vocab_size": 128256 +} diff --git a/generation_config.json b/generation_config.json new file mode 100644 index 0000000000000000000000000000000000000000..57a143f2aec52fb1e288ab850ad0d13bb74becaf --- /dev/null +++ b/generation_config.json @@ -0,0 +1,12 @@ +{ + "bos_token_id": 128000, + "do_sample": true, + "eos_token_id": [ + 128001, + 128008, + 128009 + ], + "temperature": 0.6, + "top_p": 0.9, + "transformers_version": "4.43.4" +} diff --git a/hf_quant_config.json b/hf_quant_config.json new file mode 100644 index 0000000000000000000000000000000000000000..8157c8ddcaac48c2b7cf4717c636dec9dbd8373a --- /dev/null +++ b/hf_quant_config.json @@ -0,0 +1,10 @@ +{ + "producer": { + "name": "modelopt", + "version": "0.17.0" + }, + "quantization": { + "quant_algo": "FP8", + "kv_cache_quant_algo": null + } +} diff --git a/model-00001-of-00086.safetensors b/model-00001-of-00086.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..0deba5632a5b7425e23ee00a753b30894a2afb90 --- /dev/null +++ b/model-00001-of-00086.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:abc799746b86eb40c887f3b66c1f1df1b66c8925dcd6ce372cc21193c6a54fad +size 4773119376 diff --git a/model-00002-of-00086.safetensors b/model-00002-of-00086.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..5f4443feb2f223919f1064b551349dc6a89ee2be --- /dev/null +++ b/model-00002-of-00086.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:3917a62ef49efce3331d1621f6409c5466b17fbe757cc00bdefafd2180dc8236 +size 4932570128 diff --git a/model-00003-of-00086.safetensors b/model-00003-of-00086.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..69cb7849d3e78f67e1333f2fba00bb001bc89200 --- /dev/null +++ b/model-00003-of-00086.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:a46ee529d528b6dfec63878fb7e65398fc8887ce3249504031757193c5cde7d2 +size 4630646976 diff --git a/model-00004-of-00086.safetensors b/model-00004-of-00086.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..2c1ff5ec6b6bfafe3a67b3d674228caa85eca907 --- /dev/null +++ b/model-00004-of-00086.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:7db50ebf9f0f94dede49b1b4f100cf93766f512fb3d457d85eec3f629e50319b +size 4932570128 diff --git a/model-00005-of-00086.safetensors b/model-00005-of-00086.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..b166b2df3a9b309b2e32a201ff3c901ee0363f51 --- /dev/null +++ b/model-00005-of-00086.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:4a377187b3de8e1c296b21fefb4b0e88fa9e7d0eb4cb94f2862065384c3cccb5 +size 4630646976 diff --git a/model-00006-of-00086.safetensors b/model-00006-of-00086.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..cd8dd1bac41227dc7acf88bb5a959592f33a6d22 --- /dev/null +++ b/model-00006-of-00086.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:cab4217ce8fa627e56f1c60cf4e6184c495c0be8fa18eb960cedd7a9788daeed +size 4932570128 diff --git a/model-00007-of-00086.safetensors b/model-00007-of-00086.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..547eb5d4a0680930d1caa0df532261d735a1d20b --- /dev/null +++ b/model-00007-of-00086.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:bd2a4ad3ee10bede80e5e8f3340f4a063a0d45f02a5c0c2cd77ed3708ed89f30 +size 4630646976 diff --git a/model-00008-of-00086.safetensors b/model-00008-of-00086.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..e1f2dcbecb3e165c14fbd1b80824ce287d953540 --- /dev/null +++ b/model-00008-of-00086.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:341eeb2f36f63dcd86cd081bd6b1de8355a52eecd3facddb355b5b9d2576d3dc +size 4932570144 diff --git a/model-00009-of-00086.safetensors b/model-00009-of-00086.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..53f859fdea12e0d307df8d248d2a63af7596fb4c --- /dev/null +++ b/model-00009-of-00086.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:0b32ee8cba42ec8696b9de04a206aca20d5ef7f2c77b56e804162a5c0cb4e938 +size 4630647016 diff --git a/model-00010-of-00086.safetensors b/model-00010-of-00086.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..4cfa146a6a7bcb29401c2db216ae9a61aa0140a4 --- /dev/null +++ b/model-00010-of-00086.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:4413283a100aa3138b9fddd530c585fe3a881cce69df617ebfbc6808d71a1d62 +size 4932570160 diff --git a/model-00011-of-00086.safetensors b/model-00011-of-00086.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..50b44c0e9487006cb446a5fdac96c33d5daad326 --- /dev/null +++ b/model-00011-of-00086.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:0f3df58f7599098b9d444c9e0ed1119509adcb11fba4aae66a2bd5c4ef336c32 +size 4630647016 diff --git a/model-00012-of-00086.safetensors b/model-00012-of-00086.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..05f5c8d7612d0fdf01601056367375717747b454 --- /dev/null +++ b/model-00012-of-00086.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:d7e2cf786c5195d3a59b33faf8e73fb39bb691957fab0d9811e35517e998f086 +size 4932570160 diff --git a/model-00013-of-00086.safetensors b/model-00013-of-00086.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..b47dd78268d640d11a592f951b80f07323e65580 --- /dev/null +++ b/model-00013-of-00086.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:d729c1da0bc71319f56a0dbafc6d2d7f1a0bf25f3d47dc69298a32bdcfa55b4f +size 4630647016 diff --git a/model-00014-of-00086.safetensors b/model-00014-of-00086.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..6d66db7f0ffdfeee5907d5b26a52614b2b9ff8c0 --- /dev/null +++ b/model-00014-of-00086.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:89cefccb0925f1b897097ac951c3c5d546453413d1747e18fb2377753e6da8b9 +size 4932570160 diff --git a/model-00015-of-00086.safetensors b/model-00015-of-00086.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..4571131c46a1c1badc7f0c55f7ed9aa36456b5e1 --- /dev/null +++ b/model-00015-of-00086.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:ab1e9fabbe6a02391b5e614dcfbaf30cabf5c27b973cc3692fc73f17d378d214 +size 4630647016 diff --git a/model-00016-of-00086.safetensors b/model-00016-of-00086.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..98b552528384dbe6fc20794c6f63273a820c2767 --- /dev/null +++ b/model-00016-of-00086.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:99e71eacfd0a418738f1e2e12bbd533058f9c8ef9c7fe6d58315284254abe476 +size 4932570160 diff --git a/model-00017-of-00086.safetensors b/model-00017-of-00086.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..17a8cc8c6c00915ba54c15d77c4c8fcc7213b732 --- /dev/null +++ b/model-00017-of-00086.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:3bc77d66e6ede3e0e10b225b732d7a52f53f4124eb49201698ba5741404b164a +size 4630647016 diff --git a/model-00018-of-00086.safetensors b/model-00018-of-00086.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..7e7721b6e62f1db918350a58e44a8cc0ce414593 --- /dev/null +++ b/model-00018-of-00086.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:a63590a0766446e17835a4b3d3407688b0c2c795f9eabc09097bdb0274acd42d +size 4932570160 diff --git a/model-00019-of-00086.safetensors b/model-00019-of-00086.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..b1db7e90be3d3b18c3cdac9468bae32b4645d4b5 --- /dev/null +++ b/model-00019-of-00086.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:44b7f24cc86576cc9b2ba0b6e9d06f77bfd2d99f34b9a08e139b047f1c916a48 +size 4630647016 diff --git a/model-00020-of-00086.safetensors b/model-00020-of-00086.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..a602d7f98df06e26f7ed85b4ac0bc67a87da0306 --- /dev/null +++ b/model-00020-of-00086.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:cbf3b420b673a88769315d8ce192807bc38dbc849c3ae54567d620aaec93b7ed +size 4932570160 diff --git a/model-00021-of-00086.safetensors b/model-00021-of-00086.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..d6bafd7fdb4c2c9e0fc7a7b407ea032b1ededf3f --- /dev/null +++ b/model-00021-of-00086.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:fa037061ff66d64fa0a3152c32bf2452cac8a87848a99fa50b1b894d6806d9af +size 4630647016 diff --git a/model-00022-of-00086.safetensors b/model-00022-of-00086.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..693b6998932239e191c76f4942547ed4250463f4 --- /dev/null +++ b/model-00022-of-00086.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:48e9e21b78f808c4726a896d3791a26c787c053bc2182e109ce5b3b2c5e81ca1 +size 4932570160 diff --git a/model-00023-of-00086.safetensors b/model-00023-of-00086.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..e37c7ac478094bf04c7fbad00154961ef73882be --- /dev/null +++ b/model-00023-of-00086.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:325fc87cb641c0f8632170b4d2ecbb8a790fa42cf4d52c416d86c5389049b949 +size 4630647016 diff --git a/model-00024-of-00086.safetensors b/model-00024-of-00086.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..faeb1f429ff0fe3d99d4a10e1aec7e482924466a --- /dev/null +++ b/model-00024-of-00086.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:c8faaeba1c8166d9ab926d84a05672e6adeb687e03b20724e929945d51f01ca9 +size 4932570160 diff --git a/model-00025-of-00086.safetensors b/model-00025-of-00086.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..e36b672b65107acd3d5533aeaf08f0fe9ab62949 --- /dev/null +++ b/model-00025-of-00086.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:4709e6e48efaf8c73933f06ac7e90f689ef07b80811707291e241499139dca17 +size 4630647016 diff --git a/model-00026-of-00086.safetensors b/model-00026-of-00086.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..ed255871e7e4e98f2f71ebcfa313749ce9945918 --- /dev/null +++ b/model-00026-of-00086.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:7d29e50dd7fb1d407180bf86bc641a216a77b4be74af130bb930449e3c634347 +size 4932570160 diff --git a/model-00027-of-00086.safetensors b/model-00027-of-00086.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..7d748bd6e97bf529f7d1cf28a164480c7b395bd9 --- /dev/null +++ b/model-00027-of-00086.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:6175ba8c92031a46d81ab5b815724110f155fa6efdf00e162e58b5b92dc04a5f +size 4630647016 diff --git a/model-00028-of-00086.safetensors b/model-00028-of-00086.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..240228c57078b372e59b90940f87d007a124dfb9 --- /dev/null +++ b/model-00028-of-00086.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:0b08da7afe4edfc8863829ece066ad2306d31e9190cae1f7498e0ada93a0efdc +size 4932570160 diff --git a/model-00029-of-00086.safetensors b/model-00029-of-00086.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..d18099fa61632d3ff0ce0511e186dab3852013bb --- /dev/null +++ b/model-00029-of-00086.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:375bc82d48d042ab84f72e6c986b954b865286b15b5f364b0c23c1526d7105ab +size 4630647016 diff --git a/model-00030-of-00086.safetensors b/model-00030-of-00086.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..34790a6c30de58a3780313c206c89875a6c223ab --- /dev/null +++ b/model-00030-of-00086.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:abdea2093d079a3f01f2dc2815412fb8fb58b3932448c10a5dbef3325b71d4c4 +size 4932570160 diff --git a/model-00031-of-00086.safetensors b/model-00031-of-00086.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..6ed7731cba1df8037f2538a80fb0499f3b2c6fd9 --- /dev/null +++ b/model-00031-of-00086.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:7f83ffc139abca27d301931d476347e484375f4100e0d98ca68b8cb86e52a65d +size 4630647016 diff --git a/model-00032-of-00086.safetensors b/model-00032-of-00086.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..83cbebd0b9f8760c4a824700b7056960ec5b722e --- /dev/null +++ b/model-00032-of-00086.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:71e78f25c7b2d4db79f65373d597970173872b9a3808401b61d3b28701166244 +size 4932570160 diff --git a/model-00033-of-00086.safetensors b/model-00033-of-00086.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..5b1d04038c4da4f47391928c1c7da3f716d3e0e8 --- /dev/null +++ b/model-00033-of-00086.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:fbb68bc5ef9951cf2e0182a0b5cbc3e3773de794614cd869cda015fd4c386678 +size 4630647016 diff --git a/model-00034-of-00086.safetensors b/model-00034-of-00086.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..35172d3390bfed82a376a99b0e4a43af2e10cedd --- /dev/null +++ b/model-00034-of-00086.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:964c58979eb119eb30d052da0251c14138629d12522cb664475c354cff1852d9 +size 4932570160 diff --git a/model-00035-of-00086.safetensors b/model-00035-of-00086.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..52c7edd713324886fcfd2b57a3e2c732a19059ec --- /dev/null +++ b/model-00035-of-00086.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:e2b5aa300bab7e85a480d0a7f847816a458398da82a4f8a5e512b19564809e3a +size 4630647016 diff --git a/model-00036-of-00086.safetensors b/model-00036-of-00086.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..39cbb17ffc0e17fa89d989589765d87cb3c4c922 --- /dev/null +++ b/model-00036-of-00086.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:bd9195bea0a5992ddce286acfcc2ee9de6859ae8faaaf9176fcb854c89803248 +size 4932570160 diff --git a/model-00037-of-00086.safetensors b/model-00037-of-00086.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..808fb3be3afa86523e6ff0e7f2c6b619016b048f --- /dev/null +++ b/model-00037-of-00086.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:67915574bf7aad4ee6299f33e7706ab40e0437e3014ea6c7a42eadcf582099cc +size 4630647016 diff --git a/model-00038-of-00086.safetensors b/model-00038-of-00086.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..80d92985f9f2cc678c683e2a6f8ac7cf381a63b9 --- /dev/null +++ b/model-00038-of-00086.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:dd5cdcddffcd6b4dde66e5fe7070e8a8447981471baf345fa1db79b080d8e521 +size 4932570160 diff --git a/model-00039-of-00086.safetensors b/model-00039-of-00086.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..a54b3fbd6b7a2983f3339869820eb5eee1ff837c --- /dev/null +++ b/model-00039-of-00086.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:027128272cf78c68ce3261cd8f7834f3e52e1d12bd7793fe2f3f10e33f629d3a +size 4630647016 diff --git a/model-00040-of-00086.safetensors b/model-00040-of-00086.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..6f63c0881ada9de9c080303ea5475b8d4cccd535 --- /dev/null +++ b/model-00040-of-00086.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:33fd2044a3e4a6f5557953fe7683d8fdbd37a83d56568e5aacf2435d7e8d3a5b +size 4932570160 diff --git a/model-00041-of-00086.safetensors b/model-00041-of-00086.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..415562fab3247703fa6e8de4f0d464149b3c5b51 --- /dev/null +++ b/model-00041-of-00086.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:94fb40bbc0a21692d63a2b59a3648cf5d380232e53f38860a28f5d907fc4c6a9 +size 4630647016 diff --git a/model-00042-of-00086.safetensors b/model-00042-of-00086.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..8a74a7c4f0897bd93cbddefb477be2a1289feaa7 --- /dev/null +++ b/model-00042-of-00086.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:bc8e474d63ebeb8a0704470bd45a83c4d6d3214e189b250e0c010ebcc2363ec1 +size 4932570160 diff --git a/model-00043-of-00086.safetensors b/model-00043-of-00086.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..e752921a5f47840e368f22805d3e1577f57eb02b --- /dev/null +++ b/model-00043-of-00086.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:42ddf7231acf5bc45f939b6aeba554cb6fbe4753f3ac815df5c9b50f06484e11 +size 4630647016 diff --git a/model-00044-of-00086.safetensors b/model-00044-of-00086.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..42a4ac06f861a86d1b7d11aba6a5e7cca0f5e827 --- /dev/null +++ b/model-00044-of-00086.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:25b2f8a47a783918462e4ba4547bf98645dad8203fd22b5f78b773d82d4bba3d +size 4932570160 diff --git a/model-00045-of-00086.safetensors b/model-00045-of-00086.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..bbcffccafd49af56f15641f0377193ec6ca46ba5 --- /dev/null +++ b/model-00045-of-00086.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:fadbc818e2f55c0cc1750e64502436aa7965682fe80db297a83861eda1bcd2a4 +size 4630647016 diff --git a/model-00046-of-00086.safetensors b/model-00046-of-00086.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..d447c522a63b920c6f965787c7862ff7141dec9f --- /dev/null +++ b/model-00046-of-00086.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:0c02fe1c87881687b3b0927cfbc1ec41edf3c4f7cf87f4db7661a3f89fc7ebc0 +size 4932570160 diff --git a/model-00047-of-00086.safetensors b/model-00047-of-00086.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..2dde8d8c48a86f1db8f01123b8cda796f87bcf3e --- /dev/null +++ b/model-00047-of-00086.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:21f03e7a296a843b26247ec41217d6d5ce9e6abcc251fbeae5d66519eeb8cb4b +size 4630647016 diff --git a/model-00048-of-00086.safetensors b/model-00048-of-00086.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..f46fb2ef9945f30267f9b3b784f8c5d724bddf4e --- /dev/null +++ b/model-00048-of-00086.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:d59464b9c95486a4fb27685a877e7679c39693fb16b11a0fac114f5ce2a2d80c +size 4932570160 diff --git a/model-00049-of-00086.safetensors b/model-00049-of-00086.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..6785f26f230cb4e17b5dbe6d2ff03242ae946004 --- /dev/null +++ b/model-00049-of-00086.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:d8ab572e4d3bb25764fde3e732806e94e784c28fcaa9230ccd07a7b190969cfd +size 4630647016 diff --git a/model-00050-of-00086.safetensors b/model-00050-of-00086.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..3f2594a5c439b4b8c54fa3888b49ce5803709433 --- /dev/null +++ b/model-00050-of-00086.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:db5753b295e9381ad6f7903cf7adc1ec6e07f4f59bcd71ab074ecb5647b3ac6f +size 4932570160 diff --git a/model-00051-of-00086.safetensors b/model-00051-of-00086.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..9b4331511200daab0ae05e604a0d14ada3523288 --- /dev/null +++ b/model-00051-of-00086.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:2dd55f2cd5bf927133bc5d18c5ebda6defa4983448491ba8d907e4f557e9d2ab +size 4630647016 diff --git a/model-00052-of-00086.safetensors b/model-00052-of-00086.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..c72356e525e5b5501bb920f89041909d6139fce7 --- /dev/null +++ b/model-00052-of-00086.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:357596d67c2d5d3223d1a26277775b9ca718c5698ea0eafba85bf12e60175b7f +size 4932570160 diff --git a/model-00053-of-00086.safetensors b/model-00053-of-00086.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..af1b366c11c6f990305008009b39105391e0f765 --- /dev/null +++ b/model-00053-of-00086.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:8b7acce89e3911ce01b116cc0d0de4c35f6ee8ab0d79491f05b647f66719d881 +size 4630647016 diff --git a/model-00054-of-00086.safetensors b/model-00054-of-00086.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..a6aef2d1078d159d5aa356e19fc4fdaa74031211 --- /dev/null +++ b/model-00054-of-00086.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:a06caaea84733c021bd720ca791b117394df50b568d10fe8b81b510059bfe067 +size 4932570160 diff --git a/model-00055-of-00086.safetensors b/model-00055-of-00086.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..d35f99ad52341a0d6cb5ae369879ae41ea2dff6f --- /dev/null +++ b/model-00055-of-00086.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:521ae77b8ab363add741c7d5f985d709d08ca606d09ff7147babc1712f59722c +size 4630647016 diff --git a/model-00056-of-00086.safetensors b/model-00056-of-00086.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..eaffd1dba0e14450e93e07869d6829a45b37ad2f --- /dev/null +++ b/model-00056-of-00086.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:1260ee9956bfe9c67a3a7e5138122a7c47fcda1fad23629689137f3f4ac098d9 +size 4932570160 diff --git a/model-00057-of-00086.safetensors b/model-00057-of-00086.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..8b382147989a800116b4cbf799e3b9a4e8afe72e --- /dev/null +++ b/model-00057-of-00086.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:492b7aec58f7857a6880362bdec1ba2d07bfc8e5b4ee8959b0a7b21378b33bd3 +size 4630647016 diff --git a/model-00058-of-00086.safetensors b/model-00058-of-00086.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..2f3b804700b05ef6a1b810342d16d320c5066c8c --- /dev/null +++ b/model-00058-of-00086.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:73b9a221ad00c1207a3752c1e8d25eacc621762d74942ebdbb07fe43089b18d0 +size 4932570160 diff --git a/model-00059-of-00086.safetensors b/model-00059-of-00086.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..a895f89c4b8865bb8c591117abce37958d1f581c --- /dev/null +++ b/model-00059-of-00086.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:f8eaf3290ff0ecd46b546e4c50cef1b7dd71fe552d178f096459bf09c03fae2b +size 4630647016 diff --git a/model-00060-of-00086.safetensors b/model-00060-of-00086.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..ea8dd0a3cc2baa70e7f479458b665d75c205e40f --- /dev/null +++ b/model-00060-of-00086.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:96a738f61adb669bf922e8f3fc0f3ec8ad70463b3f3375e1df618f965b63655c +size 4932570160 diff --git a/model-00061-of-00086.safetensors b/model-00061-of-00086.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..27a66b894eae6fdef35ce2385eb4444e1cb7a476 --- /dev/null +++ b/model-00061-of-00086.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:653c76c5379bf742be8fd1858191aa3cefb7b7ea7a7e2de5290a7021b79def84 +size 4630647016 diff --git a/model-00062-of-00086.safetensors b/model-00062-of-00086.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..140eb44d5836f1adf850d8ed5124c9167b114e7f --- /dev/null +++ b/model-00062-of-00086.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:6d09e5ad0a811aaaceb3a33c8c7dab7cbb681a7452097735c51b4ca49a52dc4d +size 4932570160 diff --git a/model-00063-of-00086.safetensors b/model-00063-of-00086.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..89e48d33666d6143c11db5ced3f5d9fdf28f7138 --- /dev/null +++ b/model-00063-of-00086.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:8352d626a11e1e675be89f905f42cbb3793cedfe27e13108de85857a4c37bf8b +size 4630647016 diff --git a/model-00064-of-00086.safetensors b/model-00064-of-00086.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..23b79e9523f9fd934f02f2c2702d7f5f92024522 --- /dev/null +++ b/model-00064-of-00086.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:117e1b8d7ac6b7c8d609fddf5cb2d9e1d55e96e07e5327d28bc3ff44daafe23c +size 4932570160 diff --git a/model-00065-of-00086.safetensors b/model-00065-of-00086.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..d553834c901031536e2a8ea4cb0fa0dca95a580c --- /dev/null +++ b/model-00065-of-00086.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:ca48744910e107cb36e843d456a55632b9cc2d5bf81ba3eb4937591ebdac9741 +size 4630647016 diff --git a/model-00066-of-00086.safetensors b/model-00066-of-00086.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..24666cd1ce62f5e6ee9a1152ed9c1668b8e42dd2 --- /dev/null +++ b/model-00066-of-00086.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:1523b823e45513fef3f42b2ed85cad16fd50777fda884bd0cda333bffd5f70c0 +size 4932570160 diff --git a/model-00067-of-00086.safetensors b/model-00067-of-00086.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..6a88229aecaea0c07fb5a558889227d4ac5b58db --- /dev/null +++ b/model-00067-of-00086.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:6f3fdfbad44c667366ecf5d60a4db2a7f641c6c7ec91a52295dcaa3161a2881e +size 4630647016 diff --git a/model-00068-of-00086.safetensors b/model-00068-of-00086.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..d84762ad85dee9192c287c765b23d32822ba52dc --- /dev/null +++ b/model-00068-of-00086.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:39306dc1f7bca8d06c5069bcae704839d2edb044389a76faed95014ce272928c +size 4932570176 diff --git a/model-00069-of-00086.safetensors b/model-00069-of-00086.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..46027b901326fb2067bfd40ebd66f5f13b262be7 --- /dev/null +++ b/model-00069-of-00086.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:0465b8741cb26bce90e0d6942f02d9b7708b6aefbb2d315b642cf5e4e1aa1641 +size 4630647056 diff --git a/model-00070-of-00086.safetensors b/model-00070-of-00086.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..417bc08b56b79154a788ae44259729403feb5240 --- /dev/null +++ b/model-00070-of-00086.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:d9931b7cd25138d286663f34506abf3ae533e94ea59d2fd5f48ce461d94b4460 +size 4932570184 diff --git a/model-00071-of-00086.safetensors b/model-00071-of-00086.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..0583dfb4516d8b5a6c2f0e98a6c095c985e4e799 --- /dev/null +++ b/model-00071-of-00086.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:d7ded6919d648e84e502196ffdb05374665352c54cb86208b00ff07e62f60c27 +size 4630647056 diff --git a/model-00072-of-00086.safetensors b/model-00072-of-00086.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..de9700be1618e29402a942fed58f5afe4a6d16f0 --- /dev/null +++ b/model-00072-of-00086.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:5c2230487700317f73080ddaf018492c56be7f09c7d5661d6f7878f1e45df6c9 +size 4932570184 diff --git a/model-00073-of-00086.safetensors b/model-00073-of-00086.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..62fe6f3b8a4d458bd0df953379e8b35d00a019fe --- /dev/null +++ b/model-00073-of-00086.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:b807dbc63fd28822e2db38bb492b5fd98846f6196f2e45db768f967c85182b6d +size 4630647056 diff --git a/model-00074-of-00086.safetensors b/model-00074-of-00086.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..d987ab5e1f8ef900aac03ac2933b94e7d1145f5d --- /dev/null +++ b/model-00074-of-00086.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:2296637c5aa8461c3a8650e63f373af5575a840c25adaf88ca742a27896f7f4b +size 4932570184 diff --git a/model-00075-of-00086.safetensors b/model-00075-of-00086.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..11640a76451b9500bc491ca4198a0f83d51fc298 --- /dev/null +++ b/model-00075-of-00086.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:4c016e59b03d4c5163f0edc8a2f1336b06fad7b518d63d2fff4dc2b02b48b0f4 +size 4630647056 diff --git a/model-00076-of-00086.safetensors b/model-00076-of-00086.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..0b7ca59e33eaadb5c62b0651e189a57b71075db5 --- /dev/null +++ b/model-00076-of-00086.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:0a7c5d1cd6880cb126468126fb5134cba75095356409387eea81f0f90f93db6f +size 4932570184 diff --git a/model-00077-of-00086.safetensors b/model-00077-of-00086.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..b383ced88088a879959f9f4d12ec80e6295b2432 --- /dev/null +++ b/model-00077-of-00086.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:725f9a276496c8f7c6fc5b06eb42f73c36595509221641f2d4b848e62d279896 +size 4630647056 diff --git a/model-00078-of-00086.safetensors b/model-00078-of-00086.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..a351c4a34a509c861e7675522ba4c365adf3da87 --- /dev/null +++ b/model-00078-of-00086.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:9ef47e2b85f74fa467bc2107c43671d1e8ff987ab56d382165d373dc6f66705e +size 4932570184 diff --git a/model-00079-of-00086.safetensors b/model-00079-of-00086.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..9a6d5c088a0ca9c7fc5f7090821f0d3650333d27 --- /dev/null +++ b/model-00079-of-00086.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:9b81f0962f9a0f7733fd689ef7e01324075b2eb7239f87e0b573bccd23c2c696 +size 4630647056 diff --git a/model-00080-of-00086.safetensors b/model-00080-of-00086.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..8a2adb3f52c8c1d77b1cad2d0ef69adb50dd7d63 --- /dev/null +++ b/model-00080-of-00086.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:7335ec5fdbfe67d5af062e7bf1272e06cbef2a01e9de3704ee0d51f8eced72f3 +size 4932570184 diff --git a/model-00081-of-00086.safetensors b/model-00081-of-00086.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..dce185bda0314d29fc3b0aef52826fa1af117619 --- /dev/null +++ b/model-00081-of-00086.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:73598e3741c7ce7724656903194866eb23790ff6d3d3ed2766c405a67d8a8656 +size 4630647056 diff --git a/model-00082-of-00086.safetensors b/model-00082-of-00086.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..397d9c399e0a452c1b6a9795ac103b0c930e86aa --- /dev/null +++ b/model-00082-of-00086.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:faacf479050fa5c8482737c0d4ff27c3a3c1bb3e2095247fa58488d8e046efb9 +size 4932570184 diff --git a/model-00083-of-00086.safetensors b/model-00083-of-00086.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..1066aa573c21c2dc4af0fb34cffd95bc55184db3 --- /dev/null +++ b/model-00083-of-00086.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:d1a7e31b497c3cb5ed185c5b10f129c8c330b3b1c89e5598cdd5392d9c54bcbd +size 4630647056 diff --git a/model-00084-of-00086.safetensors b/model-00084-of-00086.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..3f9c5b13c1cf93964c14dff3a9e9d98caf7f9e24 --- /dev/null +++ b/model-00084-of-00086.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:e84e77c143f9e257d8d6429abb2dd6de4cfb37658ed75d2966ba06aaea47c4d7 +size 4932570184 diff --git a/model-00085-of-00086.safetensors b/model-00085-of-00086.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..5700a2d9b765d8c4318bdfd020acb65bffcc19d7 --- /dev/null +++ b/model-00085-of-00086.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:8fd4a8e97ef72ce896c3347619a7599f9cf4712d32ea9b466989bab6d91d8eff +size 4060253248 diff --git a/model-00086-of-00086.safetensors b/model-00086-of-00086.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..189b2ba970360412785980a181e8aa7ae7624a18 --- /dev/null +++ b/model-00086-of-00086.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:71b079a02069ebe5a5f4f8135815f136afb1d6fc4423620dcb2a14fa399cdf12 +size 4202692736 diff --git a/model.safetensors.index.json b/model.safetensors.index.json new file mode 100644 index 0000000000000000000000000000000000000000..a9f556dee30b21d05d572d14cfdb00ce41a1816a --- /dev/null +++ b/model.safetensors.index.json @@ -0,0 +1,2908 @@ +{ + "metadata": { + "total_size": 410060233616 + }, + "weight_map": { + "lm_head.weight": "model-00086-of-00086.safetensors", + "model.embed_tokens.weight": "model-00001-of-00086.safetensors", + "model.layers.0.input_layernorm.weight": "model-00002-of-00086.safetensors", + "model.layers.0.mlp.down_proj.input_scale": "model-00002-of-00086.safetensors", + "model.layers.0.mlp.down_proj.weight": "model-00002-of-00086.safetensors", + "model.layers.0.mlp.down_proj.weight_scale": "model-00002-of-00086.safetensors", + "model.layers.0.mlp.gate_proj.input_scale": "model-00002-of-00086.safetensors", + "model.layers.0.mlp.gate_proj.weight": "model-00002-of-00086.safetensors", + "model.layers.0.mlp.gate_proj.weight_scale": "model-00002-of-00086.safetensors", + "model.layers.0.mlp.up_proj.input_scale": "model-00002-of-00086.safetensors", + "model.layers.0.mlp.up_proj.weight": "model-00002-of-00086.safetensors", + "model.layers.0.mlp.up_proj.weight_scale": "model-00002-of-00086.safetensors", + "model.layers.0.post_attention_layernorm.weight": "model-00002-of-00086.safetensors", + "model.layers.0.self_attn.k_proj.input_scale": "model-00001-of-00086.safetensors", + "model.layers.0.self_attn.k_proj.weight": "model-00001-of-00086.safetensors", + "model.layers.0.self_attn.k_proj.weight_scale": "model-00001-of-00086.safetensors", + "model.layers.0.self_attn.o_proj.input_scale": "model-00001-of-00086.safetensors", + "model.layers.0.self_attn.o_proj.weight": "model-00001-of-00086.safetensors", + "model.layers.0.self_attn.o_proj.weight_scale": "model-00001-of-00086.safetensors", + "model.layers.0.self_attn.q_proj.input_scale": "model-00001-of-00086.safetensors", + "model.layers.0.self_attn.q_proj.weight": "model-00001-of-00086.safetensors", + "model.layers.0.self_attn.q_proj.weight_scale": "model-00001-of-00086.safetensors", + "model.layers.0.self_attn.v_proj.input_scale": "model-00001-of-00086.safetensors", + "model.layers.0.self_attn.v_proj.weight": "model-00001-of-00086.safetensors", + "model.layers.0.self_attn.v_proj.weight_scale": "model-00001-of-00086.safetensors", + "model.layers.1.input_layernorm.weight": "model-00003-of-00086.safetensors", + "model.layers.1.mlp.down_proj.input_scale": "model-00003-of-00086.safetensors", + "model.layers.1.mlp.down_proj.weight": "model-00003-of-00086.safetensors", + "model.layers.1.mlp.down_proj.weight_scale": "model-00003-of-00086.safetensors", + "model.layers.1.mlp.gate_proj.input_scale": "model-00002-of-00086.safetensors", + "model.layers.1.mlp.gate_proj.weight": "model-00002-of-00086.safetensors", + "model.layers.1.mlp.gate_proj.weight_scale": "model-00002-of-00086.safetensors", + "model.layers.1.mlp.up_proj.input_scale": "model-00002-of-00086.safetensors", + "model.layers.1.mlp.up_proj.weight": "model-00002-of-00086.safetensors", + "model.layers.1.mlp.up_proj.weight_scale": "model-00002-of-00086.safetensors", + "model.layers.1.post_attention_layernorm.weight": "model-00003-of-00086.safetensors", + "model.layers.1.self_attn.k_proj.input_scale": "model-00002-of-00086.safetensors", + "model.layers.1.self_attn.k_proj.weight": "model-00002-of-00086.safetensors", + "model.layers.1.self_attn.k_proj.weight_scale": "model-00002-of-00086.safetensors", + "model.layers.1.self_attn.o_proj.input_scale": "model-00002-of-00086.safetensors", + "model.layers.1.self_attn.o_proj.weight": "model-00002-of-00086.safetensors", + "model.layers.1.self_attn.o_proj.weight_scale": "model-00002-of-00086.safetensors", + "model.layers.1.self_attn.q_proj.input_scale": "model-00002-of-00086.safetensors", + "model.layers.1.self_attn.q_proj.weight": "model-00002-of-00086.safetensors", + "model.layers.1.self_attn.q_proj.weight_scale": "model-00002-of-00086.safetensors", + "model.layers.1.self_attn.v_proj.input_scale": "model-00002-of-00086.safetensors", + "model.layers.1.self_attn.v_proj.weight": "model-00002-of-00086.safetensors", + "model.layers.1.self_attn.v_proj.weight_scale": "model-00002-of-00086.safetensors", + "model.layers.10.input_layernorm.weight": "model-00009-of-00086.safetensors", + "model.layers.10.mlp.down_proj.input_scale": "model-00009-of-00086.safetensors", + "model.layers.10.mlp.down_proj.weight": "model-00009-of-00086.safetensors", + "model.layers.10.mlp.down_proj.weight_scale": "model-00009-of-00086.safetensors", + "model.layers.10.mlp.gate_proj.input_scale": "model-00008-of-00086.safetensors", + "model.layers.10.mlp.gate_proj.weight": "model-00008-of-00086.safetensors", + "model.layers.10.mlp.gate_proj.weight_scale": "model-00008-of-00086.safetensors", + "model.layers.10.mlp.up_proj.input_scale": "model-00008-of-00086.safetensors", + "model.layers.10.mlp.up_proj.weight": "model-00008-of-00086.safetensors", + "model.layers.10.mlp.up_proj.weight_scale": "model-00008-of-00086.safetensors", + "model.layers.10.post_attention_layernorm.weight": "model-00009-of-00086.safetensors", + "model.layers.10.self_attn.k_proj.input_scale": "model-00008-of-00086.safetensors", + "model.layers.10.self_attn.k_proj.weight": "model-00008-of-00086.safetensors", + "model.layers.10.self_attn.k_proj.weight_scale": "model-00008-of-00086.safetensors", + "model.layers.10.self_attn.o_proj.input_scale": "model-00008-of-00086.safetensors", + "model.layers.10.self_attn.o_proj.weight": "model-00008-of-00086.safetensors", + "model.layers.10.self_attn.o_proj.weight_scale": "model-00008-of-00086.safetensors", + "model.layers.10.self_attn.q_proj.input_scale": "model-00008-of-00086.safetensors", + "model.layers.10.self_attn.q_proj.weight": "model-00008-of-00086.safetensors", + "model.layers.10.self_attn.q_proj.weight_scale": "model-00008-of-00086.safetensors", + "model.layers.10.self_attn.v_proj.input_scale": "model-00008-of-00086.safetensors", + "model.layers.10.self_attn.v_proj.weight": "model-00008-of-00086.safetensors", + "model.layers.10.self_attn.v_proj.weight_scale": "model-00008-of-00086.safetensors", + "model.layers.100.input_layernorm.weight": "model-00069-of-00086.safetensors", + "model.layers.100.mlp.down_proj.input_scale": "model-00069-of-00086.safetensors", + "model.layers.100.mlp.down_proj.weight": "model-00069-of-00086.safetensors", + "model.layers.100.mlp.down_proj.weight_scale": "model-00069-of-00086.safetensors", + "model.layers.100.mlp.gate_proj.input_scale": "model-00068-of-00086.safetensors", + "model.layers.100.mlp.gate_proj.weight": "model-00068-of-00086.safetensors", + "model.layers.100.mlp.gate_proj.weight_scale": "model-00068-of-00086.safetensors", + "model.layers.100.mlp.up_proj.input_scale": "model-00068-of-00086.safetensors", + "model.layers.100.mlp.up_proj.weight": "model-00068-of-00086.safetensors", + "model.layers.100.mlp.up_proj.weight_scale": "model-00068-of-00086.safetensors", + "model.layers.100.post_attention_layernorm.weight": "model-00069-of-00086.safetensors", + "model.layers.100.self_attn.k_proj.input_scale": "model-00068-of-00086.safetensors", + "model.layers.100.self_attn.k_proj.weight": "model-00068-of-00086.safetensors", + "model.layers.100.self_attn.k_proj.weight_scale": "model-00068-of-00086.safetensors", + "model.layers.100.self_attn.o_proj.input_scale": "model-00068-of-00086.safetensors", + "model.layers.100.self_attn.o_proj.weight": "model-00068-of-00086.safetensors", + "model.layers.100.self_attn.o_proj.weight_scale": "model-00068-of-00086.safetensors", + "model.layers.100.self_attn.q_proj.input_scale": "model-00068-of-00086.safetensors", + "model.layers.100.self_attn.q_proj.weight": "model-00068-of-00086.safetensors", + "model.layers.100.self_attn.q_proj.weight_scale": "model-00068-of-00086.safetensors", + "model.layers.100.self_attn.v_proj.input_scale": "model-00068-of-00086.safetensors", + "model.layers.100.self_attn.v_proj.weight": "model-00068-of-00086.safetensors", + "model.layers.100.self_attn.v_proj.weight_scale": "model-00068-of-00086.safetensors", + "model.layers.101.input_layernorm.weight": "model-00069-of-00086.safetensors", + "model.layers.101.mlp.down_proj.input_scale": "model-00069-of-00086.safetensors", + "model.layers.101.mlp.down_proj.weight": "model-00069-of-00086.safetensors", + "model.layers.101.mlp.down_proj.weight_scale": "model-00069-of-00086.safetensors", + "model.layers.101.mlp.gate_proj.input_scale": "model-00069-of-00086.safetensors", + "model.layers.101.mlp.gate_proj.weight": "model-00069-of-00086.safetensors", + "model.layers.101.mlp.gate_proj.weight_scale": "model-00069-of-00086.safetensors", + "model.layers.101.mlp.up_proj.input_scale": "model-00069-of-00086.safetensors", + "model.layers.101.mlp.up_proj.weight": "model-00069-of-00086.safetensors", + "model.layers.101.mlp.up_proj.weight_scale": "model-00069-of-00086.safetensors", + "model.layers.101.post_attention_layernorm.weight": "model-00069-of-00086.safetensors", + "model.layers.101.self_attn.k_proj.input_scale": "model-00069-of-00086.safetensors", + "model.layers.101.self_attn.k_proj.weight": "model-00069-of-00086.safetensors", + "model.layers.101.self_attn.k_proj.weight_scale": "model-00069-of-00086.safetensors", + "model.layers.101.self_attn.o_proj.input_scale": "model-00069-of-00086.safetensors", + "model.layers.101.self_attn.o_proj.weight": "model-00069-of-00086.safetensors", + "model.layers.101.self_attn.o_proj.weight_scale": "model-00069-of-00086.safetensors", + "model.layers.101.self_attn.q_proj.input_scale": "model-00069-of-00086.safetensors", + "model.layers.101.self_attn.q_proj.weight": "model-00069-of-00086.safetensors", + "model.layers.101.self_attn.q_proj.weight_scale": "model-00069-of-00086.safetensors", + "model.layers.101.self_attn.v_proj.input_scale": "model-00069-of-00086.safetensors", + "model.layers.101.self_attn.v_proj.weight": "model-00069-of-00086.safetensors", + "model.layers.101.self_attn.v_proj.weight_scale": "model-00069-of-00086.safetensors", + "model.layers.102.input_layernorm.weight": "model-00070-of-00086.safetensors", + "model.layers.102.mlp.down_proj.input_scale": "model-00070-of-00086.safetensors", + "model.layers.102.mlp.down_proj.weight": "model-00070-of-00086.safetensors", + "model.layers.102.mlp.down_proj.weight_scale": "model-00070-of-00086.safetensors", + "model.layers.102.mlp.gate_proj.input_scale": "model-00070-of-00086.safetensors", + "model.layers.102.mlp.gate_proj.weight": "model-00070-of-00086.safetensors", + "model.layers.102.mlp.gate_proj.weight_scale": "model-00070-of-00086.safetensors", + "model.layers.102.mlp.up_proj.input_scale": "model-00070-of-00086.safetensors", + "model.layers.102.mlp.up_proj.weight": "model-00070-of-00086.safetensors", + "model.layers.102.mlp.up_proj.weight_scale": "model-00070-of-00086.safetensors", + "model.layers.102.post_attention_layernorm.weight": "model-00070-of-00086.safetensors", + "model.layers.102.self_attn.k_proj.input_scale": "model-00069-of-00086.safetensors", + "model.layers.102.self_attn.k_proj.weight": "model-00069-of-00086.safetensors", + "model.layers.102.self_attn.k_proj.weight_scale": "model-00069-of-00086.safetensors", + "model.layers.102.self_attn.o_proj.input_scale": "model-00069-of-00086.safetensors", + "model.layers.102.self_attn.o_proj.weight": "model-00069-of-00086.safetensors", + "model.layers.102.self_attn.o_proj.weight_scale": "model-00069-of-00086.safetensors", + "model.layers.102.self_attn.q_proj.input_scale": "model-00069-of-00086.safetensors", + "model.layers.102.self_attn.q_proj.weight": "model-00069-of-00086.safetensors", + "model.layers.102.self_attn.q_proj.weight_scale": "model-00069-of-00086.safetensors", + "model.layers.102.self_attn.v_proj.input_scale": "model-00069-of-00086.safetensors", + "model.layers.102.self_attn.v_proj.weight": "model-00069-of-00086.safetensors", + "model.layers.102.self_attn.v_proj.weight_scale": "model-00069-of-00086.safetensors", + "model.layers.103.input_layernorm.weight": "model-00071-of-00086.safetensors", + "model.layers.103.mlp.down_proj.input_scale": "model-00071-of-00086.safetensors", + "model.layers.103.mlp.down_proj.weight": "model-00071-of-00086.safetensors", + "model.layers.103.mlp.down_proj.weight_scale": "model-00071-of-00086.safetensors", + "model.layers.103.mlp.gate_proj.input_scale": "model-00070-of-00086.safetensors", + "model.layers.103.mlp.gate_proj.weight": "model-00070-of-00086.safetensors", + "model.layers.103.mlp.gate_proj.weight_scale": "model-00070-of-00086.safetensors", + "model.layers.103.mlp.up_proj.input_scale": "model-00070-of-00086.safetensors", + "model.layers.103.mlp.up_proj.weight": "model-00070-of-00086.safetensors", + "model.layers.103.mlp.up_proj.weight_scale": "model-00070-of-00086.safetensors", + "model.layers.103.post_attention_layernorm.weight": "model-00071-of-00086.safetensors", + "model.layers.103.self_attn.k_proj.input_scale": "model-00070-of-00086.safetensors", + "model.layers.103.self_attn.k_proj.weight": "model-00070-of-00086.safetensors", + "model.layers.103.self_attn.k_proj.weight_scale": "model-00070-of-00086.safetensors", + "model.layers.103.self_attn.o_proj.input_scale": "model-00070-of-00086.safetensors", + "model.layers.103.self_attn.o_proj.weight": "model-00070-of-00086.safetensors", + "model.layers.103.self_attn.o_proj.weight_scale": "model-00070-of-00086.safetensors", + "model.layers.103.self_attn.q_proj.input_scale": "model-00070-of-00086.safetensors", + "model.layers.103.self_attn.q_proj.weight": "model-00070-of-00086.safetensors", + "model.layers.103.self_attn.q_proj.weight_scale": "model-00070-of-00086.safetensors", + "model.layers.103.self_attn.v_proj.input_scale": "model-00070-of-00086.safetensors", + "model.layers.103.self_attn.v_proj.weight": "model-00070-of-00086.safetensors", + "model.layers.103.self_attn.v_proj.weight_scale": "model-00070-of-00086.safetensors", + "model.layers.104.input_layernorm.weight": "model-00071-of-00086.safetensors", + "model.layers.104.mlp.down_proj.input_scale": "model-00071-of-00086.safetensors", + "model.layers.104.mlp.down_proj.weight": "model-00071-of-00086.safetensors", + "model.layers.104.mlp.down_proj.weight_scale": "model-00071-of-00086.safetensors", + "model.layers.104.mlp.gate_proj.input_scale": "model-00071-of-00086.safetensors", + "model.layers.104.mlp.gate_proj.weight": "model-00071-of-00086.safetensors", + "model.layers.104.mlp.gate_proj.weight_scale": "model-00071-of-00086.safetensors", + "model.layers.104.mlp.up_proj.input_scale": "model-00071-of-00086.safetensors", + "model.layers.104.mlp.up_proj.weight": "model-00071-of-00086.safetensors", + "model.layers.104.mlp.up_proj.weight_scale": "model-00071-of-00086.safetensors", + "model.layers.104.post_attention_layernorm.weight": "model-00071-of-00086.safetensors", + "model.layers.104.self_attn.k_proj.input_scale": "model-00071-of-00086.safetensors", + "model.layers.104.self_attn.k_proj.weight": "model-00071-of-00086.safetensors", + "model.layers.104.self_attn.k_proj.weight_scale": "model-00071-of-00086.safetensors", + "model.layers.104.self_attn.o_proj.input_scale": "model-00071-of-00086.safetensors", + "model.layers.104.self_attn.o_proj.weight": "model-00071-of-00086.safetensors", + "model.layers.104.self_attn.o_proj.weight_scale": "model-00071-of-00086.safetensors", + "model.layers.104.self_attn.q_proj.input_scale": "model-00071-of-00086.safetensors", + "model.layers.104.self_attn.q_proj.weight": "model-00071-of-00086.safetensors", + "model.layers.104.self_attn.q_proj.weight_scale": "model-00071-of-00086.safetensors", + "model.layers.104.self_attn.v_proj.input_scale": "model-00071-of-00086.safetensors", + "model.layers.104.self_attn.v_proj.weight": "model-00071-of-00086.safetensors", + "model.layers.104.self_attn.v_proj.weight_scale": "model-00071-of-00086.safetensors", + "model.layers.105.input_layernorm.weight": "model-00072-of-00086.safetensors", + "model.layers.105.mlp.down_proj.input_scale": "model-00072-of-00086.safetensors", + "model.layers.105.mlp.down_proj.weight": "model-00072-of-00086.safetensors", + "model.layers.105.mlp.down_proj.weight_scale": "model-00072-of-00086.safetensors", + "model.layers.105.mlp.gate_proj.input_scale": "model-00072-of-00086.safetensors", + "model.layers.105.mlp.gate_proj.weight": "model-00072-of-00086.safetensors", + "model.layers.105.mlp.gate_proj.weight_scale": "model-00072-of-00086.safetensors", + "model.layers.105.mlp.up_proj.input_scale": "model-00072-of-00086.safetensors", + "model.layers.105.mlp.up_proj.weight": "model-00072-of-00086.safetensors", + "model.layers.105.mlp.up_proj.weight_scale": "model-00072-of-00086.safetensors", + "model.layers.105.post_attention_layernorm.weight": "model-00072-of-00086.safetensors", + "model.layers.105.self_attn.k_proj.input_scale": "model-00071-of-00086.safetensors", + "model.layers.105.self_attn.k_proj.weight": "model-00071-of-00086.safetensors", + "model.layers.105.self_attn.k_proj.weight_scale": "model-00071-of-00086.safetensors", + "model.layers.105.self_attn.o_proj.input_scale": "model-00071-of-00086.safetensors", + "model.layers.105.self_attn.o_proj.weight": "model-00071-of-00086.safetensors", + "model.layers.105.self_attn.o_proj.weight_scale": "model-00071-of-00086.safetensors", + "model.layers.105.self_attn.q_proj.input_scale": "model-00071-of-00086.safetensors", + "model.layers.105.self_attn.q_proj.weight": "model-00071-of-00086.safetensors", + "model.layers.105.self_attn.q_proj.weight_scale": "model-00071-of-00086.safetensors", + "model.layers.105.self_attn.v_proj.input_scale": "model-00071-of-00086.safetensors", + "model.layers.105.self_attn.v_proj.weight": "model-00071-of-00086.safetensors", + "model.layers.105.self_attn.v_proj.weight_scale": "model-00071-of-00086.safetensors", + "model.layers.106.input_layernorm.weight": "model-00073-of-00086.safetensors", + "model.layers.106.mlp.down_proj.input_scale": "model-00073-of-00086.safetensors", + "model.layers.106.mlp.down_proj.weight": "model-00073-of-00086.safetensors", + "model.layers.106.mlp.down_proj.weight_scale": "model-00073-of-00086.safetensors", + "model.layers.106.mlp.gate_proj.input_scale": "model-00072-of-00086.safetensors", + "model.layers.106.mlp.gate_proj.weight": "model-00072-of-00086.safetensors", + "model.layers.106.mlp.gate_proj.weight_scale": "model-00072-of-00086.safetensors", + "model.layers.106.mlp.up_proj.input_scale": "model-00072-of-00086.safetensors", + "model.layers.106.mlp.up_proj.weight": "model-00072-of-00086.safetensors", + "model.layers.106.mlp.up_proj.weight_scale": "model-00072-of-00086.safetensors", + "model.layers.106.post_attention_layernorm.weight": "model-00073-of-00086.safetensors", + "model.layers.106.self_attn.k_proj.input_scale": "model-00072-of-00086.safetensors", + "model.layers.106.self_attn.k_proj.weight": "model-00072-of-00086.safetensors", + "model.layers.106.self_attn.k_proj.weight_scale": "model-00072-of-00086.safetensors", + "model.layers.106.self_attn.o_proj.input_scale": "model-00072-of-00086.safetensors", + "model.layers.106.self_attn.o_proj.weight": "model-00072-of-00086.safetensors", + "model.layers.106.self_attn.o_proj.weight_scale": "model-00072-of-00086.safetensors", + "model.layers.106.self_attn.q_proj.input_scale": "model-00072-of-00086.safetensors", + "model.layers.106.self_attn.q_proj.weight": "model-00072-of-00086.safetensors", + "model.layers.106.self_attn.q_proj.weight_scale": "model-00072-of-00086.safetensors", + "model.layers.106.self_attn.v_proj.input_scale": "model-00072-of-00086.safetensors", + "model.layers.106.self_attn.v_proj.weight": "model-00072-of-00086.safetensors", + "model.layers.106.self_attn.v_proj.weight_scale": "model-00072-of-00086.safetensors", + "model.layers.107.input_layernorm.weight": "model-00073-of-00086.safetensors", + "model.layers.107.mlp.down_proj.input_scale": "model-00073-of-00086.safetensors", + "model.layers.107.mlp.down_proj.weight": "model-00073-of-00086.safetensors", + "model.layers.107.mlp.down_proj.weight_scale": "model-00073-of-00086.safetensors", + "model.layers.107.mlp.gate_proj.input_scale": "model-00073-of-00086.safetensors", + "model.layers.107.mlp.gate_proj.weight": "model-00073-of-00086.safetensors", + "model.layers.107.mlp.gate_proj.weight_scale": "model-00073-of-00086.safetensors", + "model.layers.107.mlp.up_proj.input_scale": "model-00073-of-00086.safetensors", + "model.layers.107.mlp.up_proj.weight": "model-00073-of-00086.safetensors", + "model.layers.107.mlp.up_proj.weight_scale": "model-00073-of-00086.safetensors", + "model.layers.107.post_attention_layernorm.weight": "model-00073-of-00086.safetensors", + "model.layers.107.self_attn.k_proj.input_scale": "model-00073-of-00086.safetensors", + "model.layers.107.self_attn.k_proj.weight": "model-00073-of-00086.safetensors", + "model.layers.107.self_attn.k_proj.weight_scale": "model-00073-of-00086.safetensors", + "model.layers.107.self_attn.o_proj.input_scale": "model-00073-of-00086.safetensors", + "model.layers.107.self_attn.o_proj.weight": "model-00073-of-00086.safetensors", + "model.layers.107.self_attn.o_proj.weight_scale": "model-00073-of-00086.safetensors", + "model.layers.107.self_attn.q_proj.input_scale": "model-00073-of-00086.safetensors", + "model.layers.107.self_attn.q_proj.weight": "model-00073-of-00086.safetensors", + "model.layers.107.self_attn.q_proj.weight_scale": "model-00073-of-00086.safetensors", + "model.layers.107.self_attn.v_proj.input_scale": "model-00073-of-00086.safetensors", + "model.layers.107.self_attn.v_proj.weight": "model-00073-of-00086.safetensors", + "model.layers.107.self_attn.v_proj.weight_scale": "model-00073-of-00086.safetensors", + "model.layers.108.input_layernorm.weight": "model-00074-of-00086.safetensors", + "model.layers.108.mlp.down_proj.input_scale": "model-00074-of-00086.safetensors", + "model.layers.108.mlp.down_proj.weight": "model-00074-of-00086.safetensors", + "model.layers.108.mlp.down_proj.weight_scale": "model-00074-of-00086.safetensors", + "model.layers.108.mlp.gate_proj.input_scale": "model-00074-of-00086.safetensors", + "model.layers.108.mlp.gate_proj.weight": "model-00074-of-00086.safetensors", + "model.layers.108.mlp.gate_proj.weight_scale": "model-00074-of-00086.safetensors", + "model.layers.108.mlp.up_proj.input_scale": "model-00074-of-00086.safetensors", + "model.layers.108.mlp.up_proj.weight": "model-00074-of-00086.safetensors", + "model.layers.108.mlp.up_proj.weight_scale": "model-00074-of-00086.safetensors", + "model.layers.108.post_attention_layernorm.weight": "model-00074-of-00086.safetensors", + "model.layers.108.self_attn.k_proj.input_scale": "model-00073-of-00086.safetensors", + "model.layers.108.self_attn.k_proj.weight": "model-00073-of-00086.safetensors", + "model.layers.108.self_attn.k_proj.weight_scale": "model-00073-of-00086.safetensors", + "model.layers.108.self_attn.o_proj.input_scale": "model-00073-of-00086.safetensors", + "model.layers.108.self_attn.o_proj.weight": "model-00073-of-00086.safetensors", + "model.layers.108.self_attn.o_proj.weight_scale": "model-00073-of-00086.safetensors", + "model.layers.108.self_attn.q_proj.input_scale": "model-00073-of-00086.safetensors", + "model.layers.108.self_attn.q_proj.weight": "model-00073-of-00086.safetensors", + "model.layers.108.self_attn.q_proj.weight_scale": "model-00073-of-00086.safetensors", + "model.layers.108.self_attn.v_proj.input_scale": "model-00073-of-00086.safetensors", + "model.layers.108.self_attn.v_proj.weight": "model-00073-of-00086.safetensors", + "model.layers.108.self_attn.v_proj.weight_scale": "model-00073-of-00086.safetensors", + "model.layers.109.input_layernorm.weight": "model-00075-of-00086.safetensors", + "model.layers.109.mlp.down_proj.input_scale": "model-00075-of-00086.safetensors", + "model.layers.109.mlp.down_proj.weight": "model-00075-of-00086.safetensors", + "model.layers.109.mlp.down_proj.weight_scale": "model-00075-of-00086.safetensors", + "model.layers.109.mlp.gate_proj.input_scale": "model-00074-of-00086.safetensors", + "model.layers.109.mlp.gate_proj.weight": "model-00074-of-00086.safetensors", + "model.layers.109.mlp.gate_proj.weight_scale": "model-00074-of-00086.safetensors", + "model.layers.109.mlp.up_proj.input_scale": "model-00074-of-00086.safetensors", + "model.layers.109.mlp.up_proj.weight": "model-00074-of-00086.safetensors", + "model.layers.109.mlp.up_proj.weight_scale": "model-00074-of-00086.safetensors", + "model.layers.109.post_attention_layernorm.weight": "model-00075-of-00086.safetensors", + "model.layers.109.self_attn.k_proj.input_scale": "model-00074-of-00086.safetensors", + "model.layers.109.self_attn.k_proj.weight": "model-00074-of-00086.safetensors", + "model.layers.109.self_attn.k_proj.weight_scale": "model-00074-of-00086.safetensors", + "model.layers.109.self_attn.o_proj.input_scale": "model-00074-of-00086.safetensors", + "model.layers.109.self_attn.o_proj.weight": "model-00074-of-00086.safetensors", + "model.layers.109.self_attn.o_proj.weight_scale": "model-00074-of-00086.safetensors", + "model.layers.109.self_attn.q_proj.input_scale": "model-00074-of-00086.safetensors", + "model.layers.109.self_attn.q_proj.weight": "model-00074-of-00086.safetensors", + "model.layers.109.self_attn.q_proj.weight_scale": "model-00074-of-00086.safetensors", + "model.layers.109.self_attn.v_proj.input_scale": "model-00074-of-00086.safetensors", + "model.layers.109.self_attn.v_proj.weight": "model-00074-of-00086.safetensors", + "model.layers.109.self_attn.v_proj.weight_scale": "model-00074-of-00086.safetensors", + "model.layers.11.input_layernorm.weight": "model-00009-of-00086.safetensors", + "model.layers.11.mlp.down_proj.input_scale": "model-00009-of-00086.safetensors", + "model.layers.11.mlp.down_proj.weight": "model-00009-of-00086.safetensors", + "model.layers.11.mlp.down_proj.weight_scale": "model-00009-of-00086.safetensors", + "model.layers.11.mlp.gate_proj.input_scale": "model-00009-of-00086.safetensors", + "model.layers.11.mlp.gate_proj.weight": "model-00009-of-00086.safetensors", + "model.layers.11.mlp.gate_proj.weight_scale": "model-00009-of-00086.safetensors", + "model.layers.11.mlp.up_proj.input_scale": "model-00009-of-00086.safetensors", + "model.layers.11.mlp.up_proj.weight": "model-00009-of-00086.safetensors", + "model.layers.11.mlp.up_proj.weight_scale": "model-00009-of-00086.safetensors", + "model.layers.11.post_attention_layernorm.weight": "model-00009-of-00086.safetensors", + "model.layers.11.self_attn.k_proj.input_scale": "model-00009-of-00086.safetensors", + "model.layers.11.self_attn.k_proj.weight": "model-00009-of-00086.safetensors", + "model.layers.11.self_attn.k_proj.weight_scale": "model-00009-of-00086.safetensors", + "model.layers.11.self_attn.o_proj.input_scale": "model-00009-of-00086.safetensors", + "model.layers.11.self_attn.o_proj.weight": "model-00009-of-00086.safetensors", + "model.layers.11.self_attn.o_proj.weight_scale": "model-00009-of-00086.safetensors", + "model.layers.11.self_attn.q_proj.input_scale": "model-00009-of-00086.safetensors", + "model.layers.11.self_attn.q_proj.weight": "model-00009-of-00086.safetensors", + "model.layers.11.self_attn.q_proj.weight_scale": "model-00009-of-00086.safetensors", + "model.layers.11.self_attn.v_proj.input_scale": "model-00009-of-00086.safetensors", + "model.layers.11.self_attn.v_proj.weight": "model-00009-of-00086.safetensors", + "model.layers.11.self_attn.v_proj.weight_scale": "model-00009-of-00086.safetensors", + "model.layers.110.input_layernorm.weight": "model-00075-of-00086.safetensors", + "model.layers.110.mlp.down_proj.input_scale": "model-00075-of-00086.safetensors", + "model.layers.110.mlp.down_proj.weight": "model-00075-of-00086.safetensors", + "model.layers.110.mlp.down_proj.weight_scale": "model-00075-of-00086.safetensors", + "model.layers.110.mlp.gate_proj.input_scale": "model-00075-of-00086.safetensors", + "model.layers.110.mlp.gate_proj.weight": "model-00075-of-00086.safetensors", + "model.layers.110.mlp.gate_proj.weight_scale": "model-00075-of-00086.safetensors", + "model.layers.110.mlp.up_proj.input_scale": "model-00075-of-00086.safetensors", + "model.layers.110.mlp.up_proj.weight": "model-00075-of-00086.safetensors", + "model.layers.110.mlp.up_proj.weight_scale": "model-00075-of-00086.safetensors", + "model.layers.110.post_attention_layernorm.weight": "model-00075-of-00086.safetensors", + "model.layers.110.self_attn.k_proj.input_scale": "model-00075-of-00086.safetensors", + "model.layers.110.self_attn.k_proj.weight": "model-00075-of-00086.safetensors", + "model.layers.110.self_attn.k_proj.weight_scale": "model-00075-of-00086.safetensors", + "model.layers.110.self_attn.o_proj.input_scale": "model-00075-of-00086.safetensors", + "model.layers.110.self_attn.o_proj.weight": "model-00075-of-00086.safetensors", + "model.layers.110.self_attn.o_proj.weight_scale": "model-00075-of-00086.safetensors", + "model.layers.110.self_attn.q_proj.input_scale": "model-00075-of-00086.safetensors", + "model.layers.110.self_attn.q_proj.weight": "model-00075-of-00086.safetensors", + "model.layers.110.self_attn.q_proj.weight_scale": "model-00075-of-00086.safetensors", + "model.layers.110.self_attn.v_proj.input_scale": "model-00075-of-00086.safetensors", + "model.layers.110.self_attn.v_proj.weight": "model-00075-of-00086.safetensors", + "model.layers.110.self_attn.v_proj.weight_scale": "model-00075-of-00086.safetensors", + "model.layers.111.input_layernorm.weight": "model-00076-of-00086.safetensors", + "model.layers.111.mlp.down_proj.input_scale": "model-00076-of-00086.safetensors", + "model.layers.111.mlp.down_proj.weight": "model-00076-of-00086.safetensors", + "model.layers.111.mlp.down_proj.weight_scale": "model-00076-of-00086.safetensors", + "model.layers.111.mlp.gate_proj.input_scale": "model-00076-of-00086.safetensors", + "model.layers.111.mlp.gate_proj.weight": "model-00076-of-00086.safetensors", + "model.layers.111.mlp.gate_proj.weight_scale": "model-00076-of-00086.safetensors", + "model.layers.111.mlp.up_proj.input_scale": "model-00076-of-00086.safetensors", + "model.layers.111.mlp.up_proj.weight": "model-00076-of-00086.safetensors", + "model.layers.111.mlp.up_proj.weight_scale": "model-00076-of-00086.safetensors", + "model.layers.111.post_attention_layernorm.weight": "model-00076-of-00086.safetensors", + "model.layers.111.self_attn.k_proj.input_scale": "model-00075-of-00086.safetensors", + "model.layers.111.self_attn.k_proj.weight": "model-00075-of-00086.safetensors", + "model.layers.111.self_attn.k_proj.weight_scale": "model-00075-of-00086.safetensors", + "model.layers.111.self_attn.o_proj.input_scale": "model-00075-of-00086.safetensors", + "model.layers.111.self_attn.o_proj.weight": "model-00075-of-00086.safetensors", + "model.layers.111.self_attn.o_proj.weight_scale": "model-00075-of-00086.safetensors", + "model.layers.111.self_attn.q_proj.input_scale": "model-00075-of-00086.safetensors", + "model.layers.111.self_attn.q_proj.weight": "model-00075-of-00086.safetensors", + "model.layers.111.self_attn.q_proj.weight_scale": "model-00075-of-00086.safetensors", + "model.layers.111.self_attn.v_proj.input_scale": "model-00075-of-00086.safetensors", + "model.layers.111.self_attn.v_proj.weight": "model-00075-of-00086.safetensors", + "model.layers.111.self_attn.v_proj.weight_scale": "model-00075-of-00086.safetensors", + "model.layers.112.input_layernorm.weight": "model-00077-of-00086.safetensors", + "model.layers.112.mlp.down_proj.input_scale": "model-00077-of-00086.safetensors", + "model.layers.112.mlp.down_proj.weight": "model-00077-of-00086.safetensors", + "model.layers.112.mlp.down_proj.weight_scale": "model-00077-of-00086.safetensors", + "model.layers.112.mlp.gate_proj.input_scale": "model-00076-of-00086.safetensors", + "model.layers.112.mlp.gate_proj.weight": "model-00076-of-00086.safetensors", + "model.layers.112.mlp.gate_proj.weight_scale": "model-00076-of-00086.safetensors", + "model.layers.112.mlp.up_proj.input_scale": "model-00076-of-00086.safetensors", + "model.layers.112.mlp.up_proj.weight": "model-00076-of-00086.safetensors", + "model.layers.112.mlp.up_proj.weight_scale": "model-00076-of-00086.safetensors", + "model.layers.112.post_attention_layernorm.weight": "model-00077-of-00086.safetensors", + "model.layers.112.self_attn.k_proj.input_scale": "model-00076-of-00086.safetensors", + "model.layers.112.self_attn.k_proj.weight": "model-00076-of-00086.safetensors", + "model.layers.112.self_attn.k_proj.weight_scale": "model-00076-of-00086.safetensors", + "model.layers.112.self_attn.o_proj.input_scale": "model-00076-of-00086.safetensors", + "model.layers.112.self_attn.o_proj.weight": "model-00076-of-00086.safetensors", + "model.layers.112.self_attn.o_proj.weight_scale": "model-00076-of-00086.safetensors", + "model.layers.112.self_attn.q_proj.input_scale": "model-00076-of-00086.safetensors", + "model.layers.112.self_attn.q_proj.weight": "model-00076-of-00086.safetensors", + "model.layers.112.self_attn.q_proj.weight_scale": "model-00076-of-00086.safetensors", + "model.layers.112.self_attn.v_proj.input_scale": "model-00076-of-00086.safetensors", + "model.layers.112.self_attn.v_proj.weight": "model-00076-of-00086.safetensors", + "model.layers.112.self_attn.v_proj.weight_scale": "model-00076-of-00086.safetensors", + "model.layers.113.input_layernorm.weight": "model-00077-of-00086.safetensors", + "model.layers.113.mlp.down_proj.input_scale": "model-00077-of-00086.safetensors", + "model.layers.113.mlp.down_proj.weight": "model-00077-of-00086.safetensors", + "model.layers.113.mlp.down_proj.weight_scale": "model-00077-of-00086.safetensors", + "model.layers.113.mlp.gate_proj.input_scale": "model-00077-of-00086.safetensors", + "model.layers.113.mlp.gate_proj.weight": "model-00077-of-00086.safetensors", + "model.layers.113.mlp.gate_proj.weight_scale": "model-00077-of-00086.safetensors", + "model.layers.113.mlp.up_proj.input_scale": "model-00077-of-00086.safetensors", + "model.layers.113.mlp.up_proj.weight": "model-00077-of-00086.safetensors", + "model.layers.113.mlp.up_proj.weight_scale": "model-00077-of-00086.safetensors", + "model.layers.113.post_attention_layernorm.weight": "model-00077-of-00086.safetensors", + "model.layers.113.self_attn.k_proj.input_scale": "model-00077-of-00086.safetensors", + "model.layers.113.self_attn.k_proj.weight": "model-00077-of-00086.safetensors", + "model.layers.113.self_attn.k_proj.weight_scale": "model-00077-of-00086.safetensors", + "model.layers.113.self_attn.o_proj.input_scale": "model-00077-of-00086.safetensors", + "model.layers.113.self_attn.o_proj.weight": "model-00077-of-00086.safetensors", + "model.layers.113.self_attn.o_proj.weight_scale": "model-00077-of-00086.safetensors", + "model.layers.113.self_attn.q_proj.input_scale": "model-00077-of-00086.safetensors", + "model.layers.113.self_attn.q_proj.weight": "model-00077-of-00086.safetensors", + "model.layers.113.self_attn.q_proj.weight_scale": "model-00077-of-00086.safetensors", + "model.layers.113.self_attn.v_proj.input_scale": "model-00077-of-00086.safetensors", + "model.layers.113.self_attn.v_proj.weight": "model-00077-of-00086.safetensors", + "model.layers.113.self_attn.v_proj.weight_scale": "model-00077-of-00086.safetensors", + "model.layers.114.input_layernorm.weight": "model-00078-of-00086.safetensors", + "model.layers.114.mlp.down_proj.input_scale": "model-00078-of-00086.safetensors", + "model.layers.114.mlp.down_proj.weight": "model-00078-of-00086.safetensors", + "model.layers.114.mlp.down_proj.weight_scale": "model-00078-of-00086.safetensors", + "model.layers.114.mlp.gate_proj.input_scale": "model-00078-of-00086.safetensors", + "model.layers.114.mlp.gate_proj.weight": "model-00078-of-00086.safetensors", + "model.layers.114.mlp.gate_proj.weight_scale": "model-00078-of-00086.safetensors", + "model.layers.114.mlp.up_proj.input_scale": "model-00078-of-00086.safetensors", + "model.layers.114.mlp.up_proj.weight": "model-00078-of-00086.safetensors", + "model.layers.114.mlp.up_proj.weight_scale": "model-00078-of-00086.safetensors", + "model.layers.114.post_attention_layernorm.weight": "model-00078-of-00086.safetensors", + "model.layers.114.self_attn.k_proj.input_scale": "model-00077-of-00086.safetensors", + "model.layers.114.self_attn.k_proj.weight": "model-00077-of-00086.safetensors", + "model.layers.114.self_attn.k_proj.weight_scale": "model-00077-of-00086.safetensors", + "model.layers.114.self_attn.o_proj.input_scale": "model-00077-of-00086.safetensors", + "model.layers.114.self_attn.o_proj.weight": "model-00077-of-00086.safetensors", + "model.layers.114.self_attn.o_proj.weight_scale": "model-00077-of-00086.safetensors", + "model.layers.114.self_attn.q_proj.input_scale": "model-00077-of-00086.safetensors", + "model.layers.114.self_attn.q_proj.weight": "model-00077-of-00086.safetensors", + "model.layers.114.self_attn.q_proj.weight_scale": "model-00077-of-00086.safetensors", + "model.layers.114.self_attn.v_proj.input_scale": "model-00077-of-00086.safetensors", + "model.layers.114.self_attn.v_proj.weight": "model-00077-of-00086.safetensors", + "model.layers.114.self_attn.v_proj.weight_scale": "model-00077-of-00086.safetensors", + "model.layers.115.input_layernorm.weight": "model-00079-of-00086.safetensors", + "model.layers.115.mlp.down_proj.input_scale": "model-00079-of-00086.safetensors", + "model.layers.115.mlp.down_proj.weight": "model-00079-of-00086.safetensors", + "model.layers.115.mlp.down_proj.weight_scale": "model-00079-of-00086.safetensors", + "model.layers.115.mlp.gate_proj.input_scale": "model-00078-of-00086.safetensors", + "model.layers.115.mlp.gate_proj.weight": "model-00078-of-00086.safetensors", + "model.layers.115.mlp.gate_proj.weight_scale": "model-00078-of-00086.safetensors", + "model.layers.115.mlp.up_proj.input_scale": "model-00078-of-00086.safetensors", + "model.layers.115.mlp.up_proj.weight": "model-00078-of-00086.safetensors", + "model.layers.115.mlp.up_proj.weight_scale": "model-00078-of-00086.safetensors", + "model.layers.115.post_attention_layernorm.weight": "model-00079-of-00086.safetensors", + "model.layers.115.self_attn.k_proj.input_scale": "model-00078-of-00086.safetensors", + "model.layers.115.self_attn.k_proj.weight": "model-00078-of-00086.safetensors", + "model.layers.115.self_attn.k_proj.weight_scale": "model-00078-of-00086.safetensors", + "model.layers.115.self_attn.o_proj.input_scale": "model-00078-of-00086.safetensors", + "model.layers.115.self_attn.o_proj.weight": "model-00078-of-00086.safetensors", + "model.layers.115.self_attn.o_proj.weight_scale": "model-00078-of-00086.safetensors", + "model.layers.115.self_attn.q_proj.input_scale": "model-00078-of-00086.safetensors", + "model.layers.115.self_attn.q_proj.weight": "model-00078-of-00086.safetensors", + "model.layers.115.self_attn.q_proj.weight_scale": "model-00078-of-00086.safetensors", + "model.layers.115.self_attn.v_proj.input_scale": "model-00078-of-00086.safetensors", + "model.layers.115.self_attn.v_proj.weight": "model-00078-of-00086.safetensors", + "model.layers.115.self_attn.v_proj.weight_scale": "model-00078-of-00086.safetensors", + "model.layers.116.input_layernorm.weight": "model-00079-of-00086.safetensors", + "model.layers.116.mlp.down_proj.input_scale": "model-00079-of-00086.safetensors", + "model.layers.116.mlp.down_proj.weight": "model-00079-of-00086.safetensors", + "model.layers.116.mlp.down_proj.weight_scale": "model-00079-of-00086.safetensors", + "model.layers.116.mlp.gate_proj.input_scale": "model-00079-of-00086.safetensors", + "model.layers.116.mlp.gate_proj.weight": "model-00079-of-00086.safetensors", + "model.layers.116.mlp.gate_proj.weight_scale": "model-00079-of-00086.safetensors", + "model.layers.116.mlp.up_proj.input_scale": "model-00079-of-00086.safetensors", + "model.layers.116.mlp.up_proj.weight": "model-00079-of-00086.safetensors", + "model.layers.116.mlp.up_proj.weight_scale": "model-00079-of-00086.safetensors", + "model.layers.116.post_attention_layernorm.weight": "model-00079-of-00086.safetensors", + "model.layers.116.self_attn.k_proj.input_scale": "model-00079-of-00086.safetensors", + "model.layers.116.self_attn.k_proj.weight": "model-00079-of-00086.safetensors", + "model.layers.116.self_attn.k_proj.weight_scale": "model-00079-of-00086.safetensors", + "model.layers.116.self_attn.o_proj.input_scale": "model-00079-of-00086.safetensors", + "model.layers.116.self_attn.o_proj.weight": "model-00079-of-00086.safetensors", + "model.layers.116.self_attn.o_proj.weight_scale": "model-00079-of-00086.safetensors", + "model.layers.116.self_attn.q_proj.input_scale": "model-00079-of-00086.safetensors", + "model.layers.116.self_attn.q_proj.weight": "model-00079-of-00086.safetensors", + "model.layers.116.self_attn.q_proj.weight_scale": "model-00079-of-00086.safetensors", + "model.layers.116.self_attn.v_proj.input_scale": "model-00079-of-00086.safetensors", + "model.layers.116.self_attn.v_proj.weight": "model-00079-of-00086.safetensors", + "model.layers.116.self_attn.v_proj.weight_scale": "model-00079-of-00086.safetensors", + "model.layers.117.input_layernorm.weight": "model-00080-of-00086.safetensors", + "model.layers.117.mlp.down_proj.input_scale": "model-00080-of-00086.safetensors", + "model.layers.117.mlp.down_proj.weight": "model-00080-of-00086.safetensors", + "model.layers.117.mlp.down_proj.weight_scale": "model-00080-of-00086.safetensors", + "model.layers.117.mlp.gate_proj.input_scale": "model-00080-of-00086.safetensors", + "model.layers.117.mlp.gate_proj.weight": "model-00080-of-00086.safetensors", + "model.layers.117.mlp.gate_proj.weight_scale": "model-00080-of-00086.safetensors", + "model.layers.117.mlp.up_proj.input_scale": "model-00080-of-00086.safetensors", + "model.layers.117.mlp.up_proj.weight": "model-00080-of-00086.safetensors", + "model.layers.117.mlp.up_proj.weight_scale": "model-00080-of-00086.safetensors", + "model.layers.117.post_attention_layernorm.weight": "model-00080-of-00086.safetensors", + "model.layers.117.self_attn.k_proj.input_scale": "model-00079-of-00086.safetensors", + "model.layers.117.self_attn.k_proj.weight": "model-00079-of-00086.safetensors", + "model.layers.117.self_attn.k_proj.weight_scale": "model-00079-of-00086.safetensors", + "model.layers.117.self_attn.o_proj.input_scale": "model-00079-of-00086.safetensors", + "model.layers.117.self_attn.o_proj.weight": "model-00079-of-00086.safetensors", + "model.layers.117.self_attn.o_proj.weight_scale": "model-00079-of-00086.safetensors", + "model.layers.117.self_attn.q_proj.input_scale": "model-00079-of-00086.safetensors", + "model.layers.117.self_attn.q_proj.weight": "model-00079-of-00086.safetensors", + "model.layers.117.self_attn.q_proj.weight_scale": "model-00079-of-00086.safetensors", + "model.layers.117.self_attn.v_proj.input_scale": "model-00079-of-00086.safetensors", + "model.layers.117.self_attn.v_proj.weight": "model-00079-of-00086.safetensors", + "model.layers.117.self_attn.v_proj.weight_scale": "model-00079-of-00086.safetensors", + "model.layers.118.input_layernorm.weight": "model-00081-of-00086.safetensors", + "model.layers.118.mlp.down_proj.input_scale": "model-00081-of-00086.safetensors", + "model.layers.118.mlp.down_proj.weight": "model-00081-of-00086.safetensors", + "model.layers.118.mlp.down_proj.weight_scale": "model-00081-of-00086.safetensors", + "model.layers.118.mlp.gate_proj.input_scale": "model-00080-of-00086.safetensors", + "model.layers.118.mlp.gate_proj.weight": "model-00080-of-00086.safetensors", + "model.layers.118.mlp.gate_proj.weight_scale": "model-00080-of-00086.safetensors", + "model.layers.118.mlp.up_proj.input_scale": "model-00080-of-00086.safetensors", + "model.layers.118.mlp.up_proj.weight": "model-00080-of-00086.safetensors", + "model.layers.118.mlp.up_proj.weight_scale": "model-00080-of-00086.safetensors", + "model.layers.118.post_attention_layernorm.weight": "model-00081-of-00086.safetensors", + "model.layers.118.self_attn.k_proj.input_scale": "model-00080-of-00086.safetensors", + "model.layers.118.self_attn.k_proj.weight": "model-00080-of-00086.safetensors", + "model.layers.118.self_attn.k_proj.weight_scale": "model-00080-of-00086.safetensors", + "model.layers.118.self_attn.o_proj.input_scale": "model-00080-of-00086.safetensors", + "model.layers.118.self_attn.o_proj.weight": "model-00080-of-00086.safetensors", + "model.layers.118.self_attn.o_proj.weight_scale": "model-00080-of-00086.safetensors", + "model.layers.118.self_attn.q_proj.input_scale": "model-00080-of-00086.safetensors", + "model.layers.118.self_attn.q_proj.weight": "model-00080-of-00086.safetensors", + "model.layers.118.self_attn.q_proj.weight_scale": "model-00080-of-00086.safetensors", + "model.layers.118.self_attn.v_proj.input_scale": "model-00080-of-00086.safetensors", + "model.layers.118.self_attn.v_proj.weight": "model-00080-of-00086.safetensors", + "model.layers.118.self_attn.v_proj.weight_scale": "model-00080-of-00086.safetensors", + "model.layers.119.input_layernorm.weight": "model-00081-of-00086.safetensors", + "model.layers.119.mlp.down_proj.input_scale": "model-00081-of-00086.safetensors", + "model.layers.119.mlp.down_proj.weight": "model-00081-of-00086.safetensors", + "model.layers.119.mlp.down_proj.weight_scale": "model-00081-of-00086.safetensors", + "model.layers.119.mlp.gate_proj.input_scale": "model-00081-of-00086.safetensors", + "model.layers.119.mlp.gate_proj.weight": "model-00081-of-00086.safetensors", + "model.layers.119.mlp.gate_proj.weight_scale": "model-00081-of-00086.safetensors", + "model.layers.119.mlp.up_proj.input_scale": "model-00081-of-00086.safetensors", + "model.layers.119.mlp.up_proj.weight": "model-00081-of-00086.safetensors", + "model.layers.119.mlp.up_proj.weight_scale": "model-00081-of-00086.safetensors", + "model.layers.119.post_attention_layernorm.weight": "model-00081-of-00086.safetensors", + "model.layers.119.self_attn.k_proj.input_scale": "model-00081-of-00086.safetensors", + "model.layers.119.self_attn.k_proj.weight": "model-00081-of-00086.safetensors", + "model.layers.119.self_attn.k_proj.weight_scale": "model-00081-of-00086.safetensors", + "model.layers.119.self_attn.o_proj.input_scale": "model-00081-of-00086.safetensors", + "model.layers.119.self_attn.o_proj.weight": "model-00081-of-00086.safetensors", + "model.layers.119.self_attn.o_proj.weight_scale": "model-00081-of-00086.safetensors", + "model.layers.119.self_attn.q_proj.input_scale": "model-00081-of-00086.safetensors", + "model.layers.119.self_attn.q_proj.weight": "model-00081-of-00086.safetensors", + "model.layers.119.self_attn.q_proj.weight_scale": "model-00081-of-00086.safetensors", + "model.layers.119.self_attn.v_proj.input_scale": "model-00081-of-00086.safetensors", + "model.layers.119.self_attn.v_proj.weight": "model-00081-of-00086.safetensors", + "model.layers.119.self_attn.v_proj.weight_scale": "model-00081-of-00086.safetensors", + "model.layers.12.input_layernorm.weight": "model-00010-of-00086.safetensors", + "model.layers.12.mlp.down_proj.input_scale": "model-00010-of-00086.safetensors", + "model.layers.12.mlp.down_proj.weight": "model-00010-of-00086.safetensors", + "model.layers.12.mlp.down_proj.weight_scale": "model-00010-of-00086.safetensors", + "model.layers.12.mlp.gate_proj.input_scale": "model-00010-of-00086.safetensors", + "model.layers.12.mlp.gate_proj.weight": "model-00010-of-00086.safetensors", + "model.layers.12.mlp.gate_proj.weight_scale": "model-00010-of-00086.safetensors", + "model.layers.12.mlp.up_proj.input_scale": "model-00010-of-00086.safetensors", + "model.layers.12.mlp.up_proj.weight": "model-00010-of-00086.safetensors", + "model.layers.12.mlp.up_proj.weight_scale": "model-00010-of-00086.safetensors", + "model.layers.12.post_attention_layernorm.weight": "model-00010-of-00086.safetensors", + "model.layers.12.self_attn.k_proj.input_scale": "model-00009-of-00086.safetensors", + "model.layers.12.self_attn.k_proj.weight": "model-00009-of-00086.safetensors", + "model.layers.12.self_attn.k_proj.weight_scale": "model-00009-of-00086.safetensors", + "model.layers.12.self_attn.o_proj.input_scale": "model-00009-of-00086.safetensors", + "model.layers.12.self_attn.o_proj.weight": "model-00009-of-00086.safetensors", + "model.layers.12.self_attn.o_proj.weight_scale": "model-00009-of-00086.safetensors", + "model.layers.12.self_attn.q_proj.input_scale": "model-00009-of-00086.safetensors", + "model.layers.12.self_attn.q_proj.weight": "model-00009-of-00086.safetensors", + "model.layers.12.self_attn.q_proj.weight_scale": "model-00009-of-00086.safetensors", + "model.layers.12.self_attn.v_proj.input_scale": "model-00009-of-00086.safetensors", + "model.layers.12.self_attn.v_proj.weight": "model-00009-of-00086.safetensors", + "model.layers.12.self_attn.v_proj.weight_scale": "model-00009-of-00086.safetensors", + "model.layers.120.input_layernorm.weight": "model-00082-of-00086.safetensors", + "model.layers.120.mlp.down_proj.input_scale": "model-00082-of-00086.safetensors", + "model.layers.120.mlp.down_proj.weight": "model-00082-of-00086.safetensors", + "model.layers.120.mlp.down_proj.weight_scale": "model-00082-of-00086.safetensors", + "model.layers.120.mlp.gate_proj.input_scale": "model-00082-of-00086.safetensors", + "model.layers.120.mlp.gate_proj.weight": "model-00082-of-00086.safetensors", + "model.layers.120.mlp.gate_proj.weight_scale": "model-00082-of-00086.safetensors", + "model.layers.120.mlp.up_proj.input_scale": "model-00082-of-00086.safetensors", + "model.layers.120.mlp.up_proj.weight": "model-00082-of-00086.safetensors", + "model.layers.120.mlp.up_proj.weight_scale": "model-00082-of-00086.safetensors", + "model.layers.120.post_attention_layernorm.weight": "model-00082-of-00086.safetensors", + "model.layers.120.self_attn.k_proj.input_scale": "model-00081-of-00086.safetensors", + "model.layers.120.self_attn.k_proj.weight": "model-00081-of-00086.safetensors", + "model.layers.120.self_attn.k_proj.weight_scale": "model-00081-of-00086.safetensors", + "model.layers.120.self_attn.o_proj.input_scale": "model-00081-of-00086.safetensors", + "model.layers.120.self_attn.o_proj.weight": "model-00081-of-00086.safetensors", + "model.layers.120.self_attn.o_proj.weight_scale": "model-00081-of-00086.safetensors", + "model.layers.120.self_attn.q_proj.input_scale": "model-00081-of-00086.safetensors", + "model.layers.120.self_attn.q_proj.weight": "model-00081-of-00086.safetensors", + "model.layers.120.self_attn.q_proj.weight_scale": "model-00081-of-00086.safetensors", + "model.layers.120.self_attn.v_proj.input_scale": "model-00081-of-00086.safetensors", + "model.layers.120.self_attn.v_proj.weight": "model-00081-of-00086.safetensors", + "model.layers.120.self_attn.v_proj.weight_scale": "model-00081-of-00086.safetensors", + "model.layers.121.input_layernorm.weight": "model-00083-of-00086.safetensors", + "model.layers.121.mlp.down_proj.input_scale": "model-00083-of-00086.safetensors", + "model.layers.121.mlp.down_proj.weight": "model-00083-of-00086.safetensors", + "model.layers.121.mlp.down_proj.weight_scale": "model-00083-of-00086.safetensors", + "model.layers.121.mlp.gate_proj.input_scale": "model-00082-of-00086.safetensors", + "model.layers.121.mlp.gate_proj.weight": "model-00082-of-00086.safetensors", + "model.layers.121.mlp.gate_proj.weight_scale": "model-00082-of-00086.safetensors", + "model.layers.121.mlp.up_proj.input_scale": "model-00082-of-00086.safetensors", + "model.layers.121.mlp.up_proj.weight": "model-00082-of-00086.safetensors", + "model.layers.121.mlp.up_proj.weight_scale": "model-00082-of-00086.safetensors", + "model.layers.121.post_attention_layernorm.weight": "model-00083-of-00086.safetensors", + "model.layers.121.self_attn.k_proj.input_scale": "model-00082-of-00086.safetensors", + "model.layers.121.self_attn.k_proj.weight": "model-00082-of-00086.safetensors", + "model.layers.121.self_attn.k_proj.weight_scale": "model-00082-of-00086.safetensors", + "model.layers.121.self_attn.o_proj.input_scale": "model-00082-of-00086.safetensors", + "model.layers.121.self_attn.o_proj.weight": "model-00082-of-00086.safetensors", + "model.layers.121.self_attn.o_proj.weight_scale": "model-00082-of-00086.safetensors", + "model.layers.121.self_attn.q_proj.input_scale": "model-00082-of-00086.safetensors", + "model.layers.121.self_attn.q_proj.weight": "model-00082-of-00086.safetensors", + "model.layers.121.self_attn.q_proj.weight_scale": "model-00082-of-00086.safetensors", + "model.layers.121.self_attn.v_proj.input_scale": "model-00082-of-00086.safetensors", + "model.layers.121.self_attn.v_proj.weight": "model-00082-of-00086.safetensors", + "model.layers.121.self_attn.v_proj.weight_scale": "model-00082-of-00086.safetensors", + "model.layers.122.input_layernorm.weight": "model-00083-of-00086.safetensors", + "model.layers.122.mlp.down_proj.input_scale": "model-00083-of-00086.safetensors", + "model.layers.122.mlp.down_proj.weight": "model-00083-of-00086.safetensors", + "model.layers.122.mlp.down_proj.weight_scale": "model-00083-of-00086.safetensors", + "model.layers.122.mlp.gate_proj.input_scale": "model-00083-of-00086.safetensors", + "model.layers.122.mlp.gate_proj.weight": "model-00083-of-00086.safetensors", + "model.layers.122.mlp.gate_proj.weight_scale": "model-00083-of-00086.safetensors", + "model.layers.122.mlp.up_proj.input_scale": "model-00083-of-00086.safetensors", + "model.layers.122.mlp.up_proj.weight": "model-00083-of-00086.safetensors", + "model.layers.122.mlp.up_proj.weight_scale": "model-00083-of-00086.safetensors", + "model.layers.122.post_attention_layernorm.weight": "model-00083-of-00086.safetensors", + "model.layers.122.self_attn.k_proj.input_scale": "model-00083-of-00086.safetensors", + "model.layers.122.self_attn.k_proj.weight": "model-00083-of-00086.safetensors", + "model.layers.122.self_attn.k_proj.weight_scale": "model-00083-of-00086.safetensors", + "model.layers.122.self_attn.o_proj.input_scale": "model-00083-of-00086.safetensors", + "model.layers.122.self_attn.o_proj.weight": "model-00083-of-00086.safetensors", + "model.layers.122.self_attn.o_proj.weight_scale": "model-00083-of-00086.safetensors", + "model.layers.122.self_attn.q_proj.input_scale": "model-00083-of-00086.safetensors", + "model.layers.122.self_attn.q_proj.weight": "model-00083-of-00086.safetensors", + "model.layers.122.self_attn.q_proj.weight_scale": "model-00083-of-00086.safetensors", + "model.layers.122.self_attn.v_proj.input_scale": "model-00083-of-00086.safetensors", + "model.layers.122.self_attn.v_proj.weight": "model-00083-of-00086.safetensors", + "model.layers.122.self_attn.v_proj.weight_scale": "model-00083-of-00086.safetensors", + "model.layers.123.input_layernorm.weight": "model-00084-of-00086.safetensors", + "model.layers.123.mlp.down_proj.input_scale": "model-00084-of-00086.safetensors", + "model.layers.123.mlp.down_proj.weight": "model-00084-of-00086.safetensors", + "model.layers.123.mlp.down_proj.weight_scale": "model-00084-of-00086.safetensors", + "model.layers.123.mlp.gate_proj.input_scale": "model-00084-of-00086.safetensors", + "model.layers.123.mlp.gate_proj.weight": "model-00084-of-00086.safetensors", + "model.layers.123.mlp.gate_proj.weight_scale": "model-00084-of-00086.safetensors", + "model.layers.123.mlp.up_proj.input_scale": "model-00084-of-00086.safetensors", + "model.layers.123.mlp.up_proj.weight": "model-00084-of-00086.safetensors", + "model.layers.123.mlp.up_proj.weight_scale": "model-00084-of-00086.safetensors", + "model.layers.123.post_attention_layernorm.weight": "model-00084-of-00086.safetensors", + "model.layers.123.self_attn.k_proj.input_scale": "model-00083-of-00086.safetensors", + "model.layers.123.self_attn.k_proj.weight": "model-00083-of-00086.safetensors", + "model.layers.123.self_attn.k_proj.weight_scale": "model-00083-of-00086.safetensors", + "model.layers.123.self_attn.o_proj.input_scale": "model-00083-of-00086.safetensors", + "model.layers.123.self_attn.o_proj.weight": "model-00083-of-00086.safetensors", + "model.layers.123.self_attn.o_proj.weight_scale": "model-00083-of-00086.safetensors", + "model.layers.123.self_attn.q_proj.input_scale": "model-00083-of-00086.safetensors", + "model.layers.123.self_attn.q_proj.weight": "model-00083-of-00086.safetensors", + "model.layers.123.self_attn.q_proj.weight_scale": "model-00083-of-00086.safetensors", + "model.layers.123.self_attn.v_proj.input_scale": "model-00083-of-00086.safetensors", + "model.layers.123.self_attn.v_proj.weight": "model-00083-of-00086.safetensors", + "model.layers.123.self_attn.v_proj.weight_scale": "model-00083-of-00086.safetensors", + "model.layers.124.input_layernorm.weight": "model-00085-of-00086.safetensors", + "model.layers.124.mlp.down_proj.input_scale": "model-00085-of-00086.safetensors", + "model.layers.124.mlp.down_proj.weight": "model-00085-of-00086.safetensors", + "model.layers.124.mlp.down_proj.weight_scale": "model-00085-of-00086.safetensors", + "model.layers.124.mlp.gate_proj.input_scale": "model-00084-of-00086.safetensors", + "model.layers.124.mlp.gate_proj.weight": "model-00084-of-00086.safetensors", + "model.layers.124.mlp.gate_proj.weight_scale": "model-00084-of-00086.safetensors", + "model.layers.124.mlp.up_proj.input_scale": "model-00084-of-00086.safetensors", + "model.layers.124.mlp.up_proj.weight": "model-00084-of-00086.safetensors", + "model.layers.124.mlp.up_proj.weight_scale": "model-00084-of-00086.safetensors", + "model.layers.124.post_attention_layernorm.weight": "model-00085-of-00086.safetensors", + "model.layers.124.self_attn.k_proj.input_scale": "model-00084-of-00086.safetensors", + "model.layers.124.self_attn.k_proj.weight": "model-00084-of-00086.safetensors", + "model.layers.124.self_attn.k_proj.weight_scale": "model-00084-of-00086.safetensors", + "model.layers.124.self_attn.o_proj.input_scale": "model-00084-of-00086.safetensors", + "model.layers.124.self_attn.o_proj.weight": "model-00084-of-00086.safetensors", + "model.layers.124.self_attn.o_proj.weight_scale": "model-00084-of-00086.safetensors", + "model.layers.124.self_attn.q_proj.input_scale": "model-00084-of-00086.safetensors", + "model.layers.124.self_attn.q_proj.weight": "model-00084-of-00086.safetensors", + "model.layers.124.self_attn.q_proj.weight_scale": "model-00084-of-00086.safetensors", + "model.layers.124.self_attn.v_proj.input_scale": "model-00084-of-00086.safetensors", + "model.layers.124.self_attn.v_proj.weight": "model-00084-of-00086.safetensors", + "model.layers.124.self_attn.v_proj.weight_scale": "model-00084-of-00086.safetensors", + "model.layers.125.input_layernorm.weight": "model-00085-of-00086.safetensors", + "model.layers.125.mlp.down_proj.input_scale": "model-00085-of-00086.safetensors", + "model.layers.125.mlp.down_proj.weight": "model-00085-of-00086.safetensors", + "model.layers.125.mlp.down_proj.weight_scale": "model-00085-of-00086.safetensors", + "model.layers.125.mlp.gate_proj.input_scale": "model-00085-of-00086.safetensors", + "model.layers.125.mlp.gate_proj.weight": "model-00085-of-00086.safetensors", + "model.layers.125.mlp.gate_proj.weight_scale": "model-00085-of-00086.safetensors", + "model.layers.125.mlp.up_proj.input_scale": "model-00085-of-00086.safetensors", + "model.layers.125.mlp.up_proj.weight": "model-00085-of-00086.safetensors", + "model.layers.125.mlp.up_proj.weight_scale": "model-00085-of-00086.safetensors", + "model.layers.125.post_attention_layernorm.weight": "model-00085-of-00086.safetensors", + "model.layers.125.self_attn.k_proj.input_scale": "model-00085-of-00086.safetensors", + "model.layers.125.self_attn.k_proj.weight": "model-00085-of-00086.safetensors", + "model.layers.125.self_attn.k_proj.weight_scale": "model-00085-of-00086.safetensors", + "model.layers.125.self_attn.o_proj.input_scale": "model-00085-of-00086.safetensors", + "model.layers.125.self_attn.o_proj.weight": "model-00085-of-00086.safetensors", + "model.layers.125.self_attn.o_proj.weight_scale": "model-00085-of-00086.safetensors", + "model.layers.125.self_attn.q_proj.input_scale": "model-00085-of-00086.safetensors", + "model.layers.125.self_attn.q_proj.weight": "model-00085-of-00086.safetensors", + "model.layers.125.self_attn.q_proj.weight_scale": "model-00085-of-00086.safetensors", + "model.layers.125.self_attn.v_proj.input_scale": "model-00085-of-00086.safetensors", + "model.layers.125.self_attn.v_proj.weight": "model-00085-of-00086.safetensors", + "model.layers.125.self_attn.v_proj.weight_scale": "model-00085-of-00086.safetensors", + "model.layers.13.input_layernorm.weight": "model-00011-of-00086.safetensors", + "model.layers.13.mlp.down_proj.input_scale": "model-00011-of-00086.safetensors", + "model.layers.13.mlp.down_proj.weight": "model-00011-of-00086.safetensors", + "model.layers.13.mlp.down_proj.weight_scale": "model-00011-of-00086.safetensors", + "model.layers.13.mlp.gate_proj.input_scale": "model-00010-of-00086.safetensors", + "model.layers.13.mlp.gate_proj.weight": "model-00010-of-00086.safetensors", + "model.layers.13.mlp.gate_proj.weight_scale": "model-00010-of-00086.safetensors", + "model.layers.13.mlp.up_proj.input_scale": "model-00010-of-00086.safetensors", + "model.layers.13.mlp.up_proj.weight": "model-00010-of-00086.safetensors", + "model.layers.13.mlp.up_proj.weight_scale": "model-00010-of-00086.safetensors", + "model.layers.13.post_attention_layernorm.weight": "model-00011-of-00086.safetensors", + "model.layers.13.self_attn.k_proj.input_scale": "model-00010-of-00086.safetensors", + "model.layers.13.self_attn.k_proj.weight": "model-00010-of-00086.safetensors", + "model.layers.13.self_attn.k_proj.weight_scale": "model-00010-of-00086.safetensors", + "model.layers.13.self_attn.o_proj.input_scale": "model-00010-of-00086.safetensors", + "model.layers.13.self_attn.o_proj.weight": "model-00010-of-00086.safetensors", + "model.layers.13.self_attn.o_proj.weight_scale": "model-00010-of-00086.safetensors", + "model.layers.13.self_attn.q_proj.input_scale": "model-00010-of-00086.safetensors", + "model.layers.13.self_attn.q_proj.weight": "model-00010-of-00086.safetensors", + "model.layers.13.self_attn.q_proj.weight_scale": "model-00010-of-00086.safetensors", + "model.layers.13.self_attn.v_proj.input_scale": "model-00010-of-00086.safetensors", + "model.layers.13.self_attn.v_proj.weight": "model-00010-of-00086.safetensors", + "model.layers.13.self_attn.v_proj.weight_scale": "model-00010-of-00086.safetensors", + "model.layers.14.input_layernorm.weight": "model-00011-of-00086.safetensors", + "model.layers.14.mlp.down_proj.input_scale": "model-00011-of-00086.safetensors", + "model.layers.14.mlp.down_proj.weight": "model-00011-of-00086.safetensors", + "model.layers.14.mlp.down_proj.weight_scale": "model-00011-of-00086.safetensors", + "model.layers.14.mlp.gate_proj.input_scale": "model-00011-of-00086.safetensors", + "model.layers.14.mlp.gate_proj.weight": "model-00011-of-00086.safetensors", + "model.layers.14.mlp.gate_proj.weight_scale": "model-00011-of-00086.safetensors", + "model.layers.14.mlp.up_proj.input_scale": "model-00011-of-00086.safetensors", + "model.layers.14.mlp.up_proj.weight": "model-00011-of-00086.safetensors", + "model.layers.14.mlp.up_proj.weight_scale": "model-00011-of-00086.safetensors", + "model.layers.14.post_attention_layernorm.weight": "model-00011-of-00086.safetensors", + "model.layers.14.self_attn.k_proj.input_scale": "model-00011-of-00086.safetensors", + "model.layers.14.self_attn.k_proj.weight": "model-00011-of-00086.safetensors", + "model.layers.14.self_attn.k_proj.weight_scale": "model-00011-of-00086.safetensors", + "model.layers.14.self_attn.o_proj.input_scale": "model-00011-of-00086.safetensors", + "model.layers.14.self_attn.o_proj.weight": "model-00011-of-00086.safetensors", + "model.layers.14.self_attn.o_proj.weight_scale": "model-00011-of-00086.safetensors", + "model.layers.14.self_attn.q_proj.input_scale": "model-00011-of-00086.safetensors", + "model.layers.14.self_attn.q_proj.weight": "model-00011-of-00086.safetensors", + "model.layers.14.self_attn.q_proj.weight_scale": "model-00011-of-00086.safetensors", + "model.layers.14.self_attn.v_proj.input_scale": "model-00011-of-00086.safetensors", + "model.layers.14.self_attn.v_proj.weight": "model-00011-of-00086.safetensors", + "model.layers.14.self_attn.v_proj.weight_scale": "model-00011-of-00086.safetensors", + "model.layers.15.input_layernorm.weight": "model-00012-of-00086.safetensors", + "model.layers.15.mlp.down_proj.input_scale": "model-00012-of-00086.safetensors", + "model.layers.15.mlp.down_proj.weight": "model-00012-of-00086.safetensors", + "model.layers.15.mlp.down_proj.weight_scale": "model-00012-of-00086.safetensors", + "model.layers.15.mlp.gate_proj.input_scale": "model-00012-of-00086.safetensors", + "model.layers.15.mlp.gate_proj.weight": "model-00012-of-00086.safetensors", + "model.layers.15.mlp.gate_proj.weight_scale": "model-00012-of-00086.safetensors", + "model.layers.15.mlp.up_proj.input_scale": "model-00012-of-00086.safetensors", + "model.layers.15.mlp.up_proj.weight": "model-00012-of-00086.safetensors", + "model.layers.15.mlp.up_proj.weight_scale": "model-00012-of-00086.safetensors", + "model.layers.15.post_attention_layernorm.weight": "model-00012-of-00086.safetensors", + "model.layers.15.self_attn.k_proj.input_scale": "model-00011-of-00086.safetensors", + "model.layers.15.self_attn.k_proj.weight": "model-00011-of-00086.safetensors", + "model.layers.15.self_attn.k_proj.weight_scale": "model-00011-of-00086.safetensors", + "model.layers.15.self_attn.o_proj.input_scale": "model-00011-of-00086.safetensors", + "model.layers.15.self_attn.o_proj.weight": "model-00011-of-00086.safetensors", + "model.layers.15.self_attn.o_proj.weight_scale": "model-00011-of-00086.safetensors", + "model.layers.15.self_attn.q_proj.input_scale": "model-00011-of-00086.safetensors", + "model.layers.15.self_attn.q_proj.weight": "model-00011-of-00086.safetensors", + "model.layers.15.self_attn.q_proj.weight_scale": "model-00011-of-00086.safetensors", + "model.layers.15.self_attn.v_proj.input_scale": "model-00011-of-00086.safetensors", + "model.layers.15.self_attn.v_proj.weight": "model-00011-of-00086.safetensors", + "model.layers.15.self_attn.v_proj.weight_scale": "model-00011-of-00086.safetensors", + "model.layers.16.input_layernorm.weight": "model-00013-of-00086.safetensors", + "model.layers.16.mlp.down_proj.input_scale": "model-00013-of-00086.safetensors", + "model.layers.16.mlp.down_proj.weight": "model-00013-of-00086.safetensors", + "model.layers.16.mlp.down_proj.weight_scale": "model-00013-of-00086.safetensors", + "model.layers.16.mlp.gate_proj.input_scale": "model-00012-of-00086.safetensors", + "model.layers.16.mlp.gate_proj.weight": "model-00012-of-00086.safetensors", + "model.layers.16.mlp.gate_proj.weight_scale": "model-00012-of-00086.safetensors", + "model.layers.16.mlp.up_proj.input_scale": "model-00012-of-00086.safetensors", + "model.layers.16.mlp.up_proj.weight": "model-00012-of-00086.safetensors", + "model.layers.16.mlp.up_proj.weight_scale": "model-00012-of-00086.safetensors", + "model.layers.16.post_attention_layernorm.weight": "model-00013-of-00086.safetensors", + "model.layers.16.self_attn.k_proj.input_scale": "model-00012-of-00086.safetensors", + "model.layers.16.self_attn.k_proj.weight": "model-00012-of-00086.safetensors", + "model.layers.16.self_attn.k_proj.weight_scale": "model-00012-of-00086.safetensors", + "model.layers.16.self_attn.o_proj.input_scale": "model-00012-of-00086.safetensors", + "model.layers.16.self_attn.o_proj.weight": "model-00012-of-00086.safetensors", + "model.layers.16.self_attn.o_proj.weight_scale": "model-00012-of-00086.safetensors", + "model.layers.16.self_attn.q_proj.input_scale": "model-00012-of-00086.safetensors", + "model.layers.16.self_attn.q_proj.weight": "model-00012-of-00086.safetensors", + "model.layers.16.self_attn.q_proj.weight_scale": "model-00012-of-00086.safetensors", + "model.layers.16.self_attn.v_proj.input_scale": "model-00012-of-00086.safetensors", + "model.layers.16.self_attn.v_proj.weight": "model-00012-of-00086.safetensors", + "model.layers.16.self_attn.v_proj.weight_scale": "model-00012-of-00086.safetensors", + "model.layers.17.input_layernorm.weight": "model-00013-of-00086.safetensors", + "model.layers.17.mlp.down_proj.input_scale": "model-00013-of-00086.safetensors", + "model.layers.17.mlp.down_proj.weight": "model-00013-of-00086.safetensors", + "model.layers.17.mlp.down_proj.weight_scale": "model-00013-of-00086.safetensors", + "model.layers.17.mlp.gate_proj.input_scale": "model-00013-of-00086.safetensors", + "model.layers.17.mlp.gate_proj.weight": "model-00013-of-00086.safetensors", + "model.layers.17.mlp.gate_proj.weight_scale": "model-00013-of-00086.safetensors", + "model.layers.17.mlp.up_proj.input_scale": "model-00013-of-00086.safetensors", + "model.layers.17.mlp.up_proj.weight": "model-00013-of-00086.safetensors", + "model.layers.17.mlp.up_proj.weight_scale": "model-00013-of-00086.safetensors", + "model.layers.17.post_attention_layernorm.weight": "model-00013-of-00086.safetensors", + "model.layers.17.self_attn.k_proj.input_scale": "model-00013-of-00086.safetensors", + "model.layers.17.self_attn.k_proj.weight": "model-00013-of-00086.safetensors", + "model.layers.17.self_attn.k_proj.weight_scale": "model-00013-of-00086.safetensors", + "model.layers.17.self_attn.o_proj.input_scale": "model-00013-of-00086.safetensors", + "model.layers.17.self_attn.o_proj.weight": "model-00013-of-00086.safetensors", + "model.layers.17.self_attn.o_proj.weight_scale": "model-00013-of-00086.safetensors", + "model.layers.17.self_attn.q_proj.input_scale": "model-00013-of-00086.safetensors", + "model.layers.17.self_attn.q_proj.weight": "model-00013-of-00086.safetensors", + "model.layers.17.self_attn.q_proj.weight_scale": "model-00013-of-00086.safetensors", + "model.layers.17.self_attn.v_proj.input_scale": "model-00013-of-00086.safetensors", + "model.layers.17.self_attn.v_proj.weight": "model-00013-of-00086.safetensors", + "model.layers.17.self_attn.v_proj.weight_scale": "model-00013-of-00086.safetensors", + "model.layers.18.input_layernorm.weight": "model-00014-of-00086.safetensors", + "model.layers.18.mlp.down_proj.input_scale": "model-00014-of-00086.safetensors", + "model.layers.18.mlp.down_proj.weight": "model-00014-of-00086.safetensors", + "model.layers.18.mlp.down_proj.weight_scale": "model-00014-of-00086.safetensors", + "model.layers.18.mlp.gate_proj.input_scale": "model-00014-of-00086.safetensors", + "model.layers.18.mlp.gate_proj.weight": "model-00014-of-00086.safetensors", + "model.layers.18.mlp.gate_proj.weight_scale": "model-00014-of-00086.safetensors", + "model.layers.18.mlp.up_proj.input_scale": "model-00014-of-00086.safetensors", + "model.layers.18.mlp.up_proj.weight": "model-00014-of-00086.safetensors", + "model.layers.18.mlp.up_proj.weight_scale": "model-00014-of-00086.safetensors", + "model.layers.18.post_attention_layernorm.weight": "model-00014-of-00086.safetensors", + "model.layers.18.self_attn.k_proj.input_scale": "model-00013-of-00086.safetensors", + "model.layers.18.self_attn.k_proj.weight": "model-00013-of-00086.safetensors", + "model.layers.18.self_attn.k_proj.weight_scale": "model-00013-of-00086.safetensors", + "model.layers.18.self_attn.o_proj.input_scale": "model-00013-of-00086.safetensors", + "model.layers.18.self_attn.o_proj.weight": "model-00013-of-00086.safetensors", + "model.layers.18.self_attn.o_proj.weight_scale": "model-00013-of-00086.safetensors", + "model.layers.18.self_attn.q_proj.input_scale": "model-00013-of-00086.safetensors", + "model.layers.18.self_attn.q_proj.weight": "model-00013-of-00086.safetensors", + "model.layers.18.self_attn.q_proj.weight_scale": "model-00013-of-00086.safetensors", + "model.layers.18.self_attn.v_proj.input_scale": "model-00013-of-00086.safetensors", + "model.layers.18.self_attn.v_proj.weight": "model-00013-of-00086.safetensors", + "model.layers.18.self_attn.v_proj.weight_scale": "model-00013-of-00086.safetensors", + "model.layers.19.input_layernorm.weight": "model-00015-of-00086.safetensors", + "model.layers.19.mlp.down_proj.input_scale": "model-00015-of-00086.safetensors", + "model.layers.19.mlp.down_proj.weight": "model-00015-of-00086.safetensors", + "model.layers.19.mlp.down_proj.weight_scale": "model-00015-of-00086.safetensors", + "model.layers.19.mlp.gate_proj.input_scale": "model-00014-of-00086.safetensors", + "model.layers.19.mlp.gate_proj.weight": "model-00014-of-00086.safetensors", + "model.layers.19.mlp.gate_proj.weight_scale": "model-00014-of-00086.safetensors", + "model.layers.19.mlp.up_proj.input_scale": "model-00014-of-00086.safetensors", + "model.layers.19.mlp.up_proj.weight": "model-00014-of-00086.safetensors", + "model.layers.19.mlp.up_proj.weight_scale": "model-00014-of-00086.safetensors", + "model.layers.19.post_attention_layernorm.weight": "model-00015-of-00086.safetensors", + "model.layers.19.self_attn.k_proj.input_scale": "model-00014-of-00086.safetensors", + "model.layers.19.self_attn.k_proj.weight": "model-00014-of-00086.safetensors", + "model.layers.19.self_attn.k_proj.weight_scale": "model-00014-of-00086.safetensors", + "model.layers.19.self_attn.o_proj.input_scale": "model-00014-of-00086.safetensors", + "model.layers.19.self_attn.o_proj.weight": "model-00014-of-00086.safetensors", + "model.layers.19.self_attn.o_proj.weight_scale": "model-00014-of-00086.safetensors", + "model.layers.19.self_attn.q_proj.input_scale": "model-00014-of-00086.safetensors", + "model.layers.19.self_attn.q_proj.weight": "model-00014-of-00086.safetensors", + "model.layers.19.self_attn.q_proj.weight_scale": "model-00014-of-00086.safetensors", + "model.layers.19.self_attn.v_proj.input_scale": "model-00014-of-00086.safetensors", + "model.layers.19.self_attn.v_proj.weight": "model-00014-of-00086.safetensors", + "model.layers.19.self_attn.v_proj.weight_scale": "model-00014-of-00086.safetensors", + "model.layers.2.input_layernorm.weight": "model-00003-of-00086.safetensors", + "model.layers.2.mlp.down_proj.input_scale": "model-00003-of-00086.safetensors", + "model.layers.2.mlp.down_proj.weight": "model-00003-of-00086.safetensors", + "model.layers.2.mlp.down_proj.weight_scale": "model-00003-of-00086.safetensors", + "model.layers.2.mlp.gate_proj.input_scale": "model-00003-of-00086.safetensors", + "model.layers.2.mlp.gate_proj.weight": "model-00003-of-00086.safetensors", + "model.layers.2.mlp.gate_proj.weight_scale": "model-00003-of-00086.safetensors", + "model.layers.2.mlp.up_proj.input_scale": "model-00003-of-00086.safetensors", + "model.layers.2.mlp.up_proj.weight": "model-00003-of-00086.safetensors", + "model.layers.2.mlp.up_proj.weight_scale": "model-00003-of-00086.safetensors", + "model.layers.2.post_attention_layernorm.weight": "model-00003-of-00086.safetensors", + "model.layers.2.self_attn.k_proj.input_scale": "model-00003-of-00086.safetensors", + "model.layers.2.self_attn.k_proj.weight": "model-00003-of-00086.safetensors", + "model.layers.2.self_attn.k_proj.weight_scale": "model-00003-of-00086.safetensors", + "model.layers.2.self_attn.o_proj.input_scale": "model-00003-of-00086.safetensors", + "model.layers.2.self_attn.o_proj.weight": "model-00003-of-00086.safetensors", + "model.layers.2.self_attn.o_proj.weight_scale": "model-00003-of-00086.safetensors", + "model.layers.2.self_attn.q_proj.input_scale": "model-00003-of-00086.safetensors", + "model.layers.2.self_attn.q_proj.weight": "model-00003-of-00086.safetensors", + "model.layers.2.self_attn.q_proj.weight_scale": "model-00003-of-00086.safetensors", + "model.layers.2.self_attn.v_proj.input_scale": "model-00003-of-00086.safetensors", + "model.layers.2.self_attn.v_proj.weight": "model-00003-of-00086.safetensors", + "model.layers.2.self_attn.v_proj.weight_scale": "model-00003-of-00086.safetensors", + "model.layers.20.input_layernorm.weight": "model-00015-of-00086.safetensors", + "model.layers.20.mlp.down_proj.input_scale": "model-00015-of-00086.safetensors", + "model.layers.20.mlp.down_proj.weight": "model-00015-of-00086.safetensors", + "model.layers.20.mlp.down_proj.weight_scale": "model-00015-of-00086.safetensors", + "model.layers.20.mlp.gate_proj.input_scale": "model-00015-of-00086.safetensors", + "model.layers.20.mlp.gate_proj.weight": "model-00015-of-00086.safetensors", + "model.layers.20.mlp.gate_proj.weight_scale": "model-00015-of-00086.safetensors", + "model.layers.20.mlp.up_proj.input_scale": "model-00015-of-00086.safetensors", + "model.layers.20.mlp.up_proj.weight": "model-00015-of-00086.safetensors", + "model.layers.20.mlp.up_proj.weight_scale": "model-00015-of-00086.safetensors", + "model.layers.20.post_attention_layernorm.weight": "model-00015-of-00086.safetensors", + "model.layers.20.self_attn.k_proj.input_scale": "model-00015-of-00086.safetensors", + "model.layers.20.self_attn.k_proj.weight": "model-00015-of-00086.safetensors", + "model.layers.20.self_attn.k_proj.weight_scale": "model-00015-of-00086.safetensors", + "model.layers.20.self_attn.o_proj.input_scale": "model-00015-of-00086.safetensors", + "model.layers.20.self_attn.o_proj.weight": "model-00015-of-00086.safetensors", + "model.layers.20.self_attn.o_proj.weight_scale": "model-00015-of-00086.safetensors", + "model.layers.20.self_attn.q_proj.input_scale": "model-00015-of-00086.safetensors", + "model.layers.20.self_attn.q_proj.weight": "model-00015-of-00086.safetensors", + "model.layers.20.self_attn.q_proj.weight_scale": "model-00015-of-00086.safetensors", + "model.layers.20.self_attn.v_proj.input_scale": "model-00015-of-00086.safetensors", + "model.layers.20.self_attn.v_proj.weight": "model-00015-of-00086.safetensors", + "model.layers.20.self_attn.v_proj.weight_scale": "model-00015-of-00086.safetensors", + "model.layers.21.input_layernorm.weight": "model-00016-of-00086.safetensors", + "model.layers.21.mlp.down_proj.input_scale": "model-00016-of-00086.safetensors", + "model.layers.21.mlp.down_proj.weight": "model-00016-of-00086.safetensors", + "model.layers.21.mlp.down_proj.weight_scale": "model-00016-of-00086.safetensors", + "model.layers.21.mlp.gate_proj.input_scale": "model-00016-of-00086.safetensors", + "model.layers.21.mlp.gate_proj.weight": "model-00016-of-00086.safetensors", + "model.layers.21.mlp.gate_proj.weight_scale": "model-00016-of-00086.safetensors", + "model.layers.21.mlp.up_proj.input_scale": "model-00016-of-00086.safetensors", + "model.layers.21.mlp.up_proj.weight": "model-00016-of-00086.safetensors", + "model.layers.21.mlp.up_proj.weight_scale": "model-00016-of-00086.safetensors", + "model.layers.21.post_attention_layernorm.weight": "model-00016-of-00086.safetensors", + "model.layers.21.self_attn.k_proj.input_scale": "model-00015-of-00086.safetensors", + "model.layers.21.self_attn.k_proj.weight": "model-00015-of-00086.safetensors", + "model.layers.21.self_attn.k_proj.weight_scale": "model-00015-of-00086.safetensors", + "model.layers.21.self_attn.o_proj.input_scale": "model-00015-of-00086.safetensors", + "model.layers.21.self_attn.o_proj.weight": "model-00015-of-00086.safetensors", + "model.layers.21.self_attn.o_proj.weight_scale": "model-00015-of-00086.safetensors", + "model.layers.21.self_attn.q_proj.input_scale": "model-00015-of-00086.safetensors", + "model.layers.21.self_attn.q_proj.weight": "model-00015-of-00086.safetensors", + "model.layers.21.self_attn.q_proj.weight_scale": "model-00015-of-00086.safetensors", + "model.layers.21.self_attn.v_proj.input_scale": "model-00015-of-00086.safetensors", + "model.layers.21.self_attn.v_proj.weight": "model-00015-of-00086.safetensors", + "model.layers.21.self_attn.v_proj.weight_scale": "model-00015-of-00086.safetensors", + "model.layers.22.input_layernorm.weight": "model-00017-of-00086.safetensors", + "model.layers.22.mlp.down_proj.input_scale": "model-00017-of-00086.safetensors", + "model.layers.22.mlp.down_proj.weight": "model-00017-of-00086.safetensors", + "model.layers.22.mlp.down_proj.weight_scale": "model-00017-of-00086.safetensors", + "model.layers.22.mlp.gate_proj.input_scale": "model-00016-of-00086.safetensors", + "model.layers.22.mlp.gate_proj.weight": "model-00016-of-00086.safetensors", + "model.layers.22.mlp.gate_proj.weight_scale": "model-00016-of-00086.safetensors", + "model.layers.22.mlp.up_proj.input_scale": "model-00016-of-00086.safetensors", + "model.layers.22.mlp.up_proj.weight": "model-00016-of-00086.safetensors", + "model.layers.22.mlp.up_proj.weight_scale": "model-00016-of-00086.safetensors", + "model.layers.22.post_attention_layernorm.weight": "model-00017-of-00086.safetensors", + "model.layers.22.self_attn.k_proj.input_scale": "model-00016-of-00086.safetensors", + "model.layers.22.self_attn.k_proj.weight": "model-00016-of-00086.safetensors", + "model.layers.22.self_attn.k_proj.weight_scale": "model-00016-of-00086.safetensors", + "model.layers.22.self_attn.o_proj.input_scale": "model-00016-of-00086.safetensors", + "model.layers.22.self_attn.o_proj.weight": "model-00016-of-00086.safetensors", + "model.layers.22.self_attn.o_proj.weight_scale": "model-00016-of-00086.safetensors", + "model.layers.22.self_attn.q_proj.input_scale": "model-00016-of-00086.safetensors", + "model.layers.22.self_attn.q_proj.weight": "model-00016-of-00086.safetensors", + "model.layers.22.self_attn.q_proj.weight_scale": "model-00016-of-00086.safetensors", + "model.layers.22.self_attn.v_proj.input_scale": "model-00016-of-00086.safetensors", + "model.layers.22.self_attn.v_proj.weight": "model-00016-of-00086.safetensors", + "model.layers.22.self_attn.v_proj.weight_scale": "model-00016-of-00086.safetensors", + "model.layers.23.input_layernorm.weight": "model-00017-of-00086.safetensors", + "model.layers.23.mlp.down_proj.input_scale": "model-00017-of-00086.safetensors", + "model.layers.23.mlp.down_proj.weight": "model-00017-of-00086.safetensors", + "model.layers.23.mlp.down_proj.weight_scale": "model-00017-of-00086.safetensors", + "model.layers.23.mlp.gate_proj.input_scale": "model-00017-of-00086.safetensors", + "model.layers.23.mlp.gate_proj.weight": "model-00017-of-00086.safetensors", + "model.layers.23.mlp.gate_proj.weight_scale": "model-00017-of-00086.safetensors", + "model.layers.23.mlp.up_proj.input_scale": "model-00017-of-00086.safetensors", + "model.layers.23.mlp.up_proj.weight": "model-00017-of-00086.safetensors", + "model.layers.23.mlp.up_proj.weight_scale": "model-00017-of-00086.safetensors", + "model.layers.23.post_attention_layernorm.weight": "model-00017-of-00086.safetensors", + "model.layers.23.self_attn.k_proj.input_scale": "model-00017-of-00086.safetensors", + "model.layers.23.self_attn.k_proj.weight": "model-00017-of-00086.safetensors", + "model.layers.23.self_attn.k_proj.weight_scale": "model-00017-of-00086.safetensors", + "model.layers.23.self_attn.o_proj.input_scale": "model-00017-of-00086.safetensors", + "model.layers.23.self_attn.o_proj.weight": "model-00017-of-00086.safetensors", + "model.layers.23.self_attn.o_proj.weight_scale": "model-00017-of-00086.safetensors", + "model.layers.23.self_attn.q_proj.input_scale": "model-00017-of-00086.safetensors", + "model.layers.23.self_attn.q_proj.weight": "model-00017-of-00086.safetensors", + "model.layers.23.self_attn.q_proj.weight_scale": "model-00017-of-00086.safetensors", + "model.layers.23.self_attn.v_proj.input_scale": "model-00017-of-00086.safetensors", + "model.layers.23.self_attn.v_proj.weight": "model-00017-of-00086.safetensors", + "model.layers.23.self_attn.v_proj.weight_scale": "model-00017-of-00086.safetensors", + "model.layers.24.input_layernorm.weight": "model-00018-of-00086.safetensors", + "model.layers.24.mlp.down_proj.input_scale": "model-00018-of-00086.safetensors", + "model.layers.24.mlp.down_proj.weight": "model-00018-of-00086.safetensors", + "model.layers.24.mlp.down_proj.weight_scale": "model-00018-of-00086.safetensors", + "model.layers.24.mlp.gate_proj.input_scale": "model-00018-of-00086.safetensors", + "model.layers.24.mlp.gate_proj.weight": "model-00018-of-00086.safetensors", + "model.layers.24.mlp.gate_proj.weight_scale": "model-00018-of-00086.safetensors", + "model.layers.24.mlp.up_proj.input_scale": "model-00018-of-00086.safetensors", + "model.layers.24.mlp.up_proj.weight": "model-00018-of-00086.safetensors", + "model.layers.24.mlp.up_proj.weight_scale": "model-00018-of-00086.safetensors", + "model.layers.24.post_attention_layernorm.weight": "model-00018-of-00086.safetensors", + "model.layers.24.self_attn.k_proj.input_scale": "model-00017-of-00086.safetensors", + "model.layers.24.self_attn.k_proj.weight": "model-00017-of-00086.safetensors", + "model.layers.24.self_attn.k_proj.weight_scale": "model-00017-of-00086.safetensors", + "model.layers.24.self_attn.o_proj.input_scale": "model-00017-of-00086.safetensors", + "model.layers.24.self_attn.o_proj.weight": "model-00017-of-00086.safetensors", + "model.layers.24.self_attn.o_proj.weight_scale": "model-00017-of-00086.safetensors", + "model.layers.24.self_attn.q_proj.input_scale": "model-00017-of-00086.safetensors", + "model.layers.24.self_attn.q_proj.weight": "model-00017-of-00086.safetensors", + "model.layers.24.self_attn.q_proj.weight_scale": "model-00017-of-00086.safetensors", + "model.layers.24.self_attn.v_proj.input_scale": "model-00017-of-00086.safetensors", + "model.layers.24.self_attn.v_proj.weight": "model-00017-of-00086.safetensors", + "model.layers.24.self_attn.v_proj.weight_scale": "model-00017-of-00086.safetensors", + "model.layers.25.input_layernorm.weight": "model-00019-of-00086.safetensors", + "model.layers.25.mlp.down_proj.input_scale": "model-00019-of-00086.safetensors", + "model.layers.25.mlp.down_proj.weight": "model-00019-of-00086.safetensors", + "model.layers.25.mlp.down_proj.weight_scale": "model-00019-of-00086.safetensors", + "model.layers.25.mlp.gate_proj.input_scale": "model-00018-of-00086.safetensors", + "model.layers.25.mlp.gate_proj.weight": "model-00018-of-00086.safetensors", + "model.layers.25.mlp.gate_proj.weight_scale": "model-00018-of-00086.safetensors", + "model.layers.25.mlp.up_proj.input_scale": "model-00018-of-00086.safetensors", + "model.layers.25.mlp.up_proj.weight": "model-00018-of-00086.safetensors", + "model.layers.25.mlp.up_proj.weight_scale": "model-00018-of-00086.safetensors", + "model.layers.25.post_attention_layernorm.weight": "model-00019-of-00086.safetensors", + "model.layers.25.self_attn.k_proj.input_scale": "model-00018-of-00086.safetensors", + "model.layers.25.self_attn.k_proj.weight": "model-00018-of-00086.safetensors", + "model.layers.25.self_attn.k_proj.weight_scale": "model-00018-of-00086.safetensors", + "model.layers.25.self_attn.o_proj.input_scale": "model-00018-of-00086.safetensors", + "model.layers.25.self_attn.o_proj.weight": "model-00018-of-00086.safetensors", + "model.layers.25.self_attn.o_proj.weight_scale": "model-00018-of-00086.safetensors", + "model.layers.25.self_attn.q_proj.input_scale": "model-00018-of-00086.safetensors", + "model.layers.25.self_attn.q_proj.weight": "model-00018-of-00086.safetensors", + "model.layers.25.self_attn.q_proj.weight_scale": "model-00018-of-00086.safetensors", + "model.layers.25.self_attn.v_proj.input_scale": "model-00018-of-00086.safetensors", + "model.layers.25.self_attn.v_proj.weight": "model-00018-of-00086.safetensors", + "model.layers.25.self_attn.v_proj.weight_scale": "model-00018-of-00086.safetensors", + "model.layers.26.input_layernorm.weight": "model-00019-of-00086.safetensors", + "model.layers.26.mlp.down_proj.input_scale": "model-00019-of-00086.safetensors", + "model.layers.26.mlp.down_proj.weight": "model-00019-of-00086.safetensors", + "model.layers.26.mlp.down_proj.weight_scale": "model-00019-of-00086.safetensors", + "model.layers.26.mlp.gate_proj.input_scale": "model-00019-of-00086.safetensors", + "model.layers.26.mlp.gate_proj.weight": "model-00019-of-00086.safetensors", + "model.layers.26.mlp.gate_proj.weight_scale": "model-00019-of-00086.safetensors", + "model.layers.26.mlp.up_proj.input_scale": "model-00019-of-00086.safetensors", + "model.layers.26.mlp.up_proj.weight": "model-00019-of-00086.safetensors", + "model.layers.26.mlp.up_proj.weight_scale": "model-00019-of-00086.safetensors", + "model.layers.26.post_attention_layernorm.weight": "model-00019-of-00086.safetensors", + "model.layers.26.self_attn.k_proj.input_scale": "model-00019-of-00086.safetensors", + "model.layers.26.self_attn.k_proj.weight": "model-00019-of-00086.safetensors", + "model.layers.26.self_attn.k_proj.weight_scale": "model-00019-of-00086.safetensors", + "model.layers.26.self_attn.o_proj.input_scale": "model-00019-of-00086.safetensors", + "model.layers.26.self_attn.o_proj.weight": "model-00019-of-00086.safetensors", + "model.layers.26.self_attn.o_proj.weight_scale": "model-00019-of-00086.safetensors", + "model.layers.26.self_attn.q_proj.input_scale": "model-00019-of-00086.safetensors", + "model.layers.26.self_attn.q_proj.weight": "model-00019-of-00086.safetensors", + "model.layers.26.self_attn.q_proj.weight_scale": "model-00019-of-00086.safetensors", + "model.layers.26.self_attn.v_proj.input_scale": "model-00019-of-00086.safetensors", + "model.layers.26.self_attn.v_proj.weight": "model-00019-of-00086.safetensors", + "model.layers.26.self_attn.v_proj.weight_scale": "model-00019-of-00086.safetensors", + "model.layers.27.input_layernorm.weight": "model-00020-of-00086.safetensors", + "model.layers.27.mlp.down_proj.input_scale": "model-00020-of-00086.safetensors", + "model.layers.27.mlp.down_proj.weight": "model-00020-of-00086.safetensors", + "model.layers.27.mlp.down_proj.weight_scale": "model-00020-of-00086.safetensors", + "model.layers.27.mlp.gate_proj.input_scale": "model-00020-of-00086.safetensors", + "model.layers.27.mlp.gate_proj.weight": "model-00020-of-00086.safetensors", + "model.layers.27.mlp.gate_proj.weight_scale": "model-00020-of-00086.safetensors", + "model.layers.27.mlp.up_proj.input_scale": "model-00020-of-00086.safetensors", + "model.layers.27.mlp.up_proj.weight": "model-00020-of-00086.safetensors", + "model.layers.27.mlp.up_proj.weight_scale": "model-00020-of-00086.safetensors", + "model.layers.27.post_attention_layernorm.weight": "model-00020-of-00086.safetensors", + "model.layers.27.self_attn.k_proj.input_scale": "model-00019-of-00086.safetensors", + "model.layers.27.self_attn.k_proj.weight": "model-00019-of-00086.safetensors", + "model.layers.27.self_attn.k_proj.weight_scale": "model-00019-of-00086.safetensors", + "model.layers.27.self_attn.o_proj.input_scale": "model-00019-of-00086.safetensors", + "model.layers.27.self_attn.o_proj.weight": "model-00019-of-00086.safetensors", + "model.layers.27.self_attn.o_proj.weight_scale": "model-00019-of-00086.safetensors", + "model.layers.27.self_attn.q_proj.input_scale": "model-00019-of-00086.safetensors", + "model.layers.27.self_attn.q_proj.weight": "model-00019-of-00086.safetensors", + "model.layers.27.self_attn.q_proj.weight_scale": "model-00019-of-00086.safetensors", + "model.layers.27.self_attn.v_proj.input_scale": "model-00019-of-00086.safetensors", + "model.layers.27.self_attn.v_proj.weight": "model-00019-of-00086.safetensors", + "model.layers.27.self_attn.v_proj.weight_scale": "model-00019-of-00086.safetensors", + "model.layers.28.input_layernorm.weight": "model-00021-of-00086.safetensors", + "model.layers.28.mlp.down_proj.input_scale": "model-00021-of-00086.safetensors", + "model.layers.28.mlp.down_proj.weight": "model-00021-of-00086.safetensors", + "model.layers.28.mlp.down_proj.weight_scale": "model-00021-of-00086.safetensors", + "model.layers.28.mlp.gate_proj.input_scale": "model-00020-of-00086.safetensors", + "model.layers.28.mlp.gate_proj.weight": "model-00020-of-00086.safetensors", + "model.layers.28.mlp.gate_proj.weight_scale": "model-00020-of-00086.safetensors", + "model.layers.28.mlp.up_proj.input_scale": "model-00020-of-00086.safetensors", + "model.layers.28.mlp.up_proj.weight": "model-00020-of-00086.safetensors", + "model.layers.28.mlp.up_proj.weight_scale": "model-00020-of-00086.safetensors", + "model.layers.28.post_attention_layernorm.weight": "model-00021-of-00086.safetensors", + "model.layers.28.self_attn.k_proj.input_scale": "model-00020-of-00086.safetensors", + "model.layers.28.self_attn.k_proj.weight": "model-00020-of-00086.safetensors", + "model.layers.28.self_attn.k_proj.weight_scale": "model-00020-of-00086.safetensors", + "model.layers.28.self_attn.o_proj.input_scale": "model-00020-of-00086.safetensors", + "model.layers.28.self_attn.o_proj.weight": "model-00020-of-00086.safetensors", + "model.layers.28.self_attn.o_proj.weight_scale": "model-00020-of-00086.safetensors", + "model.layers.28.self_attn.q_proj.input_scale": "model-00020-of-00086.safetensors", + "model.layers.28.self_attn.q_proj.weight": "model-00020-of-00086.safetensors", + "model.layers.28.self_attn.q_proj.weight_scale": "model-00020-of-00086.safetensors", + "model.layers.28.self_attn.v_proj.input_scale": "model-00020-of-00086.safetensors", + "model.layers.28.self_attn.v_proj.weight": "model-00020-of-00086.safetensors", + "model.layers.28.self_attn.v_proj.weight_scale": "model-00020-of-00086.safetensors", + "model.layers.29.input_layernorm.weight": "model-00021-of-00086.safetensors", + "model.layers.29.mlp.down_proj.input_scale": "model-00021-of-00086.safetensors", + "model.layers.29.mlp.down_proj.weight": "model-00021-of-00086.safetensors", + "model.layers.29.mlp.down_proj.weight_scale": "model-00021-of-00086.safetensors", + "model.layers.29.mlp.gate_proj.input_scale": "model-00021-of-00086.safetensors", + "model.layers.29.mlp.gate_proj.weight": "model-00021-of-00086.safetensors", + "model.layers.29.mlp.gate_proj.weight_scale": "model-00021-of-00086.safetensors", + "model.layers.29.mlp.up_proj.input_scale": "model-00021-of-00086.safetensors", + "model.layers.29.mlp.up_proj.weight": "model-00021-of-00086.safetensors", + "model.layers.29.mlp.up_proj.weight_scale": "model-00021-of-00086.safetensors", + "model.layers.29.post_attention_layernorm.weight": "model-00021-of-00086.safetensors", + "model.layers.29.self_attn.k_proj.input_scale": "model-00021-of-00086.safetensors", + "model.layers.29.self_attn.k_proj.weight": "model-00021-of-00086.safetensors", + "model.layers.29.self_attn.k_proj.weight_scale": "model-00021-of-00086.safetensors", + "model.layers.29.self_attn.o_proj.input_scale": "model-00021-of-00086.safetensors", + "model.layers.29.self_attn.o_proj.weight": "model-00021-of-00086.safetensors", + "model.layers.29.self_attn.o_proj.weight_scale": "model-00021-of-00086.safetensors", + "model.layers.29.self_attn.q_proj.input_scale": "model-00021-of-00086.safetensors", + "model.layers.29.self_attn.q_proj.weight": "model-00021-of-00086.safetensors", + "model.layers.29.self_attn.q_proj.weight_scale": "model-00021-of-00086.safetensors", + "model.layers.29.self_attn.v_proj.input_scale": "model-00021-of-00086.safetensors", + "model.layers.29.self_attn.v_proj.weight": "model-00021-of-00086.safetensors", + "model.layers.29.self_attn.v_proj.weight_scale": "model-00021-of-00086.safetensors", + "model.layers.3.input_layernorm.weight": "model-00004-of-00086.safetensors", + "model.layers.3.mlp.down_proj.input_scale": "model-00004-of-00086.safetensors", + "model.layers.3.mlp.down_proj.weight": "model-00004-of-00086.safetensors", + "model.layers.3.mlp.down_proj.weight_scale": "model-00004-of-00086.safetensors", + "model.layers.3.mlp.gate_proj.input_scale": "model-00004-of-00086.safetensors", + "model.layers.3.mlp.gate_proj.weight": "model-00004-of-00086.safetensors", + "model.layers.3.mlp.gate_proj.weight_scale": "model-00004-of-00086.safetensors", + "model.layers.3.mlp.up_proj.input_scale": "model-00004-of-00086.safetensors", + "model.layers.3.mlp.up_proj.weight": "model-00004-of-00086.safetensors", + "model.layers.3.mlp.up_proj.weight_scale": "model-00004-of-00086.safetensors", + "model.layers.3.post_attention_layernorm.weight": "model-00004-of-00086.safetensors", + "model.layers.3.self_attn.k_proj.input_scale": "model-00003-of-00086.safetensors", + "model.layers.3.self_attn.k_proj.weight": "model-00003-of-00086.safetensors", + "model.layers.3.self_attn.k_proj.weight_scale": "model-00003-of-00086.safetensors", + "model.layers.3.self_attn.o_proj.input_scale": "model-00003-of-00086.safetensors", + "model.layers.3.self_attn.o_proj.weight": "model-00003-of-00086.safetensors", + "model.layers.3.self_attn.o_proj.weight_scale": "model-00003-of-00086.safetensors", + "model.layers.3.self_attn.q_proj.input_scale": "model-00003-of-00086.safetensors", + "model.layers.3.self_attn.q_proj.weight": "model-00003-of-00086.safetensors", + "model.layers.3.self_attn.q_proj.weight_scale": "model-00003-of-00086.safetensors", + "model.layers.3.self_attn.v_proj.input_scale": "model-00003-of-00086.safetensors", + "model.layers.3.self_attn.v_proj.weight": "model-00003-of-00086.safetensors", + "model.layers.3.self_attn.v_proj.weight_scale": "model-00003-of-00086.safetensors", + "model.layers.30.input_layernorm.weight": "model-00022-of-00086.safetensors", + "model.layers.30.mlp.down_proj.input_scale": "model-00022-of-00086.safetensors", + "model.layers.30.mlp.down_proj.weight": "model-00022-of-00086.safetensors", + "model.layers.30.mlp.down_proj.weight_scale": "model-00022-of-00086.safetensors", + "model.layers.30.mlp.gate_proj.input_scale": "model-00022-of-00086.safetensors", + "model.layers.30.mlp.gate_proj.weight": "model-00022-of-00086.safetensors", + "model.layers.30.mlp.gate_proj.weight_scale": "model-00022-of-00086.safetensors", + "model.layers.30.mlp.up_proj.input_scale": "model-00022-of-00086.safetensors", + "model.layers.30.mlp.up_proj.weight": "model-00022-of-00086.safetensors", + "model.layers.30.mlp.up_proj.weight_scale": "model-00022-of-00086.safetensors", + "model.layers.30.post_attention_layernorm.weight": "model-00022-of-00086.safetensors", + "model.layers.30.self_attn.k_proj.input_scale": "model-00021-of-00086.safetensors", + "model.layers.30.self_attn.k_proj.weight": "model-00021-of-00086.safetensors", + "model.layers.30.self_attn.k_proj.weight_scale": "model-00021-of-00086.safetensors", + "model.layers.30.self_attn.o_proj.input_scale": "model-00021-of-00086.safetensors", + "model.layers.30.self_attn.o_proj.weight": "model-00021-of-00086.safetensors", + "model.layers.30.self_attn.o_proj.weight_scale": "model-00021-of-00086.safetensors", + "model.layers.30.self_attn.q_proj.input_scale": "model-00021-of-00086.safetensors", + "model.layers.30.self_attn.q_proj.weight": "model-00021-of-00086.safetensors", + "model.layers.30.self_attn.q_proj.weight_scale": "model-00021-of-00086.safetensors", + "model.layers.30.self_attn.v_proj.input_scale": "model-00021-of-00086.safetensors", + "model.layers.30.self_attn.v_proj.weight": "model-00021-of-00086.safetensors", + "model.layers.30.self_attn.v_proj.weight_scale": "model-00021-of-00086.safetensors", + "model.layers.31.input_layernorm.weight": "model-00023-of-00086.safetensors", + "model.layers.31.mlp.down_proj.input_scale": "model-00023-of-00086.safetensors", + "model.layers.31.mlp.down_proj.weight": "model-00023-of-00086.safetensors", + "model.layers.31.mlp.down_proj.weight_scale": "model-00023-of-00086.safetensors", + "model.layers.31.mlp.gate_proj.input_scale": "model-00022-of-00086.safetensors", + "model.layers.31.mlp.gate_proj.weight": "model-00022-of-00086.safetensors", + "model.layers.31.mlp.gate_proj.weight_scale": "model-00022-of-00086.safetensors", + "model.layers.31.mlp.up_proj.input_scale": "model-00022-of-00086.safetensors", + "model.layers.31.mlp.up_proj.weight": "model-00022-of-00086.safetensors", + "model.layers.31.mlp.up_proj.weight_scale": "model-00022-of-00086.safetensors", + "model.layers.31.post_attention_layernorm.weight": "model-00023-of-00086.safetensors", + "model.layers.31.self_attn.k_proj.input_scale": "model-00022-of-00086.safetensors", + "model.layers.31.self_attn.k_proj.weight": "model-00022-of-00086.safetensors", + "model.layers.31.self_attn.k_proj.weight_scale": "model-00022-of-00086.safetensors", + "model.layers.31.self_attn.o_proj.input_scale": "model-00022-of-00086.safetensors", + "model.layers.31.self_attn.o_proj.weight": "model-00022-of-00086.safetensors", + "model.layers.31.self_attn.o_proj.weight_scale": "model-00022-of-00086.safetensors", + "model.layers.31.self_attn.q_proj.input_scale": "model-00022-of-00086.safetensors", + "model.layers.31.self_attn.q_proj.weight": "model-00022-of-00086.safetensors", + "model.layers.31.self_attn.q_proj.weight_scale": "model-00022-of-00086.safetensors", + "model.layers.31.self_attn.v_proj.input_scale": "model-00022-of-00086.safetensors", + "model.layers.31.self_attn.v_proj.weight": "model-00022-of-00086.safetensors", + "model.layers.31.self_attn.v_proj.weight_scale": "model-00022-of-00086.safetensors", + "model.layers.32.input_layernorm.weight": "model-00023-of-00086.safetensors", + "model.layers.32.mlp.down_proj.input_scale": "model-00023-of-00086.safetensors", + "model.layers.32.mlp.down_proj.weight": "model-00023-of-00086.safetensors", + "model.layers.32.mlp.down_proj.weight_scale": "model-00023-of-00086.safetensors", + "model.layers.32.mlp.gate_proj.input_scale": "model-00023-of-00086.safetensors", + "model.layers.32.mlp.gate_proj.weight": "model-00023-of-00086.safetensors", + "model.layers.32.mlp.gate_proj.weight_scale": "model-00023-of-00086.safetensors", + "model.layers.32.mlp.up_proj.input_scale": "model-00023-of-00086.safetensors", + "model.layers.32.mlp.up_proj.weight": "model-00023-of-00086.safetensors", + "model.layers.32.mlp.up_proj.weight_scale": "model-00023-of-00086.safetensors", + "model.layers.32.post_attention_layernorm.weight": "model-00023-of-00086.safetensors", + "model.layers.32.self_attn.k_proj.input_scale": "model-00023-of-00086.safetensors", + "model.layers.32.self_attn.k_proj.weight": "model-00023-of-00086.safetensors", + "model.layers.32.self_attn.k_proj.weight_scale": "model-00023-of-00086.safetensors", + "model.layers.32.self_attn.o_proj.input_scale": "model-00023-of-00086.safetensors", + "model.layers.32.self_attn.o_proj.weight": "model-00023-of-00086.safetensors", + "model.layers.32.self_attn.o_proj.weight_scale": "model-00023-of-00086.safetensors", + "model.layers.32.self_attn.q_proj.input_scale": "model-00023-of-00086.safetensors", + "model.layers.32.self_attn.q_proj.weight": "model-00023-of-00086.safetensors", + "model.layers.32.self_attn.q_proj.weight_scale": "model-00023-of-00086.safetensors", + "model.layers.32.self_attn.v_proj.input_scale": "model-00023-of-00086.safetensors", + "model.layers.32.self_attn.v_proj.weight": "model-00023-of-00086.safetensors", + "model.layers.32.self_attn.v_proj.weight_scale": "model-00023-of-00086.safetensors", + "model.layers.33.input_layernorm.weight": "model-00024-of-00086.safetensors", + "model.layers.33.mlp.down_proj.input_scale": "model-00024-of-00086.safetensors", + "model.layers.33.mlp.down_proj.weight": "model-00024-of-00086.safetensors", + "model.layers.33.mlp.down_proj.weight_scale": "model-00024-of-00086.safetensors", + "model.layers.33.mlp.gate_proj.input_scale": "model-00024-of-00086.safetensors", + "model.layers.33.mlp.gate_proj.weight": "model-00024-of-00086.safetensors", + "model.layers.33.mlp.gate_proj.weight_scale": "model-00024-of-00086.safetensors", + "model.layers.33.mlp.up_proj.input_scale": "model-00024-of-00086.safetensors", + "model.layers.33.mlp.up_proj.weight": "model-00024-of-00086.safetensors", + "model.layers.33.mlp.up_proj.weight_scale": "model-00024-of-00086.safetensors", + "model.layers.33.post_attention_layernorm.weight": "model-00024-of-00086.safetensors", + "model.layers.33.self_attn.k_proj.input_scale": "model-00023-of-00086.safetensors", + "model.layers.33.self_attn.k_proj.weight": "model-00023-of-00086.safetensors", + "model.layers.33.self_attn.k_proj.weight_scale": "model-00023-of-00086.safetensors", + "model.layers.33.self_attn.o_proj.input_scale": "model-00023-of-00086.safetensors", + "model.layers.33.self_attn.o_proj.weight": "model-00023-of-00086.safetensors", + "model.layers.33.self_attn.o_proj.weight_scale": "model-00023-of-00086.safetensors", + "model.layers.33.self_attn.q_proj.input_scale": "model-00023-of-00086.safetensors", + "model.layers.33.self_attn.q_proj.weight": "model-00023-of-00086.safetensors", + "model.layers.33.self_attn.q_proj.weight_scale": "model-00023-of-00086.safetensors", + "model.layers.33.self_attn.v_proj.input_scale": "model-00023-of-00086.safetensors", + "model.layers.33.self_attn.v_proj.weight": "model-00023-of-00086.safetensors", + "model.layers.33.self_attn.v_proj.weight_scale": "model-00023-of-00086.safetensors", + "model.layers.34.input_layernorm.weight": "model-00025-of-00086.safetensors", + "model.layers.34.mlp.down_proj.input_scale": "model-00025-of-00086.safetensors", + "model.layers.34.mlp.down_proj.weight": "model-00025-of-00086.safetensors", + "model.layers.34.mlp.down_proj.weight_scale": "model-00025-of-00086.safetensors", + "model.layers.34.mlp.gate_proj.input_scale": "model-00024-of-00086.safetensors", + "model.layers.34.mlp.gate_proj.weight": "model-00024-of-00086.safetensors", + "model.layers.34.mlp.gate_proj.weight_scale": "model-00024-of-00086.safetensors", + "model.layers.34.mlp.up_proj.input_scale": "model-00024-of-00086.safetensors", + "model.layers.34.mlp.up_proj.weight": "model-00024-of-00086.safetensors", + "model.layers.34.mlp.up_proj.weight_scale": "model-00024-of-00086.safetensors", + "model.layers.34.post_attention_layernorm.weight": "model-00025-of-00086.safetensors", + "model.layers.34.self_attn.k_proj.input_scale": "model-00024-of-00086.safetensors", + "model.layers.34.self_attn.k_proj.weight": "model-00024-of-00086.safetensors", + "model.layers.34.self_attn.k_proj.weight_scale": "model-00024-of-00086.safetensors", + "model.layers.34.self_attn.o_proj.input_scale": "model-00024-of-00086.safetensors", + "model.layers.34.self_attn.o_proj.weight": "model-00024-of-00086.safetensors", + "model.layers.34.self_attn.o_proj.weight_scale": "model-00024-of-00086.safetensors", + "model.layers.34.self_attn.q_proj.input_scale": "model-00024-of-00086.safetensors", + "model.layers.34.self_attn.q_proj.weight": "model-00024-of-00086.safetensors", + "model.layers.34.self_attn.q_proj.weight_scale": "model-00024-of-00086.safetensors", + "model.layers.34.self_attn.v_proj.input_scale": "model-00024-of-00086.safetensors", + "model.layers.34.self_attn.v_proj.weight": "model-00024-of-00086.safetensors", + "model.layers.34.self_attn.v_proj.weight_scale": "model-00024-of-00086.safetensors", + "model.layers.35.input_layernorm.weight": "model-00025-of-00086.safetensors", + "model.layers.35.mlp.down_proj.input_scale": "model-00025-of-00086.safetensors", + "model.layers.35.mlp.down_proj.weight": "model-00025-of-00086.safetensors", + "model.layers.35.mlp.down_proj.weight_scale": "model-00025-of-00086.safetensors", + "model.layers.35.mlp.gate_proj.input_scale": "model-00025-of-00086.safetensors", + "model.layers.35.mlp.gate_proj.weight": "model-00025-of-00086.safetensors", + "model.layers.35.mlp.gate_proj.weight_scale": "model-00025-of-00086.safetensors", + "model.layers.35.mlp.up_proj.input_scale": "model-00025-of-00086.safetensors", + "model.layers.35.mlp.up_proj.weight": "model-00025-of-00086.safetensors", + "model.layers.35.mlp.up_proj.weight_scale": "model-00025-of-00086.safetensors", + "model.layers.35.post_attention_layernorm.weight": "model-00025-of-00086.safetensors", + "model.layers.35.self_attn.k_proj.input_scale": "model-00025-of-00086.safetensors", + "model.layers.35.self_attn.k_proj.weight": "model-00025-of-00086.safetensors", + "model.layers.35.self_attn.k_proj.weight_scale": "model-00025-of-00086.safetensors", + "model.layers.35.self_attn.o_proj.input_scale": "model-00025-of-00086.safetensors", + "model.layers.35.self_attn.o_proj.weight": "model-00025-of-00086.safetensors", + "model.layers.35.self_attn.o_proj.weight_scale": "model-00025-of-00086.safetensors", + "model.layers.35.self_attn.q_proj.input_scale": "model-00025-of-00086.safetensors", + "model.layers.35.self_attn.q_proj.weight": "model-00025-of-00086.safetensors", + "model.layers.35.self_attn.q_proj.weight_scale": "model-00025-of-00086.safetensors", + "model.layers.35.self_attn.v_proj.input_scale": "model-00025-of-00086.safetensors", + "model.layers.35.self_attn.v_proj.weight": "model-00025-of-00086.safetensors", + "model.layers.35.self_attn.v_proj.weight_scale": "model-00025-of-00086.safetensors", + "model.layers.36.input_layernorm.weight": "model-00026-of-00086.safetensors", + "model.layers.36.mlp.down_proj.input_scale": "model-00026-of-00086.safetensors", + "model.layers.36.mlp.down_proj.weight": "model-00026-of-00086.safetensors", + "model.layers.36.mlp.down_proj.weight_scale": "model-00026-of-00086.safetensors", + "model.layers.36.mlp.gate_proj.input_scale": "model-00026-of-00086.safetensors", + "model.layers.36.mlp.gate_proj.weight": "model-00026-of-00086.safetensors", + "model.layers.36.mlp.gate_proj.weight_scale": "model-00026-of-00086.safetensors", + "model.layers.36.mlp.up_proj.input_scale": "model-00026-of-00086.safetensors", + "model.layers.36.mlp.up_proj.weight": "model-00026-of-00086.safetensors", + "model.layers.36.mlp.up_proj.weight_scale": "model-00026-of-00086.safetensors", + "model.layers.36.post_attention_layernorm.weight": "model-00026-of-00086.safetensors", + "model.layers.36.self_attn.k_proj.input_scale": "model-00025-of-00086.safetensors", + "model.layers.36.self_attn.k_proj.weight": "model-00025-of-00086.safetensors", + "model.layers.36.self_attn.k_proj.weight_scale": "model-00025-of-00086.safetensors", + "model.layers.36.self_attn.o_proj.input_scale": "model-00025-of-00086.safetensors", + "model.layers.36.self_attn.o_proj.weight": "model-00025-of-00086.safetensors", + "model.layers.36.self_attn.o_proj.weight_scale": "model-00025-of-00086.safetensors", + "model.layers.36.self_attn.q_proj.input_scale": "model-00025-of-00086.safetensors", + "model.layers.36.self_attn.q_proj.weight": "model-00025-of-00086.safetensors", + "model.layers.36.self_attn.q_proj.weight_scale": "model-00025-of-00086.safetensors", + "model.layers.36.self_attn.v_proj.input_scale": "model-00025-of-00086.safetensors", + "model.layers.36.self_attn.v_proj.weight": "model-00025-of-00086.safetensors", + "model.layers.36.self_attn.v_proj.weight_scale": "model-00025-of-00086.safetensors", + "model.layers.37.input_layernorm.weight": "model-00027-of-00086.safetensors", + "model.layers.37.mlp.down_proj.input_scale": "model-00027-of-00086.safetensors", + "model.layers.37.mlp.down_proj.weight": "model-00027-of-00086.safetensors", + "model.layers.37.mlp.down_proj.weight_scale": "model-00027-of-00086.safetensors", + "model.layers.37.mlp.gate_proj.input_scale": "model-00026-of-00086.safetensors", + "model.layers.37.mlp.gate_proj.weight": "model-00026-of-00086.safetensors", + "model.layers.37.mlp.gate_proj.weight_scale": "model-00026-of-00086.safetensors", + "model.layers.37.mlp.up_proj.input_scale": "model-00026-of-00086.safetensors", + "model.layers.37.mlp.up_proj.weight": "model-00026-of-00086.safetensors", + "model.layers.37.mlp.up_proj.weight_scale": "model-00026-of-00086.safetensors", + "model.layers.37.post_attention_layernorm.weight": "model-00027-of-00086.safetensors", + "model.layers.37.self_attn.k_proj.input_scale": "model-00026-of-00086.safetensors", + "model.layers.37.self_attn.k_proj.weight": "model-00026-of-00086.safetensors", + "model.layers.37.self_attn.k_proj.weight_scale": "model-00026-of-00086.safetensors", + "model.layers.37.self_attn.o_proj.input_scale": "model-00026-of-00086.safetensors", + "model.layers.37.self_attn.o_proj.weight": "model-00026-of-00086.safetensors", + "model.layers.37.self_attn.o_proj.weight_scale": "model-00026-of-00086.safetensors", + "model.layers.37.self_attn.q_proj.input_scale": "model-00026-of-00086.safetensors", + "model.layers.37.self_attn.q_proj.weight": "model-00026-of-00086.safetensors", + "model.layers.37.self_attn.q_proj.weight_scale": "model-00026-of-00086.safetensors", + "model.layers.37.self_attn.v_proj.input_scale": "model-00026-of-00086.safetensors", + "model.layers.37.self_attn.v_proj.weight": "model-00026-of-00086.safetensors", + "model.layers.37.self_attn.v_proj.weight_scale": "model-00026-of-00086.safetensors", + "model.layers.38.input_layernorm.weight": "model-00027-of-00086.safetensors", + "model.layers.38.mlp.down_proj.input_scale": "model-00027-of-00086.safetensors", + "model.layers.38.mlp.down_proj.weight": "model-00027-of-00086.safetensors", + "model.layers.38.mlp.down_proj.weight_scale": "model-00027-of-00086.safetensors", + "model.layers.38.mlp.gate_proj.input_scale": "model-00027-of-00086.safetensors", + "model.layers.38.mlp.gate_proj.weight": "model-00027-of-00086.safetensors", + "model.layers.38.mlp.gate_proj.weight_scale": "model-00027-of-00086.safetensors", + "model.layers.38.mlp.up_proj.input_scale": "model-00027-of-00086.safetensors", + "model.layers.38.mlp.up_proj.weight": "model-00027-of-00086.safetensors", + "model.layers.38.mlp.up_proj.weight_scale": "model-00027-of-00086.safetensors", + "model.layers.38.post_attention_layernorm.weight": "model-00027-of-00086.safetensors", + "model.layers.38.self_attn.k_proj.input_scale": "model-00027-of-00086.safetensors", + "model.layers.38.self_attn.k_proj.weight": "model-00027-of-00086.safetensors", + "model.layers.38.self_attn.k_proj.weight_scale": "model-00027-of-00086.safetensors", + "model.layers.38.self_attn.o_proj.input_scale": "model-00027-of-00086.safetensors", + "model.layers.38.self_attn.o_proj.weight": "model-00027-of-00086.safetensors", + "model.layers.38.self_attn.o_proj.weight_scale": "model-00027-of-00086.safetensors", + "model.layers.38.self_attn.q_proj.input_scale": "model-00027-of-00086.safetensors", + "model.layers.38.self_attn.q_proj.weight": "model-00027-of-00086.safetensors", + "model.layers.38.self_attn.q_proj.weight_scale": "model-00027-of-00086.safetensors", + "model.layers.38.self_attn.v_proj.input_scale": "model-00027-of-00086.safetensors", + "model.layers.38.self_attn.v_proj.weight": "model-00027-of-00086.safetensors", + "model.layers.38.self_attn.v_proj.weight_scale": "model-00027-of-00086.safetensors", + "model.layers.39.input_layernorm.weight": "model-00028-of-00086.safetensors", + "model.layers.39.mlp.down_proj.input_scale": "model-00028-of-00086.safetensors", + "model.layers.39.mlp.down_proj.weight": "model-00028-of-00086.safetensors", + "model.layers.39.mlp.down_proj.weight_scale": "model-00028-of-00086.safetensors", + "model.layers.39.mlp.gate_proj.input_scale": "model-00028-of-00086.safetensors", + "model.layers.39.mlp.gate_proj.weight": "model-00028-of-00086.safetensors", + "model.layers.39.mlp.gate_proj.weight_scale": "model-00028-of-00086.safetensors", + "model.layers.39.mlp.up_proj.input_scale": "model-00028-of-00086.safetensors", + "model.layers.39.mlp.up_proj.weight": "model-00028-of-00086.safetensors", + "model.layers.39.mlp.up_proj.weight_scale": "model-00028-of-00086.safetensors", + "model.layers.39.post_attention_layernorm.weight": "model-00028-of-00086.safetensors", + "model.layers.39.self_attn.k_proj.input_scale": "model-00027-of-00086.safetensors", + "model.layers.39.self_attn.k_proj.weight": "model-00027-of-00086.safetensors", + "model.layers.39.self_attn.k_proj.weight_scale": "model-00027-of-00086.safetensors", + "model.layers.39.self_attn.o_proj.input_scale": "model-00027-of-00086.safetensors", + "model.layers.39.self_attn.o_proj.weight": "model-00027-of-00086.safetensors", + "model.layers.39.self_attn.o_proj.weight_scale": "model-00027-of-00086.safetensors", + "model.layers.39.self_attn.q_proj.input_scale": "model-00027-of-00086.safetensors", + "model.layers.39.self_attn.q_proj.weight": "model-00027-of-00086.safetensors", + "model.layers.39.self_attn.q_proj.weight_scale": "model-00027-of-00086.safetensors", + "model.layers.39.self_attn.v_proj.input_scale": "model-00027-of-00086.safetensors", + "model.layers.39.self_attn.v_proj.weight": "model-00027-of-00086.safetensors", + "model.layers.39.self_attn.v_proj.weight_scale": "model-00027-of-00086.safetensors", + "model.layers.4.input_layernorm.weight": "model-00005-of-00086.safetensors", + "model.layers.4.mlp.down_proj.input_scale": "model-00005-of-00086.safetensors", + "model.layers.4.mlp.down_proj.weight": "model-00005-of-00086.safetensors", + "model.layers.4.mlp.down_proj.weight_scale": "model-00005-of-00086.safetensors", + "model.layers.4.mlp.gate_proj.input_scale": "model-00004-of-00086.safetensors", + "model.layers.4.mlp.gate_proj.weight": "model-00004-of-00086.safetensors", + "model.layers.4.mlp.gate_proj.weight_scale": "model-00004-of-00086.safetensors", + "model.layers.4.mlp.up_proj.input_scale": "model-00004-of-00086.safetensors", + "model.layers.4.mlp.up_proj.weight": "model-00004-of-00086.safetensors", + "model.layers.4.mlp.up_proj.weight_scale": "model-00004-of-00086.safetensors", + "model.layers.4.post_attention_layernorm.weight": "model-00005-of-00086.safetensors", + "model.layers.4.self_attn.k_proj.input_scale": "model-00004-of-00086.safetensors", + "model.layers.4.self_attn.k_proj.weight": "model-00004-of-00086.safetensors", + "model.layers.4.self_attn.k_proj.weight_scale": "model-00004-of-00086.safetensors", + "model.layers.4.self_attn.o_proj.input_scale": "model-00004-of-00086.safetensors", + "model.layers.4.self_attn.o_proj.weight": "model-00004-of-00086.safetensors", + "model.layers.4.self_attn.o_proj.weight_scale": "model-00004-of-00086.safetensors", + "model.layers.4.self_attn.q_proj.input_scale": "model-00004-of-00086.safetensors", + "model.layers.4.self_attn.q_proj.weight": "model-00004-of-00086.safetensors", + "model.layers.4.self_attn.q_proj.weight_scale": "model-00004-of-00086.safetensors", + "model.layers.4.self_attn.v_proj.input_scale": "model-00004-of-00086.safetensors", + "model.layers.4.self_attn.v_proj.weight": "model-00004-of-00086.safetensors", + "model.layers.4.self_attn.v_proj.weight_scale": "model-00004-of-00086.safetensors", + "model.layers.40.input_layernorm.weight": "model-00029-of-00086.safetensors", + "model.layers.40.mlp.down_proj.input_scale": "model-00029-of-00086.safetensors", + "model.layers.40.mlp.down_proj.weight": "model-00029-of-00086.safetensors", + "model.layers.40.mlp.down_proj.weight_scale": "model-00029-of-00086.safetensors", + "model.layers.40.mlp.gate_proj.input_scale": "model-00028-of-00086.safetensors", + "model.layers.40.mlp.gate_proj.weight": "model-00028-of-00086.safetensors", + "model.layers.40.mlp.gate_proj.weight_scale": "model-00028-of-00086.safetensors", + "model.layers.40.mlp.up_proj.input_scale": "model-00028-of-00086.safetensors", + "model.layers.40.mlp.up_proj.weight": "model-00028-of-00086.safetensors", + "model.layers.40.mlp.up_proj.weight_scale": "model-00028-of-00086.safetensors", + "model.layers.40.post_attention_layernorm.weight": "model-00029-of-00086.safetensors", + "model.layers.40.self_attn.k_proj.input_scale": "model-00028-of-00086.safetensors", + "model.layers.40.self_attn.k_proj.weight": "model-00028-of-00086.safetensors", + "model.layers.40.self_attn.k_proj.weight_scale": "model-00028-of-00086.safetensors", + "model.layers.40.self_attn.o_proj.input_scale": "model-00028-of-00086.safetensors", + "model.layers.40.self_attn.o_proj.weight": "model-00028-of-00086.safetensors", + "model.layers.40.self_attn.o_proj.weight_scale": "model-00028-of-00086.safetensors", + "model.layers.40.self_attn.q_proj.input_scale": "model-00028-of-00086.safetensors", + "model.layers.40.self_attn.q_proj.weight": "model-00028-of-00086.safetensors", + "model.layers.40.self_attn.q_proj.weight_scale": "model-00028-of-00086.safetensors", + "model.layers.40.self_attn.v_proj.input_scale": "model-00028-of-00086.safetensors", + "model.layers.40.self_attn.v_proj.weight": "model-00028-of-00086.safetensors", + "model.layers.40.self_attn.v_proj.weight_scale": "model-00028-of-00086.safetensors", + "model.layers.41.input_layernorm.weight": "model-00029-of-00086.safetensors", + "model.layers.41.mlp.down_proj.input_scale": "model-00029-of-00086.safetensors", + "model.layers.41.mlp.down_proj.weight": "model-00029-of-00086.safetensors", + "model.layers.41.mlp.down_proj.weight_scale": "model-00029-of-00086.safetensors", + "model.layers.41.mlp.gate_proj.input_scale": "model-00029-of-00086.safetensors", + "model.layers.41.mlp.gate_proj.weight": "model-00029-of-00086.safetensors", + "model.layers.41.mlp.gate_proj.weight_scale": "model-00029-of-00086.safetensors", + "model.layers.41.mlp.up_proj.input_scale": "model-00029-of-00086.safetensors", + "model.layers.41.mlp.up_proj.weight": "model-00029-of-00086.safetensors", + "model.layers.41.mlp.up_proj.weight_scale": "model-00029-of-00086.safetensors", + "model.layers.41.post_attention_layernorm.weight": "model-00029-of-00086.safetensors", + "model.layers.41.self_attn.k_proj.input_scale": "model-00029-of-00086.safetensors", + "model.layers.41.self_attn.k_proj.weight": "model-00029-of-00086.safetensors", + "model.layers.41.self_attn.k_proj.weight_scale": "model-00029-of-00086.safetensors", + "model.layers.41.self_attn.o_proj.input_scale": "model-00029-of-00086.safetensors", + "model.layers.41.self_attn.o_proj.weight": "model-00029-of-00086.safetensors", + "model.layers.41.self_attn.o_proj.weight_scale": "model-00029-of-00086.safetensors", + "model.layers.41.self_attn.q_proj.input_scale": "model-00029-of-00086.safetensors", + "model.layers.41.self_attn.q_proj.weight": "model-00029-of-00086.safetensors", + "model.layers.41.self_attn.q_proj.weight_scale": "model-00029-of-00086.safetensors", + "model.layers.41.self_attn.v_proj.input_scale": "model-00029-of-00086.safetensors", + "model.layers.41.self_attn.v_proj.weight": "model-00029-of-00086.safetensors", + "model.layers.41.self_attn.v_proj.weight_scale": "model-00029-of-00086.safetensors", + "model.layers.42.input_layernorm.weight": "model-00030-of-00086.safetensors", + "model.layers.42.mlp.down_proj.input_scale": "model-00030-of-00086.safetensors", + "model.layers.42.mlp.down_proj.weight": "model-00030-of-00086.safetensors", + "model.layers.42.mlp.down_proj.weight_scale": "model-00030-of-00086.safetensors", + "model.layers.42.mlp.gate_proj.input_scale": "model-00030-of-00086.safetensors", + "model.layers.42.mlp.gate_proj.weight": "model-00030-of-00086.safetensors", + "model.layers.42.mlp.gate_proj.weight_scale": "model-00030-of-00086.safetensors", + "model.layers.42.mlp.up_proj.input_scale": "model-00030-of-00086.safetensors", + "model.layers.42.mlp.up_proj.weight": "model-00030-of-00086.safetensors", + "model.layers.42.mlp.up_proj.weight_scale": "model-00030-of-00086.safetensors", + "model.layers.42.post_attention_layernorm.weight": "model-00030-of-00086.safetensors", + "model.layers.42.self_attn.k_proj.input_scale": "model-00029-of-00086.safetensors", + "model.layers.42.self_attn.k_proj.weight": "model-00029-of-00086.safetensors", + "model.layers.42.self_attn.k_proj.weight_scale": "model-00029-of-00086.safetensors", + "model.layers.42.self_attn.o_proj.input_scale": "model-00029-of-00086.safetensors", + "model.layers.42.self_attn.o_proj.weight": "model-00029-of-00086.safetensors", + "model.layers.42.self_attn.o_proj.weight_scale": "model-00029-of-00086.safetensors", + "model.layers.42.self_attn.q_proj.input_scale": "model-00029-of-00086.safetensors", + "model.layers.42.self_attn.q_proj.weight": "model-00029-of-00086.safetensors", + "model.layers.42.self_attn.q_proj.weight_scale": "model-00029-of-00086.safetensors", + "model.layers.42.self_attn.v_proj.input_scale": "model-00029-of-00086.safetensors", + "model.layers.42.self_attn.v_proj.weight": "model-00029-of-00086.safetensors", + "model.layers.42.self_attn.v_proj.weight_scale": "model-00029-of-00086.safetensors", + "model.layers.43.input_layernorm.weight": "model-00031-of-00086.safetensors", + "model.layers.43.mlp.down_proj.input_scale": "model-00031-of-00086.safetensors", + "model.layers.43.mlp.down_proj.weight": "model-00031-of-00086.safetensors", + "model.layers.43.mlp.down_proj.weight_scale": "model-00031-of-00086.safetensors", + "model.layers.43.mlp.gate_proj.input_scale": "model-00030-of-00086.safetensors", + "model.layers.43.mlp.gate_proj.weight": "model-00030-of-00086.safetensors", + "model.layers.43.mlp.gate_proj.weight_scale": "model-00030-of-00086.safetensors", + "model.layers.43.mlp.up_proj.input_scale": "model-00030-of-00086.safetensors", + "model.layers.43.mlp.up_proj.weight": "model-00030-of-00086.safetensors", + "model.layers.43.mlp.up_proj.weight_scale": "model-00030-of-00086.safetensors", + "model.layers.43.post_attention_layernorm.weight": "model-00031-of-00086.safetensors", + "model.layers.43.self_attn.k_proj.input_scale": "model-00030-of-00086.safetensors", + "model.layers.43.self_attn.k_proj.weight": "model-00030-of-00086.safetensors", + "model.layers.43.self_attn.k_proj.weight_scale": "model-00030-of-00086.safetensors", + "model.layers.43.self_attn.o_proj.input_scale": "model-00030-of-00086.safetensors", + "model.layers.43.self_attn.o_proj.weight": "model-00030-of-00086.safetensors", + "model.layers.43.self_attn.o_proj.weight_scale": "model-00030-of-00086.safetensors", + "model.layers.43.self_attn.q_proj.input_scale": "model-00030-of-00086.safetensors", + "model.layers.43.self_attn.q_proj.weight": "model-00030-of-00086.safetensors", + "model.layers.43.self_attn.q_proj.weight_scale": "model-00030-of-00086.safetensors", + "model.layers.43.self_attn.v_proj.input_scale": "model-00030-of-00086.safetensors", + "model.layers.43.self_attn.v_proj.weight": "model-00030-of-00086.safetensors", + "model.layers.43.self_attn.v_proj.weight_scale": "model-00030-of-00086.safetensors", + "model.layers.44.input_layernorm.weight": "model-00031-of-00086.safetensors", + "model.layers.44.mlp.down_proj.input_scale": "model-00031-of-00086.safetensors", + "model.layers.44.mlp.down_proj.weight": "model-00031-of-00086.safetensors", + "model.layers.44.mlp.down_proj.weight_scale": "model-00031-of-00086.safetensors", + "model.layers.44.mlp.gate_proj.input_scale": "model-00031-of-00086.safetensors", + "model.layers.44.mlp.gate_proj.weight": "model-00031-of-00086.safetensors", + "model.layers.44.mlp.gate_proj.weight_scale": "model-00031-of-00086.safetensors", + "model.layers.44.mlp.up_proj.input_scale": "model-00031-of-00086.safetensors", + "model.layers.44.mlp.up_proj.weight": "model-00031-of-00086.safetensors", + "model.layers.44.mlp.up_proj.weight_scale": "model-00031-of-00086.safetensors", + "model.layers.44.post_attention_layernorm.weight": "model-00031-of-00086.safetensors", + "model.layers.44.self_attn.k_proj.input_scale": "model-00031-of-00086.safetensors", + "model.layers.44.self_attn.k_proj.weight": "model-00031-of-00086.safetensors", + "model.layers.44.self_attn.k_proj.weight_scale": "model-00031-of-00086.safetensors", + "model.layers.44.self_attn.o_proj.input_scale": "model-00031-of-00086.safetensors", + "model.layers.44.self_attn.o_proj.weight": "model-00031-of-00086.safetensors", + "model.layers.44.self_attn.o_proj.weight_scale": "model-00031-of-00086.safetensors", + "model.layers.44.self_attn.q_proj.input_scale": "model-00031-of-00086.safetensors", + "model.layers.44.self_attn.q_proj.weight": "model-00031-of-00086.safetensors", + "model.layers.44.self_attn.q_proj.weight_scale": "model-00031-of-00086.safetensors", + "model.layers.44.self_attn.v_proj.input_scale": "model-00031-of-00086.safetensors", + "model.layers.44.self_attn.v_proj.weight": "model-00031-of-00086.safetensors", + "model.layers.44.self_attn.v_proj.weight_scale": "model-00031-of-00086.safetensors", + "model.layers.45.input_layernorm.weight": "model-00032-of-00086.safetensors", + "model.layers.45.mlp.down_proj.input_scale": "model-00032-of-00086.safetensors", + "model.layers.45.mlp.down_proj.weight": "model-00032-of-00086.safetensors", + "model.layers.45.mlp.down_proj.weight_scale": "model-00032-of-00086.safetensors", + "model.layers.45.mlp.gate_proj.input_scale": "model-00032-of-00086.safetensors", + "model.layers.45.mlp.gate_proj.weight": "model-00032-of-00086.safetensors", + "model.layers.45.mlp.gate_proj.weight_scale": "model-00032-of-00086.safetensors", + "model.layers.45.mlp.up_proj.input_scale": "model-00032-of-00086.safetensors", + "model.layers.45.mlp.up_proj.weight": "model-00032-of-00086.safetensors", + "model.layers.45.mlp.up_proj.weight_scale": "model-00032-of-00086.safetensors", + "model.layers.45.post_attention_layernorm.weight": "model-00032-of-00086.safetensors", + "model.layers.45.self_attn.k_proj.input_scale": "model-00031-of-00086.safetensors", + "model.layers.45.self_attn.k_proj.weight": "model-00031-of-00086.safetensors", + "model.layers.45.self_attn.k_proj.weight_scale": "model-00031-of-00086.safetensors", + "model.layers.45.self_attn.o_proj.input_scale": "model-00031-of-00086.safetensors", + "model.layers.45.self_attn.o_proj.weight": "model-00031-of-00086.safetensors", + "model.layers.45.self_attn.o_proj.weight_scale": "model-00031-of-00086.safetensors", + "model.layers.45.self_attn.q_proj.input_scale": "model-00031-of-00086.safetensors", + "model.layers.45.self_attn.q_proj.weight": "model-00031-of-00086.safetensors", + "model.layers.45.self_attn.q_proj.weight_scale": "model-00031-of-00086.safetensors", + "model.layers.45.self_attn.v_proj.input_scale": "model-00031-of-00086.safetensors", + "model.layers.45.self_attn.v_proj.weight": "model-00031-of-00086.safetensors", + "model.layers.45.self_attn.v_proj.weight_scale": "model-00031-of-00086.safetensors", + "model.layers.46.input_layernorm.weight": "model-00033-of-00086.safetensors", + "model.layers.46.mlp.down_proj.input_scale": "model-00033-of-00086.safetensors", + "model.layers.46.mlp.down_proj.weight": "model-00033-of-00086.safetensors", + "model.layers.46.mlp.down_proj.weight_scale": "model-00033-of-00086.safetensors", + "model.layers.46.mlp.gate_proj.input_scale": "model-00032-of-00086.safetensors", + "model.layers.46.mlp.gate_proj.weight": "model-00032-of-00086.safetensors", + "model.layers.46.mlp.gate_proj.weight_scale": "model-00032-of-00086.safetensors", + "model.layers.46.mlp.up_proj.input_scale": "model-00032-of-00086.safetensors", + "model.layers.46.mlp.up_proj.weight": "model-00032-of-00086.safetensors", + "model.layers.46.mlp.up_proj.weight_scale": "model-00032-of-00086.safetensors", + "model.layers.46.post_attention_layernorm.weight": "model-00033-of-00086.safetensors", + "model.layers.46.self_attn.k_proj.input_scale": "model-00032-of-00086.safetensors", + "model.layers.46.self_attn.k_proj.weight": "model-00032-of-00086.safetensors", + "model.layers.46.self_attn.k_proj.weight_scale": "model-00032-of-00086.safetensors", + "model.layers.46.self_attn.o_proj.input_scale": "model-00032-of-00086.safetensors", + "model.layers.46.self_attn.o_proj.weight": "model-00032-of-00086.safetensors", + "model.layers.46.self_attn.o_proj.weight_scale": "model-00032-of-00086.safetensors", + "model.layers.46.self_attn.q_proj.input_scale": "model-00032-of-00086.safetensors", + "model.layers.46.self_attn.q_proj.weight": "model-00032-of-00086.safetensors", + "model.layers.46.self_attn.q_proj.weight_scale": "model-00032-of-00086.safetensors", + "model.layers.46.self_attn.v_proj.input_scale": "model-00032-of-00086.safetensors", + "model.layers.46.self_attn.v_proj.weight": "model-00032-of-00086.safetensors", + "model.layers.46.self_attn.v_proj.weight_scale": "model-00032-of-00086.safetensors", + "model.layers.47.input_layernorm.weight": "model-00033-of-00086.safetensors", + "model.layers.47.mlp.down_proj.input_scale": "model-00033-of-00086.safetensors", + "model.layers.47.mlp.down_proj.weight": "model-00033-of-00086.safetensors", + "model.layers.47.mlp.down_proj.weight_scale": "model-00033-of-00086.safetensors", + "model.layers.47.mlp.gate_proj.input_scale": "model-00033-of-00086.safetensors", + "model.layers.47.mlp.gate_proj.weight": "model-00033-of-00086.safetensors", + "model.layers.47.mlp.gate_proj.weight_scale": "model-00033-of-00086.safetensors", + "model.layers.47.mlp.up_proj.input_scale": "model-00033-of-00086.safetensors", + "model.layers.47.mlp.up_proj.weight": "model-00033-of-00086.safetensors", + "model.layers.47.mlp.up_proj.weight_scale": "model-00033-of-00086.safetensors", + "model.layers.47.post_attention_layernorm.weight": "model-00033-of-00086.safetensors", + "model.layers.47.self_attn.k_proj.input_scale": "model-00033-of-00086.safetensors", + "model.layers.47.self_attn.k_proj.weight": "model-00033-of-00086.safetensors", + "model.layers.47.self_attn.k_proj.weight_scale": "model-00033-of-00086.safetensors", + "model.layers.47.self_attn.o_proj.input_scale": "model-00033-of-00086.safetensors", + "model.layers.47.self_attn.o_proj.weight": "model-00033-of-00086.safetensors", + "model.layers.47.self_attn.o_proj.weight_scale": "model-00033-of-00086.safetensors", + "model.layers.47.self_attn.q_proj.input_scale": "model-00033-of-00086.safetensors", + "model.layers.47.self_attn.q_proj.weight": "model-00033-of-00086.safetensors", + "model.layers.47.self_attn.q_proj.weight_scale": "model-00033-of-00086.safetensors", + "model.layers.47.self_attn.v_proj.input_scale": "model-00033-of-00086.safetensors", + "model.layers.47.self_attn.v_proj.weight": "model-00033-of-00086.safetensors", + "model.layers.47.self_attn.v_proj.weight_scale": "model-00033-of-00086.safetensors", + "model.layers.48.input_layernorm.weight": "model-00034-of-00086.safetensors", + "model.layers.48.mlp.down_proj.input_scale": "model-00034-of-00086.safetensors", + "model.layers.48.mlp.down_proj.weight": "model-00034-of-00086.safetensors", + "model.layers.48.mlp.down_proj.weight_scale": "model-00034-of-00086.safetensors", + "model.layers.48.mlp.gate_proj.input_scale": "model-00034-of-00086.safetensors", + "model.layers.48.mlp.gate_proj.weight": "model-00034-of-00086.safetensors", + "model.layers.48.mlp.gate_proj.weight_scale": "model-00034-of-00086.safetensors", + "model.layers.48.mlp.up_proj.input_scale": "model-00034-of-00086.safetensors", + "model.layers.48.mlp.up_proj.weight": "model-00034-of-00086.safetensors", + "model.layers.48.mlp.up_proj.weight_scale": "model-00034-of-00086.safetensors", + "model.layers.48.post_attention_layernorm.weight": "model-00034-of-00086.safetensors", + "model.layers.48.self_attn.k_proj.input_scale": "model-00033-of-00086.safetensors", + "model.layers.48.self_attn.k_proj.weight": "model-00033-of-00086.safetensors", + "model.layers.48.self_attn.k_proj.weight_scale": "model-00033-of-00086.safetensors", + "model.layers.48.self_attn.o_proj.input_scale": "model-00033-of-00086.safetensors", + "model.layers.48.self_attn.o_proj.weight": "model-00033-of-00086.safetensors", + "model.layers.48.self_attn.o_proj.weight_scale": "model-00033-of-00086.safetensors", + "model.layers.48.self_attn.q_proj.input_scale": "model-00033-of-00086.safetensors", + "model.layers.48.self_attn.q_proj.weight": "model-00033-of-00086.safetensors", + "model.layers.48.self_attn.q_proj.weight_scale": "model-00033-of-00086.safetensors", + "model.layers.48.self_attn.v_proj.input_scale": "model-00033-of-00086.safetensors", + "model.layers.48.self_attn.v_proj.weight": "model-00033-of-00086.safetensors", + "model.layers.48.self_attn.v_proj.weight_scale": "model-00033-of-00086.safetensors", + "model.layers.49.input_layernorm.weight": "model-00035-of-00086.safetensors", + "model.layers.49.mlp.down_proj.input_scale": "model-00035-of-00086.safetensors", + "model.layers.49.mlp.down_proj.weight": "model-00035-of-00086.safetensors", + "model.layers.49.mlp.down_proj.weight_scale": "model-00035-of-00086.safetensors", + "model.layers.49.mlp.gate_proj.input_scale": "model-00034-of-00086.safetensors", + "model.layers.49.mlp.gate_proj.weight": "model-00034-of-00086.safetensors", + "model.layers.49.mlp.gate_proj.weight_scale": "model-00034-of-00086.safetensors", + "model.layers.49.mlp.up_proj.input_scale": "model-00034-of-00086.safetensors", + "model.layers.49.mlp.up_proj.weight": "model-00034-of-00086.safetensors", + "model.layers.49.mlp.up_proj.weight_scale": "model-00034-of-00086.safetensors", + "model.layers.49.post_attention_layernorm.weight": "model-00035-of-00086.safetensors", + "model.layers.49.self_attn.k_proj.input_scale": "model-00034-of-00086.safetensors", + "model.layers.49.self_attn.k_proj.weight": "model-00034-of-00086.safetensors", + "model.layers.49.self_attn.k_proj.weight_scale": "model-00034-of-00086.safetensors", + "model.layers.49.self_attn.o_proj.input_scale": "model-00034-of-00086.safetensors", + "model.layers.49.self_attn.o_proj.weight": "model-00034-of-00086.safetensors", + "model.layers.49.self_attn.o_proj.weight_scale": "model-00034-of-00086.safetensors", + "model.layers.49.self_attn.q_proj.input_scale": "model-00034-of-00086.safetensors", + "model.layers.49.self_attn.q_proj.weight": "model-00034-of-00086.safetensors", + "model.layers.49.self_attn.q_proj.weight_scale": "model-00034-of-00086.safetensors", + "model.layers.49.self_attn.v_proj.input_scale": "model-00034-of-00086.safetensors", + "model.layers.49.self_attn.v_proj.weight": "model-00034-of-00086.safetensors", + "model.layers.49.self_attn.v_proj.weight_scale": "model-00034-of-00086.safetensors", + "model.layers.5.input_layernorm.weight": "model-00005-of-00086.safetensors", + "model.layers.5.mlp.down_proj.input_scale": "model-00005-of-00086.safetensors", + "model.layers.5.mlp.down_proj.weight": "model-00005-of-00086.safetensors", + "model.layers.5.mlp.down_proj.weight_scale": "model-00005-of-00086.safetensors", + "model.layers.5.mlp.gate_proj.input_scale": "model-00005-of-00086.safetensors", + "model.layers.5.mlp.gate_proj.weight": "model-00005-of-00086.safetensors", + "model.layers.5.mlp.gate_proj.weight_scale": "model-00005-of-00086.safetensors", + "model.layers.5.mlp.up_proj.input_scale": "model-00005-of-00086.safetensors", + "model.layers.5.mlp.up_proj.weight": "model-00005-of-00086.safetensors", + "model.layers.5.mlp.up_proj.weight_scale": "model-00005-of-00086.safetensors", + "model.layers.5.post_attention_layernorm.weight": "model-00005-of-00086.safetensors", + "model.layers.5.self_attn.k_proj.input_scale": "model-00005-of-00086.safetensors", + "model.layers.5.self_attn.k_proj.weight": "model-00005-of-00086.safetensors", + "model.layers.5.self_attn.k_proj.weight_scale": "model-00005-of-00086.safetensors", + "model.layers.5.self_attn.o_proj.input_scale": "model-00005-of-00086.safetensors", + "model.layers.5.self_attn.o_proj.weight": "model-00005-of-00086.safetensors", + "model.layers.5.self_attn.o_proj.weight_scale": "model-00005-of-00086.safetensors", + "model.layers.5.self_attn.q_proj.input_scale": "model-00005-of-00086.safetensors", + "model.layers.5.self_attn.q_proj.weight": "model-00005-of-00086.safetensors", + "model.layers.5.self_attn.q_proj.weight_scale": "model-00005-of-00086.safetensors", + "model.layers.5.self_attn.v_proj.input_scale": "model-00005-of-00086.safetensors", + "model.layers.5.self_attn.v_proj.weight": "model-00005-of-00086.safetensors", + "model.layers.5.self_attn.v_proj.weight_scale": "model-00005-of-00086.safetensors", + "model.layers.50.input_layernorm.weight": "model-00035-of-00086.safetensors", + "model.layers.50.mlp.down_proj.input_scale": "model-00035-of-00086.safetensors", + "model.layers.50.mlp.down_proj.weight": "model-00035-of-00086.safetensors", + "model.layers.50.mlp.down_proj.weight_scale": "model-00035-of-00086.safetensors", + "model.layers.50.mlp.gate_proj.input_scale": "model-00035-of-00086.safetensors", + "model.layers.50.mlp.gate_proj.weight": "model-00035-of-00086.safetensors", + "model.layers.50.mlp.gate_proj.weight_scale": "model-00035-of-00086.safetensors", + "model.layers.50.mlp.up_proj.input_scale": "model-00035-of-00086.safetensors", + "model.layers.50.mlp.up_proj.weight": "model-00035-of-00086.safetensors", + "model.layers.50.mlp.up_proj.weight_scale": "model-00035-of-00086.safetensors", + "model.layers.50.post_attention_layernorm.weight": "model-00035-of-00086.safetensors", + "model.layers.50.self_attn.k_proj.input_scale": "model-00035-of-00086.safetensors", + "model.layers.50.self_attn.k_proj.weight": "model-00035-of-00086.safetensors", + "model.layers.50.self_attn.k_proj.weight_scale": "model-00035-of-00086.safetensors", + "model.layers.50.self_attn.o_proj.input_scale": "model-00035-of-00086.safetensors", + "model.layers.50.self_attn.o_proj.weight": "model-00035-of-00086.safetensors", + "model.layers.50.self_attn.o_proj.weight_scale": "model-00035-of-00086.safetensors", + "model.layers.50.self_attn.q_proj.input_scale": "model-00035-of-00086.safetensors", + "model.layers.50.self_attn.q_proj.weight": "model-00035-of-00086.safetensors", + "model.layers.50.self_attn.q_proj.weight_scale": "model-00035-of-00086.safetensors", + "model.layers.50.self_attn.v_proj.input_scale": "model-00035-of-00086.safetensors", + "model.layers.50.self_attn.v_proj.weight": "model-00035-of-00086.safetensors", + "model.layers.50.self_attn.v_proj.weight_scale": "model-00035-of-00086.safetensors", + "model.layers.51.input_layernorm.weight": "model-00036-of-00086.safetensors", + "model.layers.51.mlp.down_proj.input_scale": "model-00036-of-00086.safetensors", + "model.layers.51.mlp.down_proj.weight": "model-00036-of-00086.safetensors", + "model.layers.51.mlp.down_proj.weight_scale": "model-00036-of-00086.safetensors", + "model.layers.51.mlp.gate_proj.input_scale": "model-00036-of-00086.safetensors", + "model.layers.51.mlp.gate_proj.weight": "model-00036-of-00086.safetensors", + "model.layers.51.mlp.gate_proj.weight_scale": "model-00036-of-00086.safetensors", + "model.layers.51.mlp.up_proj.input_scale": "model-00036-of-00086.safetensors", + "model.layers.51.mlp.up_proj.weight": "model-00036-of-00086.safetensors", + "model.layers.51.mlp.up_proj.weight_scale": "model-00036-of-00086.safetensors", + "model.layers.51.post_attention_layernorm.weight": "model-00036-of-00086.safetensors", + "model.layers.51.self_attn.k_proj.input_scale": "model-00035-of-00086.safetensors", + "model.layers.51.self_attn.k_proj.weight": "model-00035-of-00086.safetensors", + "model.layers.51.self_attn.k_proj.weight_scale": "model-00035-of-00086.safetensors", + "model.layers.51.self_attn.o_proj.input_scale": "model-00035-of-00086.safetensors", + "model.layers.51.self_attn.o_proj.weight": "model-00035-of-00086.safetensors", + "model.layers.51.self_attn.o_proj.weight_scale": "model-00035-of-00086.safetensors", + "model.layers.51.self_attn.q_proj.input_scale": "model-00035-of-00086.safetensors", + "model.layers.51.self_attn.q_proj.weight": "model-00035-of-00086.safetensors", + "model.layers.51.self_attn.q_proj.weight_scale": "model-00035-of-00086.safetensors", + "model.layers.51.self_attn.v_proj.input_scale": "model-00035-of-00086.safetensors", + "model.layers.51.self_attn.v_proj.weight": "model-00035-of-00086.safetensors", + "model.layers.51.self_attn.v_proj.weight_scale": "model-00035-of-00086.safetensors", + "model.layers.52.input_layernorm.weight": "model-00037-of-00086.safetensors", + "model.layers.52.mlp.down_proj.input_scale": "model-00037-of-00086.safetensors", + "model.layers.52.mlp.down_proj.weight": "model-00037-of-00086.safetensors", + "model.layers.52.mlp.down_proj.weight_scale": "model-00037-of-00086.safetensors", + "model.layers.52.mlp.gate_proj.input_scale": "model-00036-of-00086.safetensors", + "model.layers.52.mlp.gate_proj.weight": "model-00036-of-00086.safetensors", + "model.layers.52.mlp.gate_proj.weight_scale": "model-00036-of-00086.safetensors", + "model.layers.52.mlp.up_proj.input_scale": "model-00036-of-00086.safetensors", + "model.layers.52.mlp.up_proj.weight": "model-00036-of-00086.safetensors", + "model.layers.52.mlp.up_proj.weight_scale": "model-00036-of-00086.safetensors", + "model.layers.52.post_attention_layernorm.weight": "model-00037-of-00086.safetensors", + "model.layers.52.self_attn.k_proj.input_scale": "model-00036-of-00086.safetensors", + "model.layers.52.self_attn.k_proj.weight": "model-00036-of-00086.safetensors", + "model.layers.52.self_attn.k_proj.weight_scale": "model-00036-of-00086.safetensors", + "model.layers.52.self_attn.o_proj.input_scale": "model-00036-of-00086.safetensors", + "model.layers.52.self_attn.o_proj.weight": "model-00036-of-00086.safetensors", + "model.layers.52.self_attn.o_proj.weight_scale": "model-00036-of-00086.safetensors", + "model.layers.52.self_attn.q_proj.input_scale": "model-00036-of-00086.safetensors", + "model.layers.52.self_attn.q_proj.weight": "model-00036-of-00086.safetensors", + "model.layers.52.self_attn.q_proj.weight_scale": "model-00036-of-00086.safetensors", + "model.layers.52.self_attn.v_proj.input_scale": "model-00036-of-00086.safetensors", + "model.layers.52.self_attn.v_proj.weight": "model-00036-of-00086.safetensors", + "model.layers.52.self_attn.v_proj.weight_scale": "model-00036-of-00086.safetensors", + "model.layers.53.input_layernorm.weight": "model-00037-of-00086.safetensors", + "model.layers.53.mlp.down_proj.input_scale": "model-00037-of-00086.safetensors", + "model.layers.53.mlp.down_proj.weight": "model-00037-of-00086.safetensors", + "model.layers.53.mlp.down_proj.weight_scale": "model-00037-of-00086.safetensors", + "model.layers.53.mlp.gate_proj.input_scale": "model-00037-of-00086.safetensors", + "model.layers.53.mlp.gate_proj.weight": "model-00037-of-00086.safetensors", + "model.layers.53.mlp.gate_proj.weight_scale": "model-00037-of-00086.safetensors", + "model.layers.53.mlp.up_proj.input_scale": "model-00037-of-00086.safetensors", + "model.layers.53.mlp.up_proj.weight": "model-00037-of-00086.safetensors", + "model.layers.53.mlp.up_proj.weight_scale": "model-00037-of-00086.safetensors", + "model.layers.53.post_attention_layernorm.weight": "model-00037-of-00086.safetensors", + "model.layers.53.self_attn.k_proj.input_scale": "model-00037-of-00086.safetensors", + "model.layers.53.self_attn.k_proj.weight": "model-00037-of-00086.safetensors", + "model.layers.53.self_attn.k_proj.weight_scale": "model-00037-of-00086.safetensors", + "model.layers.53.self_attn.o_proj.input_scale": "model-00037-of-00086.safetensors", + "model.layers.53.self_attn.o_proj.weight": "model-00037-of-00086.safetensors", + "model.layers.53.self_attn.o_proj.weight_scale": "model-00037-of-00086.safetensors", + "model.layers.53.self_attn.q_proj.input_scale": "model-00037-of-00086.safetensors", + "model.layers.53.self_attn.q_proj.weight": "model-00037-of-00086.safetensors", + "model.layers.53.self_attn.q_proj.weight_scale": "model-00037-of-00086.safetensors", + "model.layers.53.self_attn.v_proj.input_scale": "model-00037-of-00086.safetensors", + "model.layers.53.self_attn.v_proj.weight": "model-00037-of-00086.safetensors", + "model.layers.53.self_attn.v_proj.weight_scale": "model-00037-of-00086.safetensors", + "model.layers.54.input_layernorm.weight": "model-00038-of-00086.safetensors", + "model.layers.54.mlp.down_proj.input_scale": "model-00038-of-00086.safetensors", + "model.layers.54.mlp.down_proj.weight": "model-00038-of-00086.safetensors", + "model.layers.54.mlp.down_proj.weight_scale": "model-00038-of-00086.safetensors", + "model.layers.54.mlp.gate_proj.input_scale": "model-00038-of-00086.safetensors", + "model.layers.54.mlp.gate_proj.weight": "model-00038-of-00086.safetensors", + "model.layers.54.mlp.gate_proj.weight_scale": "model-00038-of-00086.safetensors", + "model.layers.54.mlp.up_proj.input_scale": "model-00038-of-00086.safetensors", + "model.layers.54.mlp.up_proj.weight": "model-00038-of-00086.safetensors", + "model.layers.54.mlp.up_proj.weight_scale": "model-00038-of-00086.safetensors", + "model.layers.54.post_attention_layernorm.weight": "model-00038-of-00086.safetensors", + "model.layers.54.self_attn.k_proj.input_scale": "model-00037-of-00086.safetensors", + "model.layers.54.self_attn.k_proj.weight": "model-00037-of-00086.safetensors", + "model.layers.54.self_attn.k_proj.weight_scale": "model-00037-of-00086.safetensors", + "model.layers.54.self_attn.o_proj.input_scale": "model-00037-of-00086.safetensors", + "model.layers.54.self_attn.o_proj.weight": "model-00037-of-00086.safetensors", + "model.layers.54.self_attn.o_proj.weight_scale": "model-00037-of-00086.safetensors", + "model.layers.54.self_attn.q_proj.input_scale": "model-00037-of-00086.safetensors", + "model.layers.54.self_attn.q_proj.weight": "model-00037-of-00086.safetensors", + "model.layers.54.self_attn.q_proj.weight_scale": "model-00037-of-00086.safetensors", + "model.layers.54.self_attn.v_proj.input_scale": "model-00037-of-00086.safetensors", + "model.layers.54.self_attn.v_proj.weight": "model-00037-of-00086.safetensors", + "model.layers.54.self_attn.v_proj.weight_scale": "model-00037-of-00086.safetensors", + "model.layers.55.input_layernorm.weight": "model-00039-of-00086.safetensors", + "model.layers.55.mlp.down_proj.input_scale": "model-00039-of-00086.safetensors", + "model.layers.55.mlp.down_proj.weight": "model-00039-of-00086.safetensors", + "model.layers.55.mlp.down_proj.weight_scale": "model-00039-of-00086.safetensors", + "model.layers.55.mlp.gate_proj.input_scale": "model-00038-of-00086.safetensors", + "model.layers.55.mlp.gate_proj.weight": "model-00038-of-00086.safetensors", + "model.layers.55.mlp.gate_proj.weight_scale": "model-00038-of-00086.safetensors", + "model.layers.55.mlp.up_proj.input_scale": "model-00038-of-00086.safetensors", + "model.layers.55.mlp.up_proj.weight": "model-00038-of-00086.safetensors", + "model.layers.55.mlp.up_proj.weight_scale": "model-00038-of-00086.safetensors", + "model.layers.55.post_attention_layernorm.weight": "model-00039-of-00086.safetensors", + "model.layers.55.self_attn.k_proj.input_scale": "model-00038-of-00086.safetensors", + "model.layers.55.self_attn.k_proj.weight": "model-00038-of-00086.safetensors", + "model.layers.55.self_attn.k_proj.weight_scale": "model-00038-of-00086.safetensors", + "model.layers.55.self_attn.o_proj.input_scale": "model-00038-of-00086.safetensors", + "model.layers.55.self_attn.o_proj.weight": "model-00038-of-00086.safetensors", + "model.layers.55.self_attn.o_proj.weight_scale": "model-00038-of-00086.safetensors", + "model.layers.55.self_attn.q_proj.input_scale": "model-00038-of-00086.safetensors", + "model.layers.55.self_attn.q_proj.weight": "model-00038-of-00086.safetensors", + "model.layers.55.self_attn.q_proj.weight_scale": "model-00038-of-00086.safetensors", + "model.layers.55.self_attn.v_proj.input_scale": "model-00038-of-00086.safetensors", + "model.layers.55.self_attn.v_proj.weight": "model-00038-of-00086.safetensors", + "model.layers.55.self_attn.v_proj.weight_scale": "model-00038-of-00086.safetensors", + "model.layers.56.input_layernorm.weight": "model-00039-of-00086.safetensors", + "model.layers.56.mlp.down_proj.input_scale": "model-00039-of-00086.safetensors", + "model.layers.56.mlp.down_proj.weight": "model-00039-of-00086.safetensors", + "model.layers.56.mlp.down_proj.weight_scale": "model-00039-of-00086.safetensors", + "model.layers.56.mlp.gate_proj.input_scale": "model-00039-of-00086.safetensors", + "model.layers.56.mlp.gate_proj.weight": "model-00039-of-00086.safetensors", + "model.layers.56.mlp.gate_proj.weight_scale": "model-00039-of-00086.safetensors", + "model.layers.56.mlp.up_proj.input_scale": "model-00039-of-00086.safetensors", + "model.layers.56.mlp.up_proj.weight": "model-00039-of-00086.safetensors", + "model.layers.56.mlp.up_proj.weight_scale": "model-00039-of-00086.safetensors", + "model.layers.56.post_attention_layernorm.weight": "model-00039-of-00086.safetensors", + "model.layers.56.self_attn.k_proj.input_scale": "model-00039-of-00086.safetensors", + "model.layers.56.self_attn.k_proj.weight": "model-00039-of-00086.safetensors", + "model.layers.56.self_attn.k_proj.weight_scale": "model-00039-of-00086.safetensors", + "model.layers.56.self_attn.o_proj.input_scale": "model-00039-of-00086.safetensors", + "model.layers.56.self_attn.o_proj.weight": "model-00039-of-00086.safetensors", + "model.layers.56.self_attn.o_proj.weight_scale": "model-00039-of-00086.safetensors", + "model.layers.56.self_attn.q_proj.input_scale": "model-00039-of-00086.safetensors", + "model.layers.56.self_attn.q_proj.weight": "model-00039-of-00086.safetensors", + "model.layers.56.self_attn.q_proj.weight_scale": "model-00039-of-00086.safetensors", + "model.layers.56.self_attn.v_proj.input_scale": "model-00039-of-00086.safetensors", + "model.layers.56.self_attn.v_proj.weight": "model-00039-of-00086.safetensors", + "model.layers.56.self_attn.v_proj.weight_scale": "model-00039-of-00086.safetensors", + "model.layers.57.input_layernorm.weight": "model-00040-of-00086.safetensors", + "model.layers.57.mlp.down_proj.input_scale": "model-00040-of-00086.safetensors", + "model.layers.57.mlp.down_proj.weight": "model-00040-of-00086.safetensors", + "model.layers.57.mlp.down_proj.weight_scale": "model-00040-of-00086.safetensors", + "model.layers.57.mlp.gate_proj.input_scale": "model-00040-of-00086.safetensors", + "model.layers.57.mlp.gate_proj.weight": "model-00040-of-00086.safetensors", + "model.layers.57.mlp.gate_proj.weight_scale": "model-00040-of-00086.safetensors", + "model.layers.57.mlp.up_proj.input_scale": "model-00040-of-00086.safetensors", + "model.layers.57.mlp.up_proj.weight": "model-00040-of-00086.safetensors", + "model.layers.57.mlp.up_proj.weight_scale": "model-00040-of-00086.safetensors", + "model.layers.57.post_attention_layernorm.weight": "model-00040-of-00086.safetensors", + "model.layers.57.self_attn.k_proj.input_scale": "model-00039-of-00086.safetensors", + "model.layers.57.self_attn.k_proj.weight": "model-00039-of-00086.safetensors", + "model.layers.57.self_attn.k_proj.weight_scale": "model-00039-of-00086.safetensors", + "model.layers.57.self_attn.o_proj.input_scale": "model-00039-of-00086.safetensors", + "model.layers.57.self_attn.o_proj.weight": "model-00039-of-00086.safetensors", + "model.layers.57.self_attn.o_proj.weight_scale": "model-00039-of-00086.safetensors", + "model.layers.57.self_attn.q_proj.input_scale": "model-00039-of-00086.safetensors", + "model.layers.57.self_attn.q_proj.weight": "model-00039-of-00086.safetensors", + "model.layers.57.self_attn.q_proj.weight_scale": "model-00039-of-00086.safetensors", + "model.layers.57.self_attn.v_proj.input_scale": "model-00039-of-00086.safetensors", + "model.layers.57.self_attn.v_proj.weight": "model-00039-of-00086.safetensors", + "model.layers.57.self_attn.v_proj.weight_scale": "model-00039-of-00086.safetensors", + "model.layers.58.input_layernorm.weight": "model-00041-of-00086.safetensors", + "model.layers.58.mlp.down_proj.input_scale": "model-00041-of-00086.safetensors", + "model.layers.58.mlp.down_proj.weight": "model-00041-of-00086.safetensors", + "model.layers.58.mlp.down_proj.weight_scale": "model-00041-of-00086.safetensors", + "model.layers.58.mlp.gate_proj.input_scale": "model-00040-of-00086.safetensors", + "model.layers.58.mlp.gate_proj.weight": "model-00040-of-00086.safetensors", + "model.layers.58.mlp.gate_proj.weight_scale": "model-00040-of-00086.safetensors", + "model.layers.58.mlp.up_proj.input_scale": "model-00040-of-00086.safetensors", + "model.layers.58.mlp.up_proj.weight": "model-00040-of-00086.safetensors", + "model.layers.58.mlp.up_proj.weight_scale": "model-00040-of-00086.safetensors", + "model.layers.58.post_attention_layernorm.weight": "model-00041-of-00086.safetensors", + "model.layers.58.self_attn.k_proj.input_scale": "model-00040-of-00086.safetensors", + "model.layers.58.self_attn.k_proj.weight": "model-00040-of-00086.safetensors", + "model.layers.58.self_attn.k_proj.weight_scale": "model-00040-of-00086.safetensors", + "model.layers.58.self_attn.o_proj.input_scale": "model-00040-of-00086.safetensors", + "model.layers.58.self_attn.o_proj.weight": "model-00040-of-00086.safetensors", + "model.layers.58.self_attn.o_proj.weight_scale": "model-00040-of-00086.safetensors", + "model.layers.58.self_attn.q_proj.input_scale": "model-00040-of-00086.safetensors", + "model.layers.58.self_attn.q_proj.weight": "model-00040-of-00086.safetensors", + "model.layers.58.self_attn.q_proj.weight_scale": "model-00040-of-00086.safetensors", + "model.layers.58.self_attn.v_proj.input_scale": "model-00040-of-00086.safetensors", + "model.layers.58.self_attn.v_proj.weight": "model-00040-of-00086.safetensors", + "model.layers.58.self_attn.v_proj.weight_scale": "model-00040-of-00086.safetensors", + "model.layers.59.input_layernorm.weight": "model-00041-of-00086.safetensors", + "model.layers.59.mlp.down_proj.input_scale": "model-00041-of-00086.safetensors", + "model.layers.59.mlp.down_proj.weight": "model-00041-of-00086.safetensors", + "model.layers.59.mlp.down_proj.weight_scale": "model-00041-of-00086.safetensors", + "model.layers.59.mlp.gate_proj.input_scale": "model-00041-of-00086.safetensors", + "model.layers.59.mlp.gate_proj.weight": "model-00041-of-00086.safetensors", + "model.layers.59.mlp.gate_proj.weight_scale": "model-00041-of-00086.safetensors", + "model.layers.59.mlp.up_proj.input_scale": "model-00041-of-00086.safetensors", + "model.layers.59.mlp.up_proj.weight": "model-00041-of-00086.safetensors", + "model.layers.59.mlp.up_proj.weight_scale": "model-00041-of-00086.safetensors", + "model.layers.59.post_attention_layernorm.weight": "model-00041-of-00086.safetensors", + "model.layers.59.self_attn.k_proj.input_scale": "model-00041-of-00086.safetensors", + "model.layers.59.self_attn.k_proj.weight": "model-00041-of-00086.safetensors", + "model.layers.59.self_attn.k_proj.weight_scale": "model-00041-of-00086.safetensors", + "model.layers.59.self_attn.o_proj.input_scale": "model-00041-of-00086.safetensors", + "model.layers.59.self_attn.o_proj.weight": "model-00041-of-00086.safetensors", + "model.layers.59.self_attn.o_proj.weight_scale": "model-00041-of-00086.safetensors", + "model.layers.59.self_attn.q_proj.input_scale": "model-00041-of-00086.safetensors", + "model.layers.59.self_attn.q_proj.weight": "model-00041-of-00086.safetensors", + "model.layers.59.self_attn.q_proj.weight_scale": "model-00041-of-00086.safetensors", + "model.layers.59.self_attn.v_proj.input_scale": "model-00041-of-00086.safetensors", + "model.layers.59.self_attn.v_proj.weight": "model-00041-of-00086.safetensors", + "model.layers.59.self_attn.v_proj.weight_scale": "model-00041-of-00086.safetensors", + "model.layers.6.input_layernorm.weight": "model-00006-of-00086.safetensors", + "model.layers.6.mlp.down_proj.input_scale": "model-00006-of-00086.safetensors", + "model.layers.6.mlp.down_proj.weight": "model-00006-of-00086.safetensors", + "model.layers.6.mlp.down_proj.weight_scale": "model-00006-of-00086.safetensors", + "model.layers.6.mlp.gate_proj.input_scale": "model-00006-of-00086.safetensors", + "model.layers.6.mlp.gate_proj.weight": "model-00006-of-00086.safetensors", + "model.layers.6.mlp.gate_proj.weight_scale": "model-00006-of-00086.safetensors", + "model.layers.6.mlp.up_proj.input_scale": "model-00006-of-00086.safetensors", + "model.layers.6.mlp.up_proj.weight": "model-00006-of-00086.safetensors", + "model.layers.6.mlp.up_proj.weight_scale": "model-00006-of-00086.safetensors", + "model.layers.6.post_attention_layernorm.weight": "model-00006-of-00086.safetensors", + "model.layers.6.self_attn.k_proj.input_scale": "model-00005-of-00086.safetensors", + "model.layers.6.self_attn.k_proj.weight": "model-00005-of-00086.safetensors", + "model.layers.6.self_attn.k_proj.weight_scale": "model-00005-of-00086.safetensors", + "model.layers.6.self_attn.o_proj.input_scale": "model-00005-of-00086.safetensors", + "model.layers.6.self_attn.o_proj.weight": "model-00005-of-00086.safetensors", + "model.layers.6.self_attn.o_proj.weight_scale": "model-00005-of-00086.safetensors", + "model.layers.6.self_attn.q_proj.input_scale": "model-00005-of-00086.safetensors", + "model.layers.6.self_attn.q_proj.weight": "model-00005-of-00086.safetensors", + "model.layers.6.self_attn.q_proj.weight_scale": "model-00005-of-00086.safetensors", + "model.layers.6.self_attn.v_proj.input_scale": "model-00005-of-00086.safetensors", + "model.layers.6.self_attn.v_proj.weight": "model-00005-of-00086.safetensors", + "model.layers.6.self_attn.v_proj.weight_scale": "model-00005-of-00086.safetensors", + "model.layers.60.input_layernorm.weight": "model-00042-of-00086.safetensors", + "model.layers.60.mlp.down_proj.input_scale": "model-00042-of-00086.safetensors", + "model.layers.60.mlp.down_proj.weight": "model-00042-of-00086.safetensors", + "model.layers.60.mlp.down_proj.weight_scale": "model-00042-of-00086.safetensors", + "model.layers.60.mlp.gate_proj.input_scale": "model-00042-of-00086.safetensors", + "model.layers.60.mlp.gate_proj.weight": "model-00042-of-00086.safetensors", + "model.layers.60.mlp.gate_proj.weight_scale": "model-00042-of-00086.safetensors", + "model.layers.60.mlp.up_proj.input_scale": "model-00042-of-00086.safetensors", + "model.layers.60.mlp.up_proj.weight": "model-00042-of-00086.safetensors", + "model.layers.60.mlp.up_proj.weight_scale": "model-00042-of-00086.safetensors", + "model.layers.60.post_attention_layernorm.weight": "model-00042-of-00086.safetensors", + "model.layers.60.self_attn.k_proj.input_scale": "model-00041-of-00086.safetensors", + "model.layers.60.self_attn.k_proj.weight": "model-00041-of-00086.safetensors", + "model.layers.60.self_attn.k_proj.weight_scale": "model-00041-of-00086.safetensors", + "model.layers.60.self_attn.o_proj.input_scale": "model-00041-of-00086.safetensors", + "model.layers.60.self_attn.o_proj.weight": "model-00041-of-00086.safetensors", + "model.layers.60.self_attn.o_proj.weight_scale": "model-00041-of-00086.safetensors", + "model.layers.60.self_attn.q_proj.input_scale": "model-00041-of-00086.safetensors", + "model.layers.60.self_attn.q_proj.weight": "model-00041-of-00086.safetensors", + "model.layers.60.self_attn.q_proj.weight_scale": "model-00041-of-00086.safetensors", + "model.layers.60.self_attn.v_proj.input_scale": "model-00041-of-00086.safetensors", + "model.layers.60.self_attn.v_proj.weight": "model-00041-of-00086.safetensors", + "model.layers.60.self_attn.v_proj.weight_scale": "model-00041-of-00086.safetensors", + "model.layers.61.input_layernorm.weight": "model-00043-of-00086.safetensors", + "model.layers.61.mlp.down_proj.input_scale": "model-00043-of-00086.safetensors", + "model.layers.61.mlp.down_proj.weight": "model-00043-of-00086.safetensors", + "model.layers.61.mlp.down_proj.weight_scale": "model-00043-of-00086.safetensors", + "model.layers.61.mlp.gate_proj.input_scale": "model-00042-of-00086.safetensors", + "model.layers.61.mlp.gate_proj.weight": "model-00042-of-00086.safetensors", + "model.layers.61.mlp.gate_proj.weight_scale": "model-00042-of-00086.safetensors", + "model.layers.61.mlp.up_proj.input_scale": "model-00042-of-00086.safetensors", + "model.layers.61.mlp.up_proj.weight": "model-00042-of-00086.safetensors", + "model.layers.61.mlp.up_proj.weight_scale": "model-00042-of-00086.safetensors", + "model.layers.61.post_attention_layernorm.weight": "model-00043-of-00086.safetensors", + "model.layers.61.self_attn.k_proj.input_scale": "model-00042-of-00086.safetensors", + "model.layers.61.self_attn.k_proj.weight": "model-00042-of-00086.safetensors", + "model.layers.61.self_attn.k_proj.weight_scale": "model-00042-of-00086.safetensors", + "model.layers.61.self_attn.o_proj.input_scale": "model-00042-of-00086.safetensors", + "model.layers.61.self_attn.o_proj.weight": "model-00042-of-00086.safetensors", + "model.layers.61.self_attn.o_proj.weight_scale": "model-00042-of-00086.safetensors", + "model.layers.61.self_attn.q_proj.input_scale": "model-00042-of-00086.safetensors", + "model.layers.61.self_attn.q_proj.weight": "model-00042-of-00086.safetensors", + "model.layers.61.self_attn.q_proj.weight_scale": "model-00042-of-00086.safetensors", + "model.layers.61.self_attn.v_proj.input_scale": "model-00042-of-00086.safetensors", + "model.layers.61.self_attn.v_proj.weight": "model-00042-of-00086.safetensors", + "model.layers.61.self_attn.v_proj.weight_scale": "model-00042-of-00086.safetensors", + "model.layers.62.input_layernorm.weight": "model-00043-of-00086.safetensors", + "model.layers.62.mlp.down_proj.input_scale": "model-00043-of-00086.safetensors", + "model.layers.62.mlp.down_proj.weight": "model-00043-of-00086.safetensors", + "model.layers.62.mlp.down_proj.weight_scale": "model-00043-of-00086.safetensors", + "model.layers.62.mlp.gate_proj.input_scale": "model-00043-of-00086.safetensors", + "model.layers.62.mlp.gate_proj.weight": "model-00043-of-00086.safetensors", + "model.layers.62.mlp.gate_proj.weight_scale": "model-00043-of-00086.safetensors", + "model.layers.62.mlp.up_proj.input_scale": "model-00043-of-00086.safetensors", + "model.layers.62.mlp.up_proj.weight": "model-00043-of-00086.safetensors", + "model.layers.62.mlp.up_proj.weight_scale": "model-00043-of-00086.safetensors", + "model.layers.62.post_attention_layernorm.weight": "model-00043-of-00086.safetensors", + "model.layers.62.self_attn.k_proj.input_scale": "model-00043-of-00086.safetensors", + "model.layers.62.self_attn.k_proj.weight": "model-00043-of-00086.safetensors", + "model.layers.62.self_attn.k_proj.weight_scale": "model-00043-of-00086.safetensors", + "model.layers.62.self_attn.o_proj.input_scale": "model-00043-of-00086.safetensors", + "model.layers.62.self_attn.o_proj.weight": "model-00043-of-00086.safetensors", + "model.layers.62.self_attn.o_proj.weight_scale": "model-00043-of-00086.safetensors", + "model.layers.62.self_attn.q_proj.input_scale": "model-00043-of-00086.safetensors", + "model.layers.62.self_attn.q_proj.weight": "model-00043-of-00086.safetensors", + "model.layers.62.self_attn.q_proj.weight_scale": "model-00043-of-00086.safetensors", + "model.layers.62.self_attn.v_proj.input_scale": "model-00043-of-00086.safetensors", + "model.layers.62.self_attn.v_proj.weight": "model-00043-of-00086.safetensors", + "model.layers.62.self_attn.v_proj.weight_scale": "model-00043-of-00086.safetensors", + "model.layers.63.input_layernorm.weight": "model-00044-of-00086.safetensors", + "model.layers.63.mlp.down_proj.input_scale": "model-00044-of-00086.safetensors", + "model.layers.63.mlp.down_proj.weight": "model-00044-of-00086.safetensors", + "model.layers.63.mlp.down_proj.weight_scale": "model-00044-of-00086.safetensors", + "model.layers.63.mlp.gate_proj.input_scale": "model-00044-of-00086.safetensors", + "model.layers.63.mlp.gate_proj.weight": "model-00044-of-00086.safetensors", + "model.layers.63.mlp.gate_proj.weight_scale": "model-00044-of-00086.safetensors", + "model.layers.63.mlp.up_proj.input_scale": "model-00044-of-00086.safetensors", + "model.layers.63.mlp.up_proj.weight": "model-00044-of-00086.safetensors", + "model.layers.63.mlp.up_proj.weight_scale": "model-00044-of-00086.safetensors", + "model.layers.63.post_attention_layernorm.weight": "model-00044-of-00086.safetensors", + "model.layers.63.self_attn.k_proj.input_scale": "model-00043-of-00086.safetensors", + "model.layers.63.self_attn.k_proj.weight": "model-00043-of-00086.safetensors", + "model.layers.63.self_attn.k_proj.weight_scale": "model-00043-of-00086.safetensors", + "model.layers.63.self_attn.o_proj.input_scale": "model-00043-of-00086.safetensors", + "model.layers.63.self_attn.o_proj.weight": "model-00043-of-00086.safetensors", + "model.layers.63.self_attn.o_proj.weight_scale": "model-00043-of-00086.safetensors", + "model.layers.63.self_attn.q_proj.input_scale": "model-00043-of-00086.safetensors", + "model.layers.63.self_attn.q_proj.weight": "model-00043-of-00086.safetensors", + "model.layers.63.self_attn.q_proj.weight_scale": "model-00043-of-00086.safetensors", + "model.layers.63.self_attn.v_proj.input_scale": "model-00043-of-00086.safetensors", + "model.layers.63.self_attn.v_proj.weight": "model-00043-of-00086.safetensors", + "model.layers.63.self_attn.v_proj.weight_scale": "model-00043-of-00086.safetensors", + "model.layers.64.input_layernorm.weight": "model-00045-of-00086.safetensors", + "model.layers.64.mlp.down_proj.input_scale": "model-00045-of-00086.safetensors", + "model.layers.64.mlp.down_proj.weight": "model-00045-of-00086.safetensors", + "model.layers.64.mlp.down_proj.weight_scale": "model-00045-of-00086.safetensors", + "model.layers.64.mlp.gate_proj.input_scale": "model-00044-of-00086.safetensors", + "model.layers.64.mlp.gate_proj.weight": "model-00044-of-00086.safetensors", + "model.layers.64.mlp.gate_proj.weight_scale": "model-00044-of-00086.safetensors", + "model.layers.64.mlp.up_proj.input_scale": "model-00044-of-00086.safetensors", + "model.layers.64.mlp.up_proj.weight": "model-00044-of-00086.safetensors", + "model.layers.64.mlp.up_proj.weight_scale": "model-00044-of-00086.safetensors", + "model.layers.64.post_attention_layernorm.weight": "model-00045-of-00086.safetensors", + "model.layers.64.self_attn.k_proj.input_scale": "model-00044-of-00086.safetensors", + "model.layers.64.self_attn.k_proj.weight": "model-00044-of-00086.safetensors", + "model.layers.64.self_attn.k_proj.weight_scale": "model-00044-of-00086.safetensors", + "model.layers.64.self_attn.o_proj.input_scale": "model-00044-of-00086.safetensors", + "model.layers.64.self_attn.o_proj.weight": "model-00044-of-00086.safetensors", + "model.layers.64.self_attn.o_proj.weight_scale": "model-00044-of-00086.safetensors", + "model.layers.64.self_attn.q_proj.input_scale": "model-00044-of-00086.safetensors", + "model.layers.64.self_attn.q_proj.weight": "model-00044-of-00086.safetensors", + "model.layers.64.self_attn.q_proj.weight_scale": "model-00044-of-00086.safetensors", + "model.layers.64.self_attn.v_proj.input_scale": "model-00044-of-00086.safetensors", + "model.layers.64.self_attn.v_proj.weight": "model-00044-of-00086.safetensors", + "model.layers.64.self_attn.v_proj.weight_scale": "model-00044-of-00086.safetensors", + "model.layers.65.input_layernorm.weight": "model-00045-of-00086.safetensors", + "model.layers.65.mlp.down_proj.input_scale": "model-00045-of-00086.safetensors", + "model.layers.65.mlp.down_proj.weight": "model-00045-of-00086.safetensors", + "model.layers.65.mlp.down_proj.weight_scale": "model-00045-of-00086.safetensors", + "model.layers.65.mlp.gate_proj.input_scale": "model-00045-of-00086.safetensors", + "model.layers.65.mlp.gate_proj.weight": "model-00045-of-00086.safetensors", + "model.layers.65.mlp.gate_proj.weight_scale": "model-00045-of-00086.safetensors", + "model.layers.65.mlp.up_proj.input_scale": "model-00045-of-00086.safetensors", + "model.layers.65.mlp.up_proj.weight": "model-00045-of-00086.safetensors", + "model.layers.65.mlp.up_proj.weight_scale": "model-00045-of-00086.safetensors", + "model.layers.65.post_attention_layernorm.weight": "model-00045-of-00086.safetensors", + "model.layers.65.self_attn.k_proj.input_scale": "model-00045-of-00086.safetensors", + "model.layers.65.self_attn.k_proj.weight": "model-00045-of-00086.safetensors", + "model.layers.65.self_attn.k_proj.weight_scale": "model-00045-of-00086.safetensors", + "model.layers.65.self_attn.o_proj.input_scale": "model-00045-of-00086.safetensors", + "model.layers.65.self_attn.o_proj.weight": "model-00045-of-00086.safetensors", + "model.layers.65.self_attn.o_proj.weight_scale": "model-00045-of-00086.safetensors", + "model.layers.65.self_attn.q_proj.input_scale": "model-00045-of-00086.safetensors", + "model.layers.65.self_attn.q_proj.weight": "model-00045-of-00086.safetensors", + "model.layers.65.self_attn.q_proj.weight_scale": "model-00045-of-00086.safetensors", + "model.layers.65.self_attn.v_proj.input_scale": "model-00045-of-00086.safetensors", + "model.layers.65.self_attn.v_proj.weight": "model-00045-of-00086.safetensors", + "model.layers.65.self_attn.v_proj.weight_scale": "model-00045-of-00086.safetensors", + "model.layers.66.input_layernorm.weight": "model-00046-of-00086.safetensors", + "model.layers.66.mlp.down_proj.input_scale": "model-00046-of-00086.safetensors", + "model.layers.66.mlp.down_proj.weight": "model-00046-of-00086.safetensors", + "model.layers.66.mlp.down_proj.weight_scale": "model-00046-of-00086.safetensors", + "model.layers.66.mlp.gate_proj.input_scale": "model-00046-of-00086.safetensors", + "model.layers.66.mlp.gate_proj.weight": "model-00046-of-00086.safetensors", + "model.layers.66.mlp.gate_proj.weight_scale": "model-00046-of-00086.safetensors", + "model.layers.66.mlp.up_proj.input_scale": "model-00046-of-00086.safetensors", + "model.layers.66.mlp.up_proj.weight": "model-00046-of-00086.safetensors", + "model.layers.66.mlp.up_proj.weight_scale": "model-00046-of-00086.safetensors", + "model.layers.66.post_attention_layernorm.weight": "model-00046-of-00086.safetensors", + "model.layers.66.self_attn.k_proj.input_scale": "model-00045-of-00086.safetensors", + "model.layers.66.self_attn.k_proj.weight": "model-00045-of-00086.safetensors", + "model.layers.66.self_attn.k_proj.weight_scale": "model-00045-of-00086.safetensors", + "model.layers.66.self_attn.o_proj.input_scale": "model-00045-of-00086.safetensors", + "model.layers.66.self_attn.o_proj.weight": "model-00045-of-00086.safetensors", + "model.layers.66.self_attn.o_proj.weight_scale": "model-00045-of-00086.safetensors", + "model.layers.66.self_attn.q_proj.input_scale": "model-00045-of-00086.safetensors", + "model.layers.66.self_attn.q_proj.weight": "model-00045-of-00086.safetensors", + "model.layers.66.self_attn.q_proj.weight_scale": "model-00045-of-00086.safetensors", + "model.layers.66.self_attn.v_proj.input_scale": "model-00045-of-00086.safetensors", + "model.layers.66.self_attn.v_proj.weight": "model-00045-of-00086.safetensors", + "model.layers.66.self_attn.v_proj.weight_scale": "model-00045-of-00086.safetensors", + "model.layers.67.input_layernorm.weight": "model-00047-of-00086.safetensors", + "model.layers.67.mlp.down_proj.input_scale": "model-00047-of-00086.safetensors", + "model.layers.67.mlp.down_proj.weight": "model-00047-of-00086.safetensors", + "model.layers.67.mlp.down_proj.weight_scale": "model-00047-of-00086.safetensors", + "model.layers.67.mlp.gate_proj.input_scale": "model-00046-of-00086.safetensors", + "model.layers.67.mlp.gate_proj.weight": "model-00046-of-00086.safetensors", + "model.layers.67.mlp.gate_proj.weight_scale": "model-00046-of-00086.safetensors", + "model.layers.67.mlp.up_proj.input_scale": "model-00046-of-00086.safetensors", + "model.layers.67.mlp.up_proj.weight": "model-00046-of-00086.safetensors", + "model.layers.67.mlp.up_proj.weight_scale": "model-00046-of-00086.safetensors", + "model.layers.67.post_attention_layernorm.weight": "model-00047-of-00086.safetensors", + "model.layers.67.self_attn.k_proj.input_scale": "model-00046-of-00086.safetensors", + "model.layers.67.self_attn.k_proj.weight": "model-00046-of-00086.safetensors", + "model.layers.67.self_attn.k_proj.weight_scale": "model-00046-of-00086.safetensors", + "model.layers.67.self_attn.o_proj.input_scale": "model-00046-of-00086.safetensors", + "model.layers.67.self_attn.o_proj.weight": "model-00046-of-00086.safetensors", + "model.layers.67.self_attn.o_proj.weight_scale": "model-00046-of-00086.safetensors", + "model.layers.67.self_attn.q_proj.input_scale": "model-00046-of-00086.safetensors", + "model.layers.67.self_attn.q_proj.weight": "model-00046-of-00086.safetensors", + "model.layers.67.self_attn.q_proj.weight_scale": "model-00046-of-00086.safetensors", + "model.layers.67.self_attn.v_proj.input_scale": "model-00046-of-00086.safetensors", + "model.layers.67.self_attn.v_proj.weight": "model-00046-of-00086.safetensors", + "model.layers.67.self_attn.v_proj.weight_scale": "model-00046-of-00086.safetensors", + "model.layers.68.input_layernorm.weight": "model-00047-of-00086.safetensors", + "model.layers.68.mlp.down_proj.input_scale": "model-00047-of-00086.safetensors", + "model.layers.68.mlp.down_proj.weight": "model-00047-of-00086.safetensors", + "model.layers.68.mlp.down_proj.weight_scale": "model-00047-of-00086.safetensors", + "model.layers.68.mlp.gate_proj.input_scale": "model-00047-of-00086.safetensors", + "model.layers.68.mlp.gate_proj.weight": "model-00047-of-00086.safetensors", + "model.layers.68.mlp.gate_proj.weight_scale": "model-00047-of-00086.safetensors", + "model.layers.68.mlp.up_proj.input_scale": "model-00047-of-00086.safetensors", + "model.layers.68.mlp.up_proj.weight": "model-00047-of-00086.safetensors", + "model.layers.68.mlp.up_proj.weight_scale": "model-00047-of-00086.safetensors", + "model.layers.68.post_attention_layernorm.weight": "model-00047-of-00086.safetensors", + "model.layers.68.self_attn.k_proj.input_scale": "model-00047-of-00086.safetensors", + "model.layers.68.self_attn.k_proj.weight": "model-00047-of-00086.safetensors", + "model.layers.68.self_attn.k_proj.weight_scale": "model-00047-of-00086.safetensors", + "model.layers.68.self_attn.o_proj.input_scale": "model-00047-of-00086.safetensors", + "model.layers.68.self_attn.o_proj.weight": "model-00047-of-00086.safetensors", + "model.layers.68.self_attn.o_proj.weight_scale": "model-00047-of-00086.safetensors", + "model.layers.68.self_attn.q_proj.input_scale": "model-00047-of-00086.safetensors", + "model.layers.68.self_attn.q_proj.weight": "model-00047-of-00086.safetensors", + "model.layers.68.self_attn.q_proj.weight_scale": "model-00047-of-00086.safetensors", + "model.layers.68.self_attn.v_proj.input_scale": "model-00047-of-00086.safetensors", + "model.layers.68.self_attn.v_proj.weight": "model-00047-of-00086.safetensors", + "model.layers.68.self_attn.v_proj.weight_scale": "model-00047-of-00086.safetensors", + "model.layers.69.input_layernorm.weight": "model-00048-of-00086.safetensors", + "model.layers.69.mlp.down_proj.input_scale": "model-00048-of-00086.safetensors", + "model.layers.69.mlp.down_proj.weight": "model-00048-of-00086.safetensors", + "model.layers.69.mlp.down_proj.weight_scale": "model-00048-of-00086.safetensors", + "model.layers.69.mlp.gate_proj.input_scale": "model-00048-of-00086.safetensors", + "model.layers.69.mlp.gate_proj.weight": "model-00048-of-00086.safetensors", + "model.layers.69.mlp.gate_proj.weight_scale": "model-00048-of-00086.safetensors", + "model.layers.69.mlp.up_proj.input_scale": "model-00048-of-00086.safetensors", + "model.layers.69.mlp.up_proj.weight": "model-00048-of-00086.safetensors", + "model.layers.69.mlp.up_proj.weight_scale": "model-00048-of-00086.safetensors", + "model.layers.69.post_attention_layernorm.weight": "model-00048-of-00086.safetensors", + "model.layers.69.self_attn.k_proj.input_scale": "model-00047-of-00086.safetensors", + "model.layers.69.self_attn.k_proj.weight": "model-00047-of-00086.safetensors", + "model.layers.69.self_attn.k_proj.weight_scale": "model-00047-of-00086.safetensors", + "model.layers.69.self_attn.o_proj.input_scale": "model-00047-of-00086.safetensors", + "model.layers.69.self_attn.o_proj.weight": "model-00047-of-00086.safetensors", + "model.layers.69.self_attn.o_proj.weight_scale": "model-00047-of-00086.safetensors", + "model.layers.69.self_attn.q_proj.input_scale": "model-00047-of-00086.safetensors", + "model.layers.69.self_attn.q_proj.weight": "model-00047-of-00086.safetensors", + "model.layers.69.self_attn.q_proj.weight_scale": "model-00047-of-00086.safetensors", + "model.layers.69.self_attn.v_proj.input_scale": "model-00047-of-00086.safetensors", + "model.layers.69.self_attn.v_proj.weight": "model-00047-of-00086.safetensors", + "model.layers.69.self_attn.v_proj.weight_scale": "model-00047-of-00086.safetensors", + "model.layers.7.input_layernorm.weight": "model-00007-of-00086.safetensors", + "model.layers.7.mlp.down_proj.input_scale": "model-00007-of-00086.safetensors", + "model.layers.7.mlp.down_proj.weight": "model-00007-of-00086.safetensors", + "model.layers.7.mlp.down_proj.weight_scale": "model-00007-of-00086.safetensors", + "model.layers.7.mlp.gate_proj.input_scale": "model-00006-of-00086.safetensors", + "model.layers.7.mlp.gate_proj.weight": "model-00006-of-00086.safetensors", + "model.layers.7.mlp.gate_proj.weight_scale": "model-00006-of-00086.safetensors", + "model.layers.7.mlp.up_proj.input_scale": "model-00006-of-00086.safetensors", + "model.layers.7.mlp.up_proj.weight": "model-00006-of-00086.safetensors", + "model.layers.7.mlp.up_proj.weight_scale": "model-00006-of-00086.safetensors", + "model.layers.7.post_attention_layernorm.weight": "model-00007-of-00086.safetensors", + "model.layers.7.self_attn.k_proj.input_scale": "model-00006-of-00086.safetensors", + "model.layers.7.self_attn.k_proj.weight": "model-00006-of-00086.safetensors", + "model.layers.7.self_attn.k_proj.weight_scale": "model-00006-of-00086.safetensors", + "model.layers.7.self_attn.o_proj.input_scale": "model-00006-of-00086.safetensors", + "model.layers.7.self_attn.o_proj.weight": "model-00006-of-00086.safetensors", + "model.layers.7.self_attn.o_proj.weight_scale": "model-00006-of-00086.safetensors", + "model.layers.7.self_attn.q_proj.input_scale": "model-00006-of-00086.safetensors", + "model.layers.7.self_attn.q_proj.weight": "model-00006-of-00086.safetensors", + "model.layers.7.self_attn.q_proj.weight_scale": "model-00006-of-00086.safetensors", + "model.layers.7.self_attn.v_proj.input_scale": "model-00006-of-00086.safetensors", + "model.layers.7.self_attn.v_proj.weight": "model-00006-of-00086.safetensors", + "model.layers.7.self_attn.v_proj.weight_scale": "model-00006-of-00086.safetensors", + "model.layers.70.input_layernorm.weight": "model-00049-of-00086.safetensors", + "model.layers.70.mlp.down_proj.input_scale": "model-00049-of-00086.safetensors", + "model.layers.70.mlp.down_proj.weight": "model-00049-of-00086.safetensors", + "model.layers.70.mlp.down_proj.weight_scale": "model-00049-of-00086.safetensors", + "model.layers.70.mlp.gate_proj.input_scale": "model-00048-of-00086.safetensors", + "model.layers.70.mlp.gate_proj.weight": "model-00048-of-00086.safetensors", + "model.layers.70.mlp.gate_proj.weight_scale": "model-00048-of-00086.safetensors", + "model.layers.70.mlp.up_proj.input_scale": "model-00048-of-00086.safetensors", + "model.layers.70.mlp.up_proj.weight": "model-00048-of-00086.safetensors", + "model.layers.70.mlp.up_proj.weight_scale": "model-00048-of-00086.safetensors", + "model.layers.70.post_attention_layernorm.weight": "model-00049-of-00086.safetensors", + "model.layers.70.self_attn.k_proj.input_scale": "model-00048-of-00086.safetensors", + "model.layers.70.self_attn.k_proj.weight": "model-00048-of-00086.safetensors", + "model.layers.70.self_attn.k_proj.weight_scale": "model-00048-of-00086.safetensors", + "model.layers.70.self_attn.o_proj.input_scale": "model-00048-of-00086.safetensors", + "model.layers.70.self_attn.o_proj.weight": "model-00048-of-00086.safetensors", + "model.layers.70.self_attn.o_proj.weight_scale": "model-00048-of-00086.safetensors", + "model.layers.70.self_attn.q_proj.input_scale": "model-00048-of-00086.safetensors", + "model.layers.70.self_attn.q_proj.weight": "model-00048-of-00086.safetensors", + "model.layers.70.self_attn.q_proj.weight_scale": "model-00048-of-00086.safetensors", + "model.layers.70.self_attn.v_proj.input_scale": "model-00048-of-00086.safetensors", + "model.layers.70.self_attn.v_proj.weight": "model-00048-of-00086.safetensors", + "model.layers.70.self_attn.v_proj.weight_scale": "model-00048-of-00086.safetensors", + "model.layers.71.input_layernorm.weight": "model-00049-of-00086.safetensors", + "model.layers.71.mlp.down_proj.input_scale": "model-00049-of-00086.safetensors", + "model.layers.71.mlp.down_proj.weight": "model-00049-of-00086.safetensors", + "model.layers.71.mlp.down_proj.weight_scale": "model-00049-of-00086.safetensors", + "model.layers.71.mlp.gate_proj.input_scale": "model-00049-of-00086.safetensors", + "model.layers.71.mlp.gate_proj.weight": "model-00049-of-00086.safetensors", + "model.layers.71.mlp.gate_proj.weight_scale": "model-00049-of-00086.safetensors", + "model.layers.71.mlp.up_proj.input_scale": "model-00049-of-00086.safetensors", + "model.layers.71.mlp.up_proj.weight": "model-00049-of-00086.safetensors", + "model.layers.71.mlp.up_proj.weight_scale": "model-00049-of-00086.safetensors", + "model.layers.71.post_attention_layernorm.weight": "model-00049-of-00086.safetensors", + "model.layers.71.self_attn.k_proj.input_scale": "model-00049-of-00086.safetensors", + "model.layers.71.self_attn.k_proj.weight": "model-00049-of-00086.safetensors", + "model.layers.71.self_attn.k_proj.weight_scale": "model-00049-of-00086.safetensors", + "model.layers.71.self_attn.o_proj.input_scale": "model-00049-of-00086.safetensors", + "model.layers.71.self_attn.o_proj.weight": "model-00049-of-00086.safetensors", + "model.layers.71.self_attn.o_proj.weight_scale": "model-00049-of-00086.safetensors", + "model.layers.71.self_attn.q_proj.input_scale": "model-00049-of-00086.safetensors", + "model.layers.71.self_attn.q_proj.weight": "model-00049-of-00086.safetensors", + "model.layers.71.self_attn.q_proj.weight_scale": "model-00049-of-00086.safetensors", + "model.layers.71.self_attn.v_proj.input_scale": "model-00049-of-00086.safetensors", + "model.layers.71.self_attn.v_proj.weight": "model-00049-of-00086.safetensors", + "model.layers.71.self_attn.v_proj.weight_scale": "model-00049-of-00086.safetensors", + "model.layers.72.input_layernorm.weight": "model-00050-of-00086.safetensors", + "model.layers.72.mlp.down_proj.input_scale": "model-00050-of-00086.safetensors", + "model.layers.72.mlp.down_proj.weight": "model-00050-of-00086.safetensors", + "model.layers.72.mlp.down_proj.weight_scale": "model-00050-of-00086.safetensors", + "model.layers.72.mlp.gate_proj.input_scale": "model-00050-of-00086.safetensors", + "model.layers.72.mlp.gate_proj.weight": "model-00050-of-00086.safetensors", + "model.layers.72.mlp.gate_proj.weight_scale": "model-00050-of-00086.safetensors", + "model.layers.72.mlp.up_proj.input_scale": "model-00050-of-00086.safetensors", + "model.layers.72.mlp.up_proj.weight": "model-00050-of-00086.safetensors", + "model.layers.72.mlp.up_proj.weight_scale": "model-00050-of-00086.safetensors", + "model.layers.72.post_attention_layernorm.weight": "model-00050-of-00086.safetensors", + "model.layers.72.self_attn.k_proj.input_scale": "model-00049-of-00086.safetensors", + "model.layers.72.self_attn.k_proj.weight": "model-00049-of-00086.safetensors", + "model.layers.72.self_attn.k_proj.weight_scale": "model-00049-of-00086.safetensors", + "model.layers.72.self_attn.o_proj.input_scale": "model-00049-of-00086.safetensors", + "model.layers.72.self_attn.o_proj.weight": "model-00049-of-00086.safetensors", + "model.layers.72.self_attn.o_proj.weight_scale": "model-00049-of-00086.safetensors", + "model.layers.72.self_attn.q_proj.input_scale": "model-00049-of-00086.safetensors", + "model.layers.72.self_attn.q_proj.weight": "model-00049-of-00086.safetensors", + "model.layers.72.self_attn.q_proj.weight_scale": "model-00049-of-00086.safetensors", + "model.layers.72.self_attn.v_proj.input_scale": "model-00049-of-00086.safetensors", + "model.layers.72.self_attn.v_proj.weight": "model-00049-of-00086.safetensors", + "model.layers.72.self_attn.v_proj.weight_scale": "model-00049-of-00086.safetensors", + "model.layers.73.input_layernorm.weight": "model-00051-of-00086.safetensors", + "model.layers.73.mlp.down_proj.input_scale": "model-00051-of-00086.safetensors", + "model.layers.73.mlp.down_proj.weight": "model-00051-of-00086.safetensors", + "model.layers.73.mlp.down_proj.weight_scale": "model-00051-of-00086.safetensors", + "model.layers.73.mlp.gate_proj.input_scale": "model-00050-of-00086.safetensors", + "model.layers.73.mlp.gate_proj.weight": "model-00050-of-00086.safetensors", + "model.layers.73.mlp.gate_proj.weight_scale": "model-00050-of-00086.safetensors", + "model.layers.73.mlp.up_proj.input_scale": "model-00050-of-00086.safetensors", + "model.layers.73.mlp.up_proj.weight": "model-00050-of-00086.safetensors", + "model.layers.73.mlp.up_proj.weight_scale": "model-00050-of-00086.safetensors", + "model.layers.73.post_attention_layernorm.weight": "model-00051-of-00086.safetensors", + "model.layers.73.self_attn.k_proj.input_scale": "model-00050-of-00086.safetensors", + "model.layers.73.self_attn.k_proj.weight": "model-00050-of-00086.safetensors", + "model.layers.73.self_attn.k_proj.weight_scale": "model-00050-of-00086.safetensors", + "model.layers.73.self_attn.o_proj.input_scale": "model-00050-of-00086.safetensors", + "model.layers.73.self_attn.o_proj.weight": "model-00050-of-00086.safetensors", + "model.layers.73.self_attn.o_proj.weight_scale": "model-00050-of-00086.safetensors", + "model.layers.73.self_attn.q_proj.input_scale": "model-00050-of-00086.safetensors", + "model.layers.73.self_attn.q_proj.weight": "model-00050-of-00086.safetensors", + "model.layers.73.self_attn.q_proj.weight_scale": "model-00050-of-00086.safetensors", + "model.layers.73.self_attn.v_proj.input_scale": "model-00050-of-00086.safetensors", + "model.layers.73.self_attn.v_proj.weight": "model-00050-of-00086.safetensors", + "model.layers.73.self_attn.v_proj.weight_scale": "model-00050-of-00086.safetensors", + "model.layers.74.input_layernorm.weight": "model-00051-of-00086.safetensors", + "model.layers.74.mlp.down_proj.input_scale": "model-00051-of-00086.safetensors", + "model.layers.74.mlp.down_proj.weight": "model-00051-of-00086.safetensors", + "model.layers.74.mlp.down_proj.weight_scale": "model-00051-of-00086.safetensors", + "model.layers.74.mlp.gate_proj.input_scale": "model-00051-of-00086.safetensors", + "model.layers.74.mlp.gate_proj.weight": "model-00051-of-00086.safetensors", + "model.layers.74.mlp.gate_proj.weight_scale": "model-00051-of-00086.safetensors", + "model.layers.74.mlp.up_proj.input_scale": "model-00051-of-00086.safetensors", + "model.layers.74.mlp.up_proj.weight": "model-00051-of-00086.safetensors", + "model.layers.74.mlp.up_proj.weight_scale": "model-00051-of-00086.safetensors", + "model.layers.74.post_attention_layernorm.weight": "model-00051-of-00086.safetensors", + "model.layers.74.self_attn.k_proj.input_scale": "model-00051-of-00086.safetensors", + "model.layers.74.self_attn.k_proj.weight": "model-00051-of-00086.safetensors", + "model.layers.74.self_attn.k_proj.weight_scale": "model-00051-of-00086.safetensors", + "model.layers.74.self_attn.o_proj.input_scale": "model-00051-of-00086.safetensors", + "model.layers.74.self_attn.o_proj.weight": "model-00051-of-00086.safetensors", + "model.layers.74.self_attn.o_proj.weight_scale": "model-00051-of-00086.safetensors", + "model.layers.74.self_attn.q_proj.input_scale": "model-00051-of-00086.safetensors", + "model.layers.74.self_attn.q_proj.weight": "model-00051-of-00086.safetensors", + "model.layers.74.self_attn.q_proj.weight_scale": "model-00051-of-00086.safetensors", + "model.layers.74.self_attn.v_proj.input_scale": "model-00051-of-00086.safetensors", + "model.layers.74.self_attn.v_proj.weight": "model-00051-of-00086.safetensors", + "model.layers.74.self_attn.v_proj.weight_scale": "model-00051-of-00086.safetensors", + "model.layers.75.input_layernorm.weight": "model-00052-of-00086.safetensors", + "model.layers.75.mlp.down_proj.input_scale": "model-00052-of-00086.safetensors", + "model.layers.75.mlp.down_proj.weight": "model-00052-of-00086.safetensors", + "model.layers.75.mlp.down_proj.weight_scale": "model-00052-of-00086.safetensors", + "model.layers.75.mlp.gate_proj.input_scale": "model-00052-of-00086.safetensors", + "model.layers.75.mlp.gate_proj.weight": "model-00052-of-00086.safetensors", + "model.layers.75.mlp.gate_proj.weight_scale": "model-00052-of-00086.safetensors", + "model.layers.75.mlp.up_proj.input_scale": "model-00052-of-00086.safetensors", + "model.layers.75.mlp.up_proj.weight": "model-00052-of-00086.safetensors", + "model.layers.75.mlp.up_proj.weight_scale": "model-00052-of-00086.safetensors", + "model.layers.75.post_attention_layernorm.weight": "model-00052-of-00086.safetensors", + "model.layers.75.self_attn.k_proj.input_scale": "model-00051-of-00086.safetensors", + "model.layers.75.self_attn.k_proj.weight": "model-00051-of-00086.safetensors", + "model.layers.75.self_attn.k_proj.weight_scale": "model-00051-of-00086.safetensors", + "model.layers.75.self_attn.o_proj.input_scale": "model-00051-of-00086.safetensors", + "model.layers.75.self_attn.o_proj.weight": "model-00051-of-00086.safetensors", + "model.layers.75.self_attn.o_proj.weight_scale": "model-00051-of-00086.safetensors", + "model.layers.75.self_attn.q_proj.input_scale": "model-00051-of-00086.safetensors", + "model.layers.75.self_attn.q_proj.weight": "model-00051-of-00086.safetensors", + "model.layers.75.self_attn.q_proj.weight_scale": "model-00051-of-00086.safetensors", + "model.layers.75.self_attn.v_proj.input_scale": "model-00051-of-00086.safetensors", + "model.layers.75.self_attn.v_proj.weight": "model-00051-of-00086.safetensors", + "model.layers.75.self_attn.v_proj.weight_scale": "model-00051-of-00086.safetensors", + "model.layers.76.input_layernorm.weight": "model-00053-of-00086.safetensors", + "model.layers.76.mlp.down_proj.input_scale": "model-00053-of-00086.safetensors", + "model.layers.76.mlp.down_proj.weight": "model-00053-of-00086.safetensors", + "model.layers.76.mlp.down_proj.weight_scale": "model-00053-of-00086.safetensors", + "model.layers.76.mlp.gate_proj.input_scale": "model-00052-of-00086.safetensors", + "model.layers.76.mlp.gate_proj.weight": "model-00052-of-00086.safetensors", + "model.layers.76.mlp.gate_proj.weight_scale": "model-00052-of-00086.safetensors", + "model.layers.76.mlp.up_proj.input_scale": "model-00052-of-00086.safetensors", + "model.layers.76.mlp.up_proj.weight": "model-00052-of-00086.safetensors", + "model.layers.76.mlp.up_proj.weight_scale": "model-00052-of-00086.safetensors", + "model.layers.76.post_attention_layernorm.weight": "model-00053-of-00086.safetensors", + "model.layers.76.self_attn.k_proj.input_scale": "model-00052-of-00086.safetensors", + "model.layers.76.self_attn.k_proj.weight": "model-00052-of-00086.safetensors", + "model.layers.76.self_attn.k_proj.weight_scale": "model-00052-of-00086.safetensors", + "model.layers.76.self_attn.o_proj.input_scale": "model-00052-of-00086.safetensors", + "model.layers.76.self_attn.o_proj.weight": "model-00052-of-00086.safetensors", + "model.layers.76.self_attn.o_proj.weight_scale": "model-00052-of-00086.safetensors", + "model.layers.76.self_attn.q_proj.input_scale": "model-00052-of-00086.safetensors", + "model.layers.76.self_attn.q_proj.weight": "model-00052-of-00086.safetensors", + "model.layers.76.self_attn.q_proj.weight_scale": "model-00052-of-00086.safetensors", + "model.layers.76.self_attn.v_proj.input_scale": "model-00052-of-00086.safetensors", + "model.layers.76.self_attn.v_proj.weight": "model-00052-of-00086.safetensors", + "model.layers.76.self_attn.v_proj.weight_scale": "model-00052-of-00086.safetensors", + "model.layers.77.input_layernorm.weight": "model-00053-of-00086.safetensors", + "model.layers.77.mlp.down_proj.input_scale": "model-00053-of-00086.safetensors", + "model.layers.77.mlp.down_proj.weight": "model-00053-of-00086.safetensors", + "model.layers.77.mlp.down_proj.weight_scale": "model-00053-of-00086.safetensors", + "model.layers.77.mlp.gate_proj.input_scale": "model-00053-of-00086.safetensors", + "model.layers.77.mlp.gate_proj.weight": "model-00053-of-00086.safetensors", + "model.layers.77.mlp.gate_proj.weight_scale": "model-00053-of-00086.safetensors", + "model.layers.77.mlp.up_proj.input_scale": "model-00053-of-00086.safetensors", + "model.layers.77.mlp.up_proj.weight": "model-00053-of-00086.safetensors", + "model.layers.77.mlp.up_proj.weight_scale": "model-00053-of-00086.safetensors", + "model.layers.77.post_attention_layernorm.weight": "model-00053-of-00086.safetensors", + "model.layers.77.self_attn.k_proj.input_scale": "model-00053-of-00086.safetensors", + "model.layers.77.self_attn.k_proj.weight": "model-00053-of-00086.safetensors", + "model.layers.77.self_attn.k_proj.weight_scale": "model-00053-of-00086.safetensors", + "model.layers.77.self_attn.o_proj.input_scale": "model-00053-of-00086.safetensors", + "model.layers.77.self_attn.o_proj.weight": "model-00053-of-00086.safetensors", + "model.layers.77.self_attn.o_proj.weight_scale": "model-00053-of-00086.safetensors", + "model.layers.77.self_attn.q_proj.input_scale": "model-00053-of-00086.safetensors", + "model.layers.77.self_attn.q_proj.weight": "model-00053-of-00086.safetensors", + "model.layers.77.self_attn.q_proj.weight_scale": "model-00053-of-00086.safetensors", + "model.layers.77.self_attn.v_proj.input_scale": "model-00053-of-00086.safetensors", + "model.layers.77.self_attn.v_proj.weight": "model-00053-of-00086.safetensors", + "model.layers.77.self_attn.v_proj.weight_scale": "model-00053-of-00086.safetensors", + "model.layers.78.input_layernorm.weight": "model-00054-of-00086.safetensors", + "model.layers.78.mlp.down_proj.input_scale": "model-00054-of-00086.safetensors", + "model.layers.78.mlp.down_proj.weight": "model-00054-of-00086.safetensors", + "model.layers.78.mlp.down_proj.weight_scale": "model-00054-of-00086.safetensors", + "model.layers.78.mlp.gate_proj.input_scale": "model-00054-of-00086.safetensors", + "model.layers.78.mlp.gate_proj.weight": "model-00054-of-00086.safetensors", + "model.layers.78.mlp.gate_proj.weight_scale": "model-00054-of-00086.safetensors", + "model.layers.78.mlp.up_proj.input_scale": "model-00054-of-00086.safetensors", + "model.layers.78.mlp.up_proj.weight": "model-00054-of-00086.safetensors", + "model.layers.78.mlp.up_proj.weight_scale": "model-00054-of-00086.safetensors", + "model.layers.78.post_attention_layernorm.weight": "model-00054-of-00086.safetensors", + "model.layers.78.self_attn.k_proj.input_scale": "model-00053-of-00086.safetensors", + "model.layers.78.self_attn.k_proj.weight": "model-00053-of-00086.safetensors", + "model.layers.78.self_attn.k_proj.weight_scale": "model-00053-of-00086.safetensors", + "model.layers.78.self_attn.o_proj.input_scale": "model-00053-of-00086.safetensors", + "model.layers.78.self_attn.o_proj.weight": "model-00053-of-00086.safetensors", + "model.layers.78.self_attn.o_proj.weight_scale": "model-00053-of-00086.safetensors", + "model.layers.78.self_attn.q_proj.input_scale": "model-00053-of-00086.safetensors", + "model.layers.78.self_attn.q_proj.weight": "model-00053-of-00086.safetensors", + "model.layers.78.self_attn.q_proj.weight_scale": "model-00053-of-00086.safetensors", + "model.layers.78.self_attn.v_proj.input_scale": "model-00053-of-00086.safetensors", + "model.layers.78.self_attn.v_proj.weight": "model-00053-of-00086.safetensors", + "model.layers.78.self_attn.v_proj.weight_scale": "model-00053-of-00086.safetensors", + "model.layers.79.input_layernorm.weight": "model-00055-of-00086.safetensors", + "model.layers.79.mlp.down_proj.input_scale": "model-00055-of-00086.safetensors", + "model.layers.79.mlp.down_proj.weight": "model-00055-of-00086.safetensors", + "model.layers.79.mlp.down_proj.weight_scale": "model-00055-of-00086.safetensors", + "model.layers.79.mlp.gate_proj.input_scale": "model-00054-of-00086.safetensors", + "model.layers.79.mlp.gate_proj.weight": "model-00054-of-00086.safetensors", + "model.layers.79.mlp.gate_proj.weight_scale": "model-00054-of-00086.safetensors", + "model.layers.79.mlp.up_proj.input_scale": "model-00054-of-00086.safetensors", + "model.layers.79.mlp.up_proj.weight": "model-00054-of-00086.safetensors", + "model.layers.79.mlp.up_proj.weight_scale": "model-00054-of-00086.safetensors", + "model.layers.79.post_attention_layernorm.weight": "model-00055-of-00086.safetensors", + "model.layers.79.self_attn.k_proj.input_scale": "model-00054-of-00086.safetensors", + "model.layers.79.self_attn.k_proj.weight": "model-00054-of-00086.safetensors", + "model.layers.79.self_attn.k_proj.weight_scale": "model-00054-of-00086.safetensors", + "model.layers.79.self_attn.o_proj.input_scale": "model-00054-of-00086.safetensors", + "model.layers.79.self_attn.o_proj.weight": "model-00054-of-00086.safetensors", + "model.layers.79.self_attn.o_proj.weight_scale": "model-00054-of-00086.safetensors", + "model.layers.79.self_attn.q_proj.input_scale": "model-00054-of-00086.safetensors", + "model.layers.79.self_attn.q_proj.weight": "model-00054-of-00086.safetensors", + "model.layers.79.self_attn.q_proj.weight_scale": "model-00054-of-00086.safetensors", + "model.layers.79.self_attn.v_proj.input_scale": "model-00054-of-00086.safetensors", + "model.layers.79.self_attn.v_proj.weight": "model-00054-of-00086.safetensors", + "model.layers.79.self_attn.v_proj.weight_scale": "model-00054-of-00086.safetensors", + "model.layers.8.input_layernorm.weight": "model-00007-of-00086.safetensors", + "model.layers.8.mlp.down_proj.input_scale": "model-00007-of-00086.safetensors", + "model.layers.8.mlp.down_proj.weight": "model-00007-of-00086.safetensors", + "model.layers.8.mlp.down_proj.weight_scale": "model-00007-of-00086.safetensors", + "model.layers.8.mlp.gate_proj.input_scale": "model-00007-of-00086.safetensors", + "model.layers.8.mlp.gate_proj.weight": "model-00007-of-00086.safetensors", + "model.layers.8.mlp.gate_proj.weight_scale": "model-00007-of-00086.safetensors", + "model.layers.8.mlp.up_proj.input_scale": "model-00007-of-00086.safetensors", + "model.layers.8.mlp.up_proj.weight": "model-00007-of-00086.safetensors", + "model.layers.8.mlp.up_proj.weight_scale": "model-00007-of-00086.safetensors", + "model.layers.8.post_attention_layernorm.weight": "model-00007-of-00086.safetensors", + "model.layers.8.self_attn.k_proj.input_scale": "model-00007-of-00086.safetensors", + "model.layers.8.self_attn.k_proj.weight": "model-00007-of-00086.safetensors", + "model.layers.8.self_attn.k_proj.weight_scale": "model-00007-of-00086.safetensors", + "model.layers.8.self_attn.o_proj.input_scale": "model-00007-of-00086.safetensors", + "model.layers.8.self_attn.o_proj.weight": "model-00007-of-00086.safetensors", + "model.layers.8.self_attn.o_proj.weight_scale": "model-00007-of-00086.safetensors", + "model.layers.8.self_attn.q_proj.input_scale": "model-00007-of-00086.safetensors", + "model.layers.8.self_attn.q_proj.weight": "model-00007-of-00086.safetensors", + "model.layers.8.self_attn.q_proj.weight_scale": "model-00007-of-00086.safetensors", + "model.layers.8.self_attn.v_proj.input_scale": "model-00007-of-00086.safetensors", + "model.layers.8.self_attn.v_proj.weight": "model-00007-of-00086.safetensors", + "model.layers.8.self_attn.v_proj.weight_scale": "model-00007-of-00086.safetensors", + "model.layers.80.input_layernorm.weight": "model-00055-of-00086.safetensors", + "model.layers.80.mlp.down_proj.input_scale": "model-00055-of-00086.safetensors", + "model.layers.80.mlp.down_proj.weight": "model-00055-of-00086.safetensors", + "model.layers.80.mlp.down_proj.weight_scale": "model-00055-of-00086.safetensors", + "model.layers.80.mlp.gate_proj.input_scale": "model-00055-of-00086.safetensors", + "model.layers.80.mlp.gate_proj.weight": "model-00055-of-00086.safetensors", + "model.layers.80.mlp.gate_proj.weight_scale": "model-00055-of-00086.safetensors", + "model.layers.80.mlp.up_proj.input_scale": "model-00055-of-00086.safetensors", + "model.layers.80.mlp.up_proj.weight": "model-00055-of-00086.safetensors", + "model.layers.80.mlp.up_proj.weight_scale": "model-00055-of-00086.safetensors", + "model.layers.80.post_attention_layernorm.weight": "model-00055-of-00086.safetensors", + "model.layers.80.self_attn.k_proj.input_scale": "model-00055-of-00086.safetensors", + "model.layers.80.self_attn.k_proj.weight": "model-00055-of-00086.safetensors", + "model.layers.80.self_attn.k_proj.weight_scale": "model-00055-of-00086.safetensors", + "model.layers.80.self_attn.o_proj.input_scale": "model-00055-of-00086.safetensors", + "model.layers.80.self_attn.o_proj.weight": "model-00055-of-00086.safetensors", + "model.layers.80.self_attn.o_proj.weight_scale": "model-00055-of-00086.safetensors", + "model.layers.80.self_attn.q_proj.input_scale": "model-00055-of-00086.safetensors", + "model.layers.80.self_attn.q_proj.weight": "model-00055-of-00086.safetensors", + "model.layers.80.self_attn.q_proj.weight_scale": "model-00055-of-00086.safetensors", + "model.layers.80.self_attn.v_proj.input_scale": "model-00055-of-00086.safetensors", + "model.layers.80.self_attn.v_proj.weight": "model-00055-of-00086.safetensors", + "model.layers.80.self_attn.v_proj.weight_scale": "model-00055-of-00086.safetensors", + "model.layers.81.input_layernorm.weight": "model-00056-of-00086.safetensors", + "model.layers.81.mlp.down_proj.input_scale": "model-00056-of-00086.safetensors", + "model.layers.81.mlp.down_proj.weight": "model-00056-of-00086.safetensors", + "model.layers.81.mlp.down_proj.weight_scale": "model-00056-of-00086.safetensors", + "model.layers.81.mlp.gate_proj.input_scale": "model-00056-of-00086.safetensors", + "model.layers.81.mlp.gate_proj.weight": "model-00056-of-00086.safetensors", + "model.layers.81.mlp.gate_proj.weight_scale": "model-00056-of-00086.safetensors", + "model.layers.81.mlp.up_proj.input_scale": "model-00056-of-00086.safetensors", + "model.layers.81.mlp.up_proj.weight": "model-00056-of-00086.safetensors", + "model.layers.81.mlp.up_proj.weight_scale": "model-00056-of-00086.safetensors", + "model.layers.81.post_attention_layernorm.weight": "model-00056-of-00086.safetensors", + "model.layers.81.self_attn.k_proj.input_scale": "model-00055-of-00086.safetensors", + "model.layers.81.self_attn.k_proj.weight": "model-00055-of-00086.safetensors", + "model.layers.81.self_attn.k_proj.weight_scale": "model-00055-of-00086.safetensors", + "model.layers.81.self_attn.o_proj.input_scale": "model-00055-of-00086.safetensors", + "model.layers.81.self_attn.o_proj.weight": "model-00055-of-00086.safetensors", + "model.layers.81.self_attn.o_proj.weight_scale": "model-00055-of-00086.safetensors", + "model.layers.81.self_attn.q_proj.input_scale": "model-00055-of-00086.safetensors", + "model.layers.81.self_attn.q_proj.weight": "model-00055-of-00086.safetensors", + "model.layers.81.self_attn.q_proj.weight_scale": "model-00055-of-00086.safetensors", + "model.layers.81.self_attn.v_proj.input_scale": "model-00055-of-00086.safetensors", + "model.layers.81.self_attn.v_proj.weight": "model-00055-of-00086.safetensors", + "model.layers.81.self_attn.v_proj.weight_scale": "model-00055-of-00086.safetensors", + "model.layers.82.input_layernorm.weight": "model-00057-of-00086.safetensors", + "model.layers.82.mlp.down_proj.input_scale": "model-00057-of-00086.safetensors", + "model.layers.82.mlp.down_proj.weight": "model-00057-of-00086.safetensors", + "model.layers.82.mlp.down_proj.weight_scale": "model-00057-of-00086.safetensors", + "model.layers.82.mlp.gate_proj.input_scale": "model-00056-of-00086.safetensors", + "model.layers.82.mlp.gate_proj.weight": "model-00056-of-00086.safetensors", + "model.layers.82.mlp.gate_proj.weight_scale": "model-00056-of-00086.safetensors", + "model.layers.82.mlp.up_proj.input_scale": "model-00056-of-00086.safetensors", + "model.layers.82.mlp.up_proj.weight": "model-00056-of-00086.safetensors", + "model.layers.82.mlp.up_proj.weight_scale": "model-00056-of-00086.safetensors", + "model.layers.82.post_attention_layernorm.weight": "model-00057-of-00086.safetensors", + "model.layers.82.self_attn.k_proj.input_scale": "model-00056-of-00086.safetensors", + "model.layers.82.self_attn.k_proj.weight": "model-00056-of-00086.safetensors", + "model.layers.82.self_attn.k_proj.weight_scale": "model-00056-of-00086.safetensors", + "model.layers.82.self_attn.o_proj.input_scale": "model-00056-of-00086.safetensors", + "model.layers.82.self_attn.o_proj.weight": "model-00056-of-00086.safetensors", + "model.layers.82.self_attn.o_proj.weight_scale": "model-00056-of-00086.safetensors", + "model.layers.82.self_attn.q_proj.input_scale": "model-00056-of-00086.safetensors", + "model.layers.82.self_attn.q_proj.weight": "model-00056-of-00086.safetensors", + "model.layers.82.self_attn.q_proj.weight_scale": "model-00056-of-00086.safetensors", + "model.layers.82.self_attn.v_proj.input_scale": "model-00056-of-00086.safetensors", + "model.layers.82.self_attn.v_proj.weight": "model-00056-of-00086.safetensors", + "model.layers.82.self_attn.v_proj.weight_scale": "model-00056-of-00086.safetensors", + "model.layers.83.input_layernorm.weight": "model-00057-of-00086.safetensors", + "model.layers.83.mlp.down_proj.input_scale": "model-00057-of-00086.safetensors", + "model.layers.83.mlp.down_proj.weight": "model-00057-of-00086.safetensors", + "model.layers.83.mlp.down_proj.weight_scale": "model-00057-of-00086.safetensors", + "model.layers.83.mlp.gate_proj.input_scale": "model-00057-of-00086.safetensors", + "model.layers.83.mlp.gate_proj.weight": "model-00057-of-00086.safetensors", + "model.layers.83.mlp.gate_proj.weight_scale": "model-00057-of-00086.safetensors", + "model.layers.83.mlp.up_proj.input_scale": "model-00057-of-00086.safetensors", + "model.layers.83.mlp.up_proj.weight": "model-00057-of-00086.safetensors", + "model.layers.83.mlp.up_proj.weight_scale": "model-00057-of-00086.safetensors", + "model.layers.83.post_attention_layernorm.weight": "model-00057-of-00086.safetensors", + "model.layers.83.self_attn.k_proj.input_scale": "model-00057-of-00086.safetensors", + "model.layers.83.self_attn.k_proj.weight": "model-00057-of-00086.safetensors", + "model.layers.83.self_attn.k_proj.weight_scale": "model-00057-of-00086.safetensors", + "model.layers.83.self_attn.o_proj.input_scale": "model-00057-of-00086.safetensors", + "model.layers.83.self_attn.o_proj.weight": "model-00057-of-00086.safetensors", + "model.layers.83.self_attn.o_proj.weight_scale": "model-00057-of-00086.safetensors", + "model.layers.83.self_attn.q_proj.input_scale": "model-00057-of-00086.safetensors", + "model.layers.83.self_attn.q_proj.weight": "model-00057-of-00086.safetensors", + "model.layers.83.self_attn.q_proj.weight_scale": "model-00057-of-00086.safetensors", + "model.layers.83.self_attn.v_proj.input_scale": "model-00057-of-00086.safetensors", + "model.layers.83.self_attn.v_proj.weight": "model-00057-of-00086.safetensors", + "model.layers.83.self_attn.v_proj.weight_scale": "model-00057-of-00086.safetensors", + "model.layers.84.input_layernorm.weight": "model-00058-of-00086.safetensors", + "model.layers.84.mlp.down_proj.input_scale": "model-00058-of-00086.safetensors", + "model.layers.84.mlp.down_proj.weight": "model-00058-of-00086.safetensors", + "model.layers.84.mlp.down_proj.weight_scale": "model-00058-of-00086.safetensors", + "model.layers.84.mlp.gate_proj.input_scale": "model-00058-of-00086.safetensors", + "model.layers.84.mlp.gate_proj.weight": "model-00058-of-00086.safetensors", + "model.layers.84.mlp.gate_proj.weight_scale": "model-00058-of-00086.safetensors", + "model.layers.84.mlp.up_proj.input_scale": "model-00058-of-00086.safetensors", + "model.layers.84.mlp.up_proj.weight": "model-00058-of-00086.safetensors", + "model.layers.84.mlp.up_proj.weight_scale": "model-00058-of-00086.safetensors", + "model.layers.84.post_attention_layernorm.weight": "model-00058-of-00086.safetensors", + "model.layers.84.self_attn.k_proj.input_scale": "model-00057-of-00086.safetensors", + "model.layers.84.self_attn.k_proj.weight": "model-00057-of-00086.safetensors", + "model.layers.84.self_attn.k_proj.weight_scale": "model-00057-of-00086.safetensors", + "model.layers.84.self_attn.o_proj.input_scale": "model-00057-of-00086.safetensors", + "model.layers.84.self_attn.o_proj.weight": "model-00057-of-00086.safetensors", + "model.layers.84.self_attn.o_proj.weight_scale": "model-00057-of-00086.safetensors", + "model.layers.84.self_attn.q_proj.input_scale": "model-00057-of-00086.safetensors", + "model.layers.84.self_attn.q_proj.weight": "model-00057-of-00086.safetensors", + "model.layers.84.self_attn.q_proj.weight_scale": "model-00057-of-00086.safetensors", + "model.layers.84.self_attn.v_proj.input_scale": "model-00057-of-00086.safetensors", + "model.layers.84.self_attn.v_proj.weight": "model-00057-of-00086.safetensors", + "model.layers.84.self_attn.v_proj.weight_scale": "model-00057-of-00086.safetensors", + "model.layers.85.input_layernorm.weight": "model-00059-of-00086.safetensors", + "model.layers.85.mlp.down_proj.input_scale": "model-00059-of-00086.safetensors", + "model.layers.85.mlp.down_proj.weight": "model-00059-of-00086.safetensors", + "model.layers.85.mlp.down_proj.weight_scale": "model-00059-of-00086.safetensors", + "model.layers.85.mlp.gate_proj.input_scale": "model-00058-of-00086.safetensors", + "model.layers.85.mlp.gate_proj.weight": "model-00058-of-00086.safetensors", + "model.layers.85.mlp.gate_proj.weight_scale": "model-00058-of-00086.safetensors", + "model.layers.85.mlp.up_proj.input_scale": "model-00058-of-00086.safetensors", + "model.layers.85.mlp.up_proj.weight": "model-00058-of-00086.safetensors", + "model.layers.85.mlp.up_proj.weight_scale": "model-00058-of-00086.safetensors", + "model.layers.85.post_attention_layernorm.weight": "model-00059-of-00086.safetensors", + "model.layers.85.self_attn.k_proj.input_scale": "model-00058-of-00086.safetensors", + "model.layers.85.self_attn.k_proj.weight": "model-00058-of-00086.safetensors", + "model.layers.85.self_attn.k_proj.weight_scale": "model-00058-of-00086.safetensors", + "model.layers.85.self_attn.o_proj.input_scale": "model-00058-of-00086.safetensors", + "model.layers.85.self_attn.o_proj.weight": "model-00058-of-00086.safetensors", + "model.layers.85.self_attn.o_proj.weight_scale": "model-00058-of-00086.safetensors", + "model.layers.85.self_attn.q_proj.input_scale": "model-00058-of-00086.safetensors", + "model.layers.85.self_attn.q_proj.weight": "model-00058-of-00086.safetensors", + "model.layers.85.self_attn.q_proj.weight_scale": "model-00058-of-00086.safetensors", + "model.layers.85.self_attn.v_proj.input_scale": "model-00058-of-00086.safetensors", + "model.layers.85.self_attn.v_proj.weight": "model-00058-of-00086.safetensors", + "model.layers.85.self_attn.v_proj.weight_scale": "model-00058-of-00086.safetensors", + "model.layers.86.input_layernorm.weight": "model-00059-of-00086.safetensors", + "model.layers.86.mlp.down_proj.input_scale": "model-00059-of-00086.safetensors", + "model.layers.86.mlp.down_proj.weight": "model-00059-of-00086.safetensors", + "model.layers.86.mlp.down_proj.weight_scale": "model-00059-of-00086.safetensors", + "model.layers.86.mlp.gate_proj.input_scale": "model-00059-of-00086.safetensors", + "model.layers.86.mlp.gate_proj.weight": "model-00059-of-00086.safetensors", + "model.layers.86.mlp.gate_proj.weight_scale": "model-00059-of-00086.safetensors", + "model.layers.86.mlp.up_proj.input_scale": "model-00059-of-00086.safetensors", + "model.layers.86.mlp.up_proj.weight": "model-00059-of-00086.safetensors", + "model.layers.86.mlp.up_proj.weight_scale": "model-00059-of-00086.safetensors", + "model.layers.86.post_attention_layernorm.weight": "model-00059-of-00086.safetensors", + "model.layers.86.self_attn.k_proj.input_scale": "model-00059-of-00086.safetensors", + "model.layers.86.self_attn.k_proj.weight": "model-00059-of-00086.safetensors", + "model.layers.86.self_attn.k_proj.weight_scale": "model-00059-of-00086.safetensors", + "model.layers.86.self_attn.o_proj.input_scale": "model-00059-of-00086.safetensors", + "model.layers.86.self_attn.o_proj.weight": "model-00059-of-00086.safetensors", + "model.layers.86.self_attn.o_proj.weight_scale": "model-00059-of-00086.safetensors", + "model.layers.86.self_attn.q_proj.input_scale": "model-00059-of-00086.safetensors", + "model.layers.86.self_attn.q_proj.weight": "model-00059-of-00086.safetensors", + "model.layers.86.self_attn.q_proj.weight_scale": "model-00059-of-00086.safetensors", + "model.layers.86.self_attn.v_proj.input_scale": "model-00059-of-00086.safetensors", + "model.layers.86.self_attn.v_proj.weight": "model-00059-of-00086.safetensors", + "model.layers.86.self_attn.v_proj.weight_scale": "model-00059-of-00086.safetensors", + "model.layers.87.input_layernorm.weight": "model-00060-of-00086.safetensors", + "model.layers.87.mlp.down_proj.input_scale": "model-00060-of-00086.safetensors", + "model.layers.87.mlp.down_proj.weight": "model-00060-of-00086.safetensors", + "model.layers.87.mlp.down_proj.weight_scale": "model-00060-of-00086.safetensors", + "model.layers.87.mlp.gate_proj.input_scale": "model-00060-of-00086.safetensors", + "model.layers.87.mlp.gate_proj.weight": "model-00060-of-00086.safetensors", + "model.layers.87.mlp.gate_proj.weight_scale": "model-00060-of-00086.safetensors", + "model.layers.87.mlp.up_proj.input_scale": "model-00060-of-00086.safetensors", + "model.layers.87.mlp.up_proj.weight": "model-00060-of-00086.safetensors", + "model.layers.87.mlp.up_proj.weight_scale": "model-00060-of-00086.safetensors", + "model.layers.87.post_attention_layernorm.weight": "model-00060-of-00086.safetensors", + "model.layers.87.self_attn.k_proj.input_scale": "model-00059-of-00086.safetensors", + "model.layers.87.self_attn.k_proj.weight": "model-00059-of-00086.safetensors", + "model.layers.87.self_attn.k_proj.weight_scale": "model-00059-of-00086.safetensors", + "model.layers.87.self_attn.o_proj.input_scale": "model-00059-of-00086.safetensors", + "model.layers.87.self_attn.o_proj.weight": "model-00059-of-00086.safetensors", + "model.layers.87.self_attn.o_proj.weight_scale": "model-00059-of-00086.safetensors", + "model.layers.87.self_attn.q_proj.input_scale": "model-00059-of-00086.safetensors", + "model.layers.87.self_attn.q_proj.weight": "model-00059-of-00086.safetensors", + "model.layers.87.self_attn.q_proj.weight_scale": "model-00059-of-00086.safetensors", + "model.layers.87.self_attn.v_proj.input_scale": "model-00059-of-00086.safetensors", + "model.layers.87.self_attn.v_proj.weight": "model-00059-of-00086.safetensors", + "model.layers.87.self_attn.v_proj.weight_scale": "model-00059-of-00086.safetensors", + "model.layers.88.input_layernorm.weight": "model-00061-of-00086.safetensors", + "model.layers.88.mlp.down_proj.input_scale": "model-00061-of-00086.safetensors", + "model.layers.88.mlp.down_proj.weight": "model-00061-of-00086.safetensors", + "model.layers.88.mlp.down_proj.weight_scale": "model-00061-of-00086.safetensors", + "model.layers.88.mlp.gate_proj.input_scale": "model-00060-of-00086.safetensors", + "model.layers.88.mlp.gate_proj.weight": "model-00060-of-00086.safetensors", + "model.layers.88.mlp.gate_proj.weight_scale": "model-00060-of-00086.safetensors", + "model.layers.88.mlp.up_proj.input_scale": "model-00060-of-00086.safetensors", + "model.layers.88.mlp.up_proj.weight": "model-00060-of-00086.safetensors", + "model.layers.88.mlp.up_proj.weight_scale": "model-00060-of-00086.safetensors", + "model.layers.88.post_attention_layernorm.weight": "model-00061-of-00086.safetensors", + "model.layers.88.self_attn.k_proj.input_scale": "model-00060-of-00086.safetensors", + "model.layers.88.self_attn.k_proj.weight": "model-00060-of-00086.safetensors", + "model.layers.88.self_attn.k_proj.weight_scale": "model-00060-of-00086.safetensors", + "model.layers.88.self_attn.o_proj.input_scale": "model-00060-of-00086.safetensors", + "model.layers.88.self_attn.o_proj.weight": "model-00060-of-00086.safetensors", + "model.layers.88.self_attn.o_proj.weight_scale": "model-00060-of-00086.safetensors", + "model.layers.88.self_attn.q_proj.input_scale": "model-00060-of-00086.safetensors", + "model.layers.88.self_attn.q_proj.weight": "model-00060-of-00086.safetensors", + "model.layers.88.self_attn.q_proj.weight_scale": "model-00060-of-00086.safetensors", + "model.layers.88.self_attn.v_proj.input_scale": "model-00060-of-00086.safetensors", + "model.layers.88.self_attn.v_proj.weight": "model-00060-of-00086.safetensors", + "model.layers.88.self_attn.v_proj.weight_scale": "model-00060-of-00086.safetensors", + "model.layers.89.input_layernorm.weight": "model-00061-of-00086.safetensors", + "model.layers.89.mlp.down_proj.input_scale": "model-00061-of-00086.safetensors", + "model.layers.89.mlp.down_proj.weight": "model-00061-of-00086.safetensors", + "model.layers.89.mlp.down_proj.weight_scale": "model-00061-of-00086.safetensors", + "model.layers.89.mlp.gate_proj.input_scale": "model-00061-of-00086.safetensors", + "model.layers.89.mlp.gate_proj.weight": "model-00061-of-00086.safetensors", + "model.layers.89.mlp.gate_proj.weight_scale": "model-00061-of-00086.safetensors", + "model.layers.89.mlp.up_proj.input_scale": "model-00061-of-00086.safetensors", + "model.layers.89.mlp.up_proj.weight": "model-00061-of-00086.safetensors", + "model.layers.89.mlp.up_proj.weight_scale": "model-00061-of-00086.safetensors", + "model.layers.89.post_attention_layernorm.weight": "model-00061-of-00086.safetensors", + "model.layers.89.self_attn.k_proj.input_scale": "model-00061-of-00086.safetensors", + "model.layers.89.self_attn.k_proj.weight": "model-00061-of-00086.safetensors", + "model.layers.89.self_attn.k_proj.weight_scale": "model-00061-of-00086.safetensors", + "model.layers.89.self_attn.o_proj.input_scale": "model-00061-of-00086.safetensors", + "model.layers.89.self_attn.o_proj.weight": "model-00061-of-00086.safetensors", + "model.layers.89.self_attn.o_proj.weight_scale": "model-00061-of-00086.safetensors", + "model.layers.89.self_attn.q_proj.input_scale": "model-00061-of-00086.safetensors", + "model.layers.89.self_attn.q_proj.weight": "model-00061-of-00086.safetensors", + "model.layers.89.self_attn.q_proj.weight_scale": "model-00061-of-00086.safetensors", + "model.layers.89.self_attn.v_proj.input_scale": "model-00061-of-00086.safetensors", + "model.layers.89.self_attn.v_proj.weight": "model-00061-of-00086.safetensors", + "model.layers.89.self_attn.v_proj.weight_scale": "model-00061-of-00086.safetensors", + "model.layers.9.input_layernorm.weight": "model-00008-of-00086.safetensors", + "model.layers.9.mlp.down_proj.input_scale": "model-00008-of-00086.safetensors", + "model.layers.9.mlp.down_proj.weight": "model-00008-of-00086.safetensors", + "model.layers.9.mlp.down_proj.weight_scale": "model-00008-of-00086.safetensors", + "model.layers.9.mlp.gate_proj.input_scale": "model-00008-of-00086.safetensors", + "model.layers.9.mlp.gate_proj.weight": "model-00008-of-00086.safetensors", + "model.layers.9.mlp.gate_proj.weight_scale": "model-00008-of-00086.safetensors", + "model.layers.9.mlp.up_proj.input_scale": "model-00008-of-00086.safetensors", + "model.layers.9.mlp.up_proj.weight": "model-00008-of-00086.safetensors", + "model.layers.9.mlp.up_proj.weight_scale": "model-00008-of-00086.safetensors", + "model.layers.9.post_attention_layernorm.weight": "model-00008-of-00086.safetensors", + "model.layers.9.self_attn.k_proj.input_scale": "model-00007-of-00086.safetensors", + "model.layers.9.self_attn.k_proj.weight": "model-00007-of-00086.safetensors", + "model.layers.9.self_attn.k_proj.weight_scale": "model-00007-of-00086.safetensors", + "model.layers.9.self_attn.o_proj.input_scale": "model-00007-of-00086.safetensors", + "model.layers.9.self_attn.o_proj.weight": "model-00007-of-00086.safetensors", + "model.layers.9.self_attn.o_proj.weight_scale": "model-00007-of-00086.safetensors", + "model.layers.9.self_attn.q_proj.input_scale": "model-00007-of-00086.safetensors", + "model.layers.9.self_attn.q_proj.weight": "model-00007-of-00086.safetensors", + "model.layers.9.self_attn.q_proj.weight_scale": "model-00007-of-00086.safetensors", + "model.layers.9.self_attn.v_proj.input_scale": "model-00007-of-00086.safetensors", + "model.layers.9.self_attn.v_proj.weight": "model-00007-of-00086.safetensors", + "model.layers.9.self_attn.v_proj.weight_scale": "model-00007-of-00086.safetensors", + "model.layers.90.input_layernorm.weight": "model-00062-of-00086.safetensors", + "model.layers.90.mlp.down_proj.input_scale": "model-00062-of-00086.safetensors", + "model.layers.90.mlp.down_proj.weight": "model-00062-of-00086.safetensors", + "model.layers.90.mlp.down_proj.weight_scale": "model-00062-of-00086.safetensors", + "model.layers.90.mlp.gate_proj.input_scale": "model-00062-of-00086.safetensors", + "model.layers.90.mlp.gate_proj.weight": "model-00062-of-00086.safetensors", + "model.layers.90.mlp.gate_proj.weight_scale": "model-00062-of-00086.safetensors", + "model.layers.90.mlp.up_proj.input_scale": "model-00062-of-00086.safetensors", + "model.layers.90.mlp.up_proj.weight": "model-00062-of-00086.safetensors", + "model.layers.90.mlp.up_proj.weight_scale": "model-00062-of-00086.safetensors", + "model.layers.90.post_attention_layernorm.weight": "model-00062-of-00086.safetensors", + "model.layers.90.self_attn.k_proj.input_scale": "model-00061-of-00086.safetensors", + "model.layers.90.self_attn.k_proj.weight": "model-00061-of-00086.safetensors", + "model.layers.90.self_attn.k_proj.weight_scale": "model-00061-of-00086.safetensors", + "model.layers.90.self_attn.o_proj.input_scale": "model-00061-of-00086.safetensors", + "model.layers.90.self_attn.o_proj.weight": "model-00061-of-00086.safetensors", + "model.layers.90.self_attn.o_proj.weight_scale": "model-00061-of-00086.safetensors", + "model.layers.90.self_attn.q_proj.input_scale": "model-00061-of-00086.safetensors", + "model.layers.90.self_attn.q_proj.weight": "model-00061-of-00086.safetensors", + "model.layers.90.self_attn.q_proj.weight_scale": "model-00061-of-00086.safetensors", + "model.layers.90.self_attn.v_proj.input_scale": "model-00061-of-00086.safetensors", + "model.layers.90.self_attn.v_proj.weight": "model-00061-of-00086.safetensors", + "model.layers.90.self_attn.v_proj.weight_scale": "model-00061-of-00086.safetensors", + "model.layers.91.input_layernorm.weight": "model-00063-of-00086.safetensors", + "model.layers.91.mlp.down_proj.input_scale": "model-00063-of-00086.safetensors", + "model.layers.91.mlp.down_proj.weight": "model-00063-of-00086.safetensors", + "model.layers.91.mlp.down_proj.weight_scale": "model-00063-of-00086.safetensors", + "model.layers.91.mlp.gate_proj.input_scale": "model-00062-of-00086.safetensors", + "model.layers.91.mlp.gate_proj.weight": "model-00062-of-00086.safetensors", + "model.layers.91.mlp.gate_proj.weight_scale": "model-00062-of-00086.safetensors", + "model.layers.91.mlp.up_proj.input_scale": "model-00062-of-00086.safetensors", + "model.layers.91.mlp.up_proj.weight": "model-00062-of-00086.safetensors", + "model.layers.91.mlp.up_proj.weight_scale": "model-00062-of-00086.safetensors", + "model.layers.91.post_attention_layernorm.weight": "model-00063-of-00086.safetensors", + "model.layers.91.self_attn.k_proj.input_scale": "model-00062-of-00086.safetensors", + "model.layers.91.self_attn.k_proj.weight": "model-00062-of-00086.safetensors", + "model.layers.91.self_attn.k_proj.weight_scale": "model-00062-of-00086.safetensors", + "model.layers.91.self_attn.o_proj.input_scale": "model-00062-of-00086.safetensors", + "model.layers.91.self_attn.o_proj.weight": "model-00062-of-00086.safetensors", + "model.layers.91.self_attn.o_proj.weight_scale": "model-00062-of-00086.safetensors", + "model.layers.91.self_attn.q_proj.input_scale": "model-00062-of-00086.safetensors", + "model.layers.91.self_attn.q_proj.weight": "model-00062-of-00086.safetensors", + "model.layers.91.self_attn.q_proj.weight_scale": "model-00062-of-00086.safetensors", + "model.layers.91.self_attn.v_proj.input_scale": "model-00062-of-00086.safetensors", + "model.layers.91.self_attn.v_proj.weight": "model-00062-of-00086.safetensors", + "model.layers.91.self_attn.v_proj.weight_scale": "model-00062-of-00086.safetensors", + "model.layers.92.input_layernorm.weight": "model-00063-of-00086.safetensors", + "model.layers.92.mlp.down_proj.input_scale": "model-00063-of-00086.safetensors", + "model.layers.92.mlp.down_proj.weight": "model-00063-of-00086.safetensors", + "model.layers.92.mlp.down_proj.weight_scale": "model-00063-of-00086.safetensors", + "model.layers.92.mlp.gate_proj.input_scale": "model-00063-of-00086.safetensors", + "model.layers.92.mlp.gate_proj.weight": "model-00063-of-00086.safetensors", + "model.layers.92.mlp.gate_proj.weight_scale": "model-00063-of-00086.safetensors", + "model.layers.92.mlp.up_proj.input_scale": "model-00063-of-00086.safetensors", + "model.layers.92.mlp.up_proj.weight": "model-00063-of-00086.safetensors", + "model.layers.92.mlp.up_proj.weight_scale": "model-00063-of-00086.safetensors", + "model.layers.92.post_attention_layernorm.weight": "model-00063-of-00086.safetensors", + "model.layers.92.self_attn.k_proj.input_scale": "model-00063-of-00086.safetensors", + "model.layers.92.self_attn.k_proj.weight": "model-00063-of-00086.safetensors", + "model.layers.92.self_attn.k_proj.weight_scale": "model-00063-of-00086.safetensors", + "model.layers.92.self_attn.o_proj.input_scale": "model-00063-of-00086.safetensors", + "model.layers.92.self_attn.o_proj.weight": "model-00063-of-00086.safetensors", + "model.layers.92.self_attn.o_proj.weight_scale": "model-00063-of-00086.safetensors", + "model.layers.92.self_attn.q_proj.input_scale": "model-00063-of-00086.safetensors", + "model.layers.92.self_attn.q_proj.weight": "model-00063-of-00086.safetensors", + "model.layers.92.self_attn.q_proj.weight_scale": "model-00063-of-00086.safetensors", + "model.layers.92.self_attn.v_proj.input_scale": "model-00063-of-00086.safetensors", + "model.layers.92.self_attn.v_proj.weight": "model-00063-of-00086.safetensors", + "model.layers.92.self_attn.v_proj.weight_scale": "model-00063-of-00086.safetensors", + "model.layers.93.input_layernorm.weight": "model-00064-of-00086.safetensors", + "model.layers.93.mlp.down_proj.input_scale": "model-00064-of-00086.safetensors", + "model.layers.93.mlp.down_proj.weight": "model-00064-of-00086.safetensors", + "model.layers.93.mlp.down_proj.weight_scale": "model-00064-of-00086.safetensors", + "model.layers.93.mlp.gate_proj.input_scale": "model-00064-of-00086.safetensors", + "model.layers.93.mlp.gate_proj.weight": "model-00064-of-00086.safetensors", + "model.layers.93.mlp.gate_proj.weight_scale": "model-00064-of-00086.safetensors", + "model.layers.93.mlp.up_proj.input_scale": "model-00064-of-00086.safetensors", + "model.layers.93.mlp.up_proj.weight": "model-00064-of-00086.safetensors", + "model.layers.93.mlp.up_proj.weight_scale": "model-00064-of-00086.safetensors", + "model.layers.93.post_attention_layernorm.weight": "model-00064-of-00086.safetensors", + "model.layers.93.self_attn.k_proj.input_scale": "model-00063-of-00086.safetensors", + "model.layers.93.self_attn.k_proj.weight": "model-00063-of-00086.safetensors", + "model.layers.93.self_attn.k_proj.weight_scale": "model-00063-of-00086.safetensors", + "model.layers.93.self_attn.o_proj.input_scale": "model-00063-of-00086.safetensors", + "model.layers.93.self_attn.o_proj.weight": "model-00063-of-00086.safetensors", + "model.layers.93.self_attn.o_proj.weight_scale": "model-00063-of-00086.safetensors", + "model.layers.93.self_attn.q_proj.input_scale": "model-00063-of-00086.safetensors", + "model.layers.93.self_attn.q_proj.weight": "model-00063-of-00086.safetensors", + "model.layers.93.self_attn.q_proj.weight_scale": "model-00063-of-00086.safetensors", + "model.layers.93.self_attn.v_proj.input_scale": "model-00063-of-00086.safetensors", + "model.layers.93.self_attn.v_proj.weight": "model-00063-of-00086.safetensors", + "model.layers.93.self_attn.v_proj.weight_scale": "model-00063-of-00086.safetensors", + "model.layers.94.input_layernorm.weight": "model-00065-of-00086.safetensors", + "model.layers.94.mlp.down_proj.input_scale": "model-00065-of-00086.safetensors", + "model.layers.94.mlp.down_proj.weight": "model-00065-of-00086.safetensors", + "model.layers.94.mlp.down_proj.weight_scale": "model-00065-of-00086.safetensors", + "model.layers.94.mlp.gate_proj.input_scale": "model-00064-of-00086.safetensors", + "model.layers.94.mlp.gate_proj.weight": "model-00064-of-00086.safetensors", + "model.layers.94.mlp.gate_proj.weight_scale": "model-00064-of-00086.safetensors", + "model.layers.94.mlp.up_proj.input_scale": "model-00064-of-00086.safetensors", + "model.layers.94.mlp.up_proj.weight": "model-00064-of-00086.safetensors", + "model.layers.94.mlp.up_proj.weight_scale": "model-00064-of-00086.safetensors", + "model.layers.94.post_attention_layernorm.weight": "model-00065-of-00086.safetensors", + "model.layers.94.self_attn.k_proj.input_scale": "model-00064-of-00086.safetensors", + "model.layers.94.self_attn.k_proj.weight": "model-00064-of-00086.safetensors", + "model.layers.94.self_attn.k_proj.weight_scale": "model-00064-of-00086.safetensors", + "model.layers.94.self_attn.o_proj.input_scale": "model-00064-of-00086.safetensors", + "model.layers.94.self_attn.o_proj.weight": "model-00064-of-00086.safetensors", + "model.layers.94.self_attn.o_proj.weight_scale": "model-00064-of-00086.safetensors", + "model.layers.94.self_attn.q_proj.input_scale": "model-00064-of-00086.safetensors", + "model.layers.94.self_attn.q_proj.weight": "model-00064-of-00086.safetensors", + "model.layers.94.self_attn.q_proj.weight_scale": "model-00064-of-00086.safetensors", + "model.layers.94.self_attn.v_proj.input_scale": "model-00064-of-00086.safetensors", + "model.layers.94.self_attn.v_proj.weight": "model-00064-of-00086.safetensors", + "model.layers.94.self_attn.v_proj.weight_scale": "model-00064-of-00086.safetensors", + "model.layers.95.input_layernorm.weight": "model-00065-of-00086.safetensors", + "model.layers.95.mlp.down_proj.input_scale": "model-00065-of-00086.safetensors", + "model.layers.95.mlp.down_proj.weight": "model-00065-of-00086.safetensors", + "model.layers.95.mlp.down_proj.weight_scale": "model-00065-of-00086.safetensors", + "model.layers.95.mlp.gate_proj.input_scale": "model-00065-of-00086.safetensors", + "model.layers.95.mlp.gate_proj.weight": "model-00065-of-00086.safetensors", + "model.layers.95.mlp.gate_proj.weight_scale": "model-00065-of-00086.safetensors", + "model.layers.95.mlp.up_proj.input_scale": "model-00065-of-00086.safetensors", + "model.layers.95.mlp.up_proj.weight": "model-00065-of-00086.safetensors", + "model.layers.95.mlp.up_proj.weight_scale": "model-00065-of-00086.safetensors", + "model.layers.95.post_attention_layernorm.weight": "model-00065-of-00086.safetensors", + "model.layers.95.self_attn.k_proj.input_scale": "model-00065-of-00086.safetensors", + "model.layers.95.self_attn.k_proj.weight": "model-00065-of-00086.safetensors", + "model.layers.95.self_attn.k_proj.weight_scale": "model-00065-of-00086.safetensors", + "model.layers.95.self_attn.o_proj.input_scale": "model-00065-of-00086.safetensors", + "model.layers.95.self_attn.o_proj.weight": "model-00065-of-00086.safetensors", + "model.layers.95.self_attn.o_proj.weight_scale": "model-00065-of-00086.safetensors", + "model.layers.95.self_attn.q_proj.input_scale": "model-00065-of-00086.safetensors", + "model.layers.95.self_attn.q_proj.weight": "model-00065-of-00086.safetensors", + "model.layers.95.self_attn.q_proj.weight_scale": "model-00065-of-00086.safetensors", + "model.layers.95.self_attn.v_proj.input_scale": "model-00065-of-00086.safetensors", + "model.layers.95.self_attn.v_proj.weight": "model-00065-of-00086.safetensors", + "model.layers.95.self_attn.v_proj.weight_scale": "model-00065-of-00086.safetensors", + "model.layers.96.input_layernorm.weight": "model-00066-of-00086.safetensors", + "model.layers.96.mlp.down_proj.input_scale": "model-00066-of-00086.safetensors", + "model.layers.96.mlp.down_proj.weight": "model-00066-of-00086.safetensors", + "model.layers.96.mlp.down_proj.weight_scale": "model-00066-of-00086.safetensors", + "model.layers.96.mlp.gate_proj.input_scale": "model-00066-of-00086.safetensors", + "model.layers.96.mlp.gate_proj.weight": "model-00066-of-00086.safetensors", + "model.layers.96.mlp.gate_proj.weight_scale": "model-00066-of-00086.safetensors", + "model.layers.96.mlp.up_proj.input_scale": "model-00066-of-00086.safetensors", + "model.layers.96.mlp.up_proj.weight": "model-00066-of-00086.safetensors", + "model.layers.96.mlp.up_proj.weight_scale": "model-00066-of-00086.safetensors", + "model.layers.96.post_attention_layernorm.weight": "model-00066-of-00086.safetensors", + "model.layers.96.self_attn.k_proj.input_scale": "model-00065-of-00086.safetensors", + "model.layers.96.self_attn.k_proj.weight": "model-00065-of-00086.safetensors", + "model.layers.96.self_attn.k_proj.weight_scale": "model-00065-of-00086.safetensors", + "model.layers.96.self_attn.o_proj.input_scale": "model-00065-of-00086.safetensors", + "model.layers.96.self_attn.o_proj.weight": "model-00065-of-00086.safetensors", + "model.layers.96.self_attn.o_proj.weight_scale": "model-00065-of-00086.safetensors", + "model.layers.96.self_attn.q_proj.input_scale": "model-00065-of-00086.safetensors", + "model.layers.96.self_attn.q_proj.weight": "model-00065-of-00086.safetensors", + "model.layers.96.self_attn.q_proj.weight_scale": "model-00065-of-00086.safetensors", + "model.layers.96.self_attn.v_proj.input_scale": "model-00065-of-00086.safetensors", + "model.layers.96.self_attn.v_proj.weight": "model-00065-of-00086.safetensors", + "model.layers.96.self_attn.v_proj.weight_scale": "model-00065-of-00086.safetensors", + "model.layers.97.input_layernorm.weight": "model-00067-of-00086.safetensors", + "model.layers.97.mlp.down_proj.input_scale": "model-00067-of-00086.safetensors", + "model.layers.97.mlp.down_proj.weight": "model-00067-of-00086.safetensors", + "model.layers.97.mlp.down_proj.weight_scale": "model-00067-of-00086.safetensors", + "model.layers.97.mlp.gate_proj.input_scale": "model-00066-of-00086.safetensors", + "model.layers.97.mlp.gate_proj.weight": "model-00066-of-00086.safetensors", + "model.layers.97.mlp.gate_proj.weight_scale": "model-00066-of-00086.safetensors", + "model.layers.97.mlp.up_proj.input_scale": "model-00066-of-00086.safetensors", + "model.layers.97.mlp.up_proj.weight": "model-00066-of-00086.safetensors", + "model.layers.97.mlp.up_proj.weight_scale": "model-00066-of-00086.safetensors", + "model.layers.97.post_attention_layernorm.weight": "model-00067-of-00086.safetensors", + "model.layers.97.self_attn.k_proj.input_scale": "model-00066-of-00086.safetensors", + "model.layers.97.self_attn.k_proj.weight": "model-00066-of-00086.safetensors", + "model.layers.97.self_attn.k_proj.weight_scale": "model-00066-of-00086.safetensors", + "model.layers.97.self_attn.o_proj.input_scale": "model-00066-of-00086.safetensors", + "model.layers.97.self_attn.o_proj.weight": "model-00066-of-00086.safetensors", + "model.layers.97.self_attn.o_proj.weight_scale": "model-00066-of-00086.safetensors", + "model.layers.97.self_attn.q_proj.input_scale": "model-00066-of-00086.safetensors", + "model.layers.97.self_attn.q_proj.weight": "model-00066-of-00086.safetensors", + "model.layers.97.self_attn.q_proj.weight_scale": "model-00066-of-00086.safetensors", + "model.layers.97.self_attn.v_proj.input_scale": "model-00066-of-00086.safetensors", + "model.layers.97.self_attn.v_proj.weight": "model-00066-of-00086.safetensors", + "model.layers.97.self_attn.v_proj.weight_scale": "model-00066-of-00086.safetensors", + "model.layers.98.input_layernorm.weight": "model-00067-of-00086.safetensors", + "model.layers.98.mlp.down_proj.input_scale": "model-00067-of-00086.safetensors", + "model.layers.98.mlp.down_proj.weight": "model-00067-of-00086.safetensors", + "model.layers.98.mlp.down_proj.weight_scale": "model-00067-of-00086.safetensors", + "model.layers.98.mlp.gate_proj.input_scale": "model-00067-of-00086.safetensors", + "model.layers.98.mlp.gate_proj.weight": "model-00067-of-00086.safetensors", + "model.layers.98.mlp.gate_proj.weight_scale": "model-00067-of-00086.safetensors", + "model.layers.98.mlp.up_proj.input_scale": "model-00067-of-00086.safetensors", + "model.layers.98.mlp.up_proj.weight": "model-00067-of-00086.safetensors", + "model.layers.98.mlp.up_proj.weight_scale": "model-00067-of-00086.safetensors", + "model.layers.98.post_attention_layernorm.weight": "model-00067-of-00086.safetensors", + "model.layers.98.self_attn.k_proj.input_scale": "model-00067-of-00086.safetensors", + "model.layers.98.self_attn.k_proj.weight": "model-00067-of-00086.safetensors", + "model.layers.98.self_attn.k_proj.weight_scale": "model-00067-of-00086.safetensors", + "model.layers.98.self_attn.o_proj.input_scale": "model-00067-of-00086.safetensors", + "model.layers.98.self_attn.o_proj.weight": "model-00067-of-00086.safetensors", + "model.layers.98.self_attn.o_proj.weight_scale": "model-00067-of-00086.safetensors", + "model.layers.98.self_attn.q_proj.input_scale": "model-00067-of-00086.safetensors", + "model.layers.98.self_attn.q_proj.weight": "model-00067-of-00086.safetensors", + "model.layers.98.self_attn.q_proj.weight_scale": "model-00067-of-00086.safetensors", + "model.layers.98.self_attn.v_proj.input_scale": "model-00067-of-00086.safetensors", + "model.layers.98.self_attn.v_proj.weight": "model-00067-of-00086.safetensors", + "model.layers.98.self_attn.v_proj.weight_scale": "model-00067-of-00086.safetensors", + "model.layers.99.input_layernorm.weight": "model-00068-of-00086.safetensors", + "model.layers.99.mlp.down_proj.input_scale": "model-00068-of-00086.safetensors", + "model.layers.99.mlp.down_proj.weight": "model-00068-of-00086.safetensors", + "model.layers.99.mlp.down_proj.weight_scale": "model-00068-of-00086.safetensors", + "model.layers.99.mlp.gate_proj.input_scale": "model-00068-of-00086.safetensors", + "model.layers.99.mlp.gate_proj.weight": "model-00068-of-00086.safetensors", + "model.layers.99.mlp.gate_proj.weight_scale": "model-00068-of-00086.safetensors", + "model.layers.99.mlp.up_proj.input_scale": "model-00068-of-00086.safetensors", + "model.layers.99.mlp.up_proj.weight": "model-00068-of-00086.safetensors", + "model.layers.99.mlp.up_proj.weight_scale": "model-00068-of-00086.safetensors", + "model.layers.99.post_attention_layernorm.weight": "model-00068-of-00086.safetensors", + "model.layers.99.self_attn.k_proj.input_scale": "model-00067-of-00086.safetensors", + "model.layers.99.self_attn.k_proj.weight": "model-00067-of-00086.safetensors", + "model.layers.99.self_attn.k_proj.weight_scale": "model-00067-of-00086.safetensors", + "model.layers.99.self_attn.o_proj.input_scale": "model-00067-of-00086.safetensors", + "model.layers.99.self_attn.o_proj.weight": "model-00067-of-00086.safetensors", + "model.layers.99.self_attn.o_proj.weight_scale": "model-00067-of-00086.safetensors", + "model.layers.99.self_attn.q_proj.input_scale": "model-00067-of-00086.safetensors", + "model.layers.99.self_attn.q_proj.weight": "model-00067-of-00086.safetensors", + "model.layers.99.self_attn.q_proj.weight_scale": "model-00067-of-00086.safetensors", + "model.layers.99.self_attn.v_proj.input_scale": "model-00067-of-00086.safetensors", + "model.layers.99.self_attn.v_proj.weight": "model-00067-of-00086.safetensors", + "model.layers.99.self_attn.v_proj.weight_scale": "model-00067-of-00086.safetensors", + "model.norm.weight": "model-00085-of-00086.safetensors" + } +} diff --git a/special_tokens_map.json b/special_tokens_map.json new file mode 100644 index 0000000000000000000000000000000000000000..04829afa78a2d2df203ac846968db37269b01f7f --- /dev/null +++ b/special_tokens_map.json @@ -0,0 +1,17 @@ +{ + "bos_token": { + "content": "<|begin_of_text|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false + }, + "eos_token": { + "content": "<|end_of_text|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false + }, + "pad_token": "<|end_of_text|>" +} diff --git a/tokenizer.json b/tokenizer.json new file mode 100644 index 0000000000000000000000000000000000000000..2bac1a86a87dd66d9e953215a4c76f7151c65994 --- /dev/null +++ b/tokenizer.json @@ -0,0 +1,410575 @@ +{ + "version": "1.0", + "truncation": { + "direction": "Right", + "max_length": 512, + "strategy": "LongestFirst", + "stride": 0 + }, + "padding": { + "strategy": "BatchLongest", + "direction": "Left", + "pad_to_multiple_of": null, + "pad_id": 128001, + "pad_type_id": 0, + "pad_token": "<|end_of_text|>" + }, + "added_tokens": [ + { + "id": 128000, + "content": "<|begin_of_text|>", + "single_word": false, + "lstrip": false, + "rstrip": false, + "normalized": false, + "special": true + }, + { + "id": 128001, + "content": "<|end_of_text|>", + "single_word": false, + "lstrip": false, + "rstrip": false, + "normalized": false, + "special": true + }, + { + "id": 128002, + "content": "<|reserved_special_token_0|>", + "single_word": false, + "lstrip": false, + "rstrip": false, + "normalized": false, + "special": true + }, + { + "id": 128003, + "content": "<|reserved_special_token_1|>", + "single_word": false, + "lstrip": false, + "rstrip": false, + "normalized": false, + "special": true + }, + { + "id": 128004, + "content": "<|finetune_right_pad_id|>", + "single_word": false, + "lstrip": false, + "rstrip": false, + "normalized": false, + "special": true + }, + { + "id": 128005, + "content": "<|reserved_special_token_2|>", + "single_word": false, + "lstrip": false, + "rstrip": false, + "normalized": false, + "special": true + }, + { + "id": 128006, + "content": "<|start_header_id|>", + "single_word": false, + "lstrip": false, + "rstrip": false, + "normalized": false, + "special": true + }, + { + "id": 128007, + "content": "<|end_header_id|>", + "single_word": false, + "lstrip": false, + "rstrip": false, + "normalized": false, + "special": true + }, + { + "id": 128008, + "content": "<|eom_id|>", + "single_word": false, + "lstrip": false, + "rstrip": false, + "normalized": false, + "special": true + }, + { + "id": 128009, + "content": "<|eot_id|>", + "single_word": false, + "lstrip": false, + "rstrip": false, + "normalized": false, + "special": true + }, + { + "id": 128010, + "content": "<|python_tag|>", + "single_word": false, + "lstrip": false, + "rstrip": false, + "normalized": false, + "special": true + }, + { + "id": 128011, + "content": "<|reserved_special_token_3|>", + "single_word": false, + "lstrip": false, + "rstrip": false, + "normalized": false, + "special": true + }, + { + "id": 128012, + "content": "<|reserved_special_token_4|>", + "single_word": false, + "lstrip": false, + "rstrip": false, + "normalized": false, + "special": true + }, + { + "id": 128013, + "content": "<|reserved_special_token_5|>", + "single_word": false, + "lstrip": false, + "rstrip": false, + "normalized": false, + "special": true + }, + { + "id": 128014, + "content": "<|reserved_special_token_6|>", + "single_word": false, + "lstrip": false, + "rstrip": false, + "normalized": false, + "special": true + }, + { + "id": 128015, + "content": "<|reserved_special_token_7|>", + "single_word": false, + "lstrip": false, + "rstrip": false, + "normalized": false, + "special": true + }, + { + "id": 128016, + "content": "<|reserved_special_token_8|>", + "single_word": false, + "lstrip": false, + "rstrip": false, + "normalized": false, + "special": true + }, + { + "id": 128017, + "content": "<|reserved_special_token_9|>", + "single_word": false, + "lstrip": false, + "rstrip": false, + "normalized": false, + "special": true + }, + { + "id": 128018, + "content": "<|reserved_special_token_10|>", + "single_word": false, + "lstrip": false, + "rstrip": false, + "normalized": false, + "special": true + }, + { + "id": 128019, + "content": "<|reserved_special_token_11|>", + "single_word": false, + "lstrip": false, + "rstrip": false, + "normalized": false, + "special": true + }, + { + "id": 128020, + "content": "<|reserved_special_token_12|>", + "single_word": false, + "lstrip": false, + "rstrip": false, + "normalized": false, + "special": true + }, + { + "id": 128021, + "content": "<|reserved_special_token_13|>", + "single_word": false, + "lstrip": false, + "rstrip": false, + "normalized": false, + "special": true + }, + { + "id": 128022, + "content": "<|reserved_special_token_14|>", + "single_word": false, + "lstrip": false, + "rstrip": false, + "normalized": false, + "special": true + }, + { + "id": 128023, + "content": "<|reserved_special_token_15|>", + "single_word": false, + "lstrip": false, + "rstrip": false, + "normalized": false, + "special": true + }, + { + "id": 128024, + "content": "<|reserved_special_token_16|>", + "single_word": false, + "lstrip": false, + "rstrip": false, + "normalized": false, + "special": true + }, + { + "id": 128025, + "content": "<|reserved_special_token_17|>", + "single_word": false, + "lstrip": false, + "rstrip": false, + "normalized": false, + "special": true + }, + { + "id": 128026, + "content": "<|reserved_special_token_18|>", + "single_word": false, + "lstrip": false, + "rstrip": false, + "normalized": false, + "special": true + }, + { + "id": 128027, + "content": "<|reserved_special_token_19|>", + "single_word": false, + "lstrip": false, + "rstrip": false, + "normalized": false, + "special": true + }, + { + "id": 128028, + "content": "<|reserved_special_token_20|>", + "single_word": false, + "lstrip": false, + "rstrip": false, + "normalized": false, + "special": true + }, + { + "id": 128029, + "content": "<|reserved_special_token_21|>", + "single_word": false, + "lstrip": false, + "rstrip": false, + "normalized": false, + "special": true + }, + { + "id": 128030, + "content": "<|reserved_special_token_22|>", + "single_word": false, + "lstrip": false, + "rstrip": false, + "normalized": false, + "special": true + }, + { + "id": 128031, + "content": "<|reserved_special_token_23|>", + "single_word": false, + "lstrip": false, + "rstrip": false, + "normalized": false, + "special": true + }, + { + "id": 128032, + "content": "<|reserved_special_token_24|>", + "single_word": false, + "lstrip": false, + "rstrip": false, + "normalized": false, + "special": true + }, + { + "id": 128033, + "content": "<|reserved_special_token_25|>", + "single_word": false, + "lstrip": false, + "rstrip": false, + "normalized": false, + "special": true + }, + { + "id": 128034, + "content": "<|reserved_special_token_26|>", + "single_word": false, + "lstrip": false, + "rstrip": false, + "normalized": false, + "special": true + }, + { + "id": 128035, + "content": "<|reserved_special_token_27|>", + "single_word": false, + "lstrip": false, + "rstrip": false, + "normalized": false, + "special": true + }, + { + "id": 128036, + "content": "<|reserved_special_token_28|>", + "single_word": false, + "lstrip": false, + "rstrip": false, + "normalized": false, + "special": true + }, + { + "id": 128037, + "content": "<|reserved_special_token_29|>", + "single_word": false, + "lstrip": false, + "rstrip": false, + "normalized": false, + "special": true + }, + { + "id": 128038, + "content": "<|reserved_special_token_30|>", + "single_word": false, + "lstrip": false, + "rstrip": false, + "normalized": false, + "special": true + }, + { + "id": 128039, + "content": "<|reserved_special_token_31|>", + "single_word": false, + "lstrip": false, + "rstrip": false, + "normalized": false, + "special": true + }, + { + "id": 128040, + "content": "<|reserved_special_token_32|>", + "single_word": false, + "lstrip": false, + "rstrip": false, + "normalized": false, + "special": true + }, + { + "id": 128041, + "content": "<|reserved_special_token_33|>", + "single_word": false, + "lstrip": false, + "rstrip": false, + "normalized": false, + "special": true + }, + { + "id": 128042, + "content": "<|reserved_special_token_34|>", + "single_word": false, + "lstrip": false, + "rstrip": false, + "normalized": false, + "special": true + }, + { + "id": 128043, + "content": "<|reserved_special_token_35|>", + "single_word": false, + "lstrip": false, + "rstrip": false, + "normalized": false, + "special": true + }, + { + "id": 128044, + "content": "<|reserved_special_token_36|>", + "single_word": false, + "lstrip": false, + "rstrip": false, + "normalized": false, + "special": true + }, + { + "id": 128045, + "content": "<|reserved_special_token_37|>", + "single_word": false, + "lstrip": false, + "rstrip": false, + "normalized": false, + "special": true + }, + { + "id": 128046, + "content": "<|reserved_special_token_38|>", + "single_word": false, + "lstrip": false, + "rstrip": false, + "normalized": false, + "special": true + }, + { + "id": 128047, + "content": "<|reserved_special_token_39|>", + "single_word": false, + "lstrip": false, + "rstrip": false, + "normalized": false, + "special": true + }, + { + "id": 128048, + "content": "<|reserved_special_token_40|>", + "single_word": false, + "lstrip": false, + "rstrip": false, + "normalized": false, + "special": true + }, + { + "id": 128049, + "content": "<|reserved_special_token_41|>", + "single_word": false, + "lstrip": false, + "rstrip": false, + "normalized": false, + "special": true + }, + { + "id": 128050, + "content": "<|reserved_special_token_42|>", + "single_word": false, + "lstrip": false, + "rstrip": false, + "normalized": false, + "special": true + }, + { + "id": 128051, + "content": "<|reserved_special_token_43|>", + "single_word": false, + "lstrip": false, + "rstrip": false, + "normalized": false, + "special": true + }, + { + "id": 128052, + "content": "<|reserved_special_token_44|>", + "single_word": false, + "lstrip": false, + "rstrip": false, + "normalized": false, + "special": true + }, + { + "id": 128053, + "content": "<|reserved_special_token_45|>", + "single_word": false, + "lstrip": false, + "rstrip": false, + "normalized": false, + "special": true + }, + { + "id": 128054, + "content": "<|reserved_special_token_46|>", + "single_word": false, + "lstrip": false, + "rstrip": false, + "normalized": false, + "special": true + }, + { + "id": 128055, + "content": "<|reserved_special_token_47|>", + "single_word": false, + "lstrip": false, + "rstrip": false, + "normalized": false, + "special": true + }, + { + "id": 128056, + "content": "<|reserved_special_token_48|>", + "single_word": false, + "lstrip": false, + "rstrip": false, + "normalized": false, + "special": true + }, + { + "id": 128057, + "content": "<|reserved_special_token_49|>", + "single_word": false, + "lstrip": false, + "rstrip": false, + "normalized": false, + "special": true + }, + { + "id": 128058, + "content": "<|reserved_special_token_50|>", + "single_word": false, + "lstrip": false, + "rstrip": false, + "normalized": false, + "special": true + }, + { + "id": 128059, + "content": "<|reserved_special_token_51|>", + "single_word": false, + "lstrip": false, + "rstrip": false, + "normalized": false, + "special": true + }, + { + "id": 128060, + "content": "<|reserved_special_token_52|>", + "single_word": false, + "lstrip": false, + "rstrip": false, + "normalized": false, + "special": true + }, + { + "id": 128061, + "content": "<|reserved_special_token_53|>", + "single_word": false, + "lstrip": false, + "rstrip": false, + "normalized": false, + "special": true + }, + { + "id": 128062, + "content": "<|reserved_special_token_54|>", + "single_word": false, + "lstrip": false, + "rstrip": false, + "normalized": false, + "special": true + }, + { + "id": 128063, + "content": "<|reserved_special_token_55|>", + "single_word": false, + "lstrip": false, + "rstrip": false, + "normalized": false, + "special": true + }, + { + "id": 128064, + "content": "<|reserved_special_token_56|>", + "single_word": false, + "lstrip": false, + "rstrip": false, + "normalized": false, + "special": true + }, + { + "id": 128065, + "content": "<|reserved_special_token_57|>", + "single_word": false, + "lstrip": false, + "rstrip": false, + "normalized": false, + "special": true + }, + { + "id": 128066, + "content": "<|reserved_special_token_58|>", + "single_word": false, + "lstrip": false, + "rstrip": false, + "normalized": false, + "special": true + }, + { + "id": 128067, + "content": "<|reserved_special_token_59|>", + "single_word": false, + "lstrip": false, + "rstrip": false, + "normalized": false, + "special": true + }, + { + "id": 128068, + "content": "<|reserved_special_token_60|>", + "single_word": false, + "lstrip": false, + "rstrip": false, + "normalized": false, + "special": true + }, + { + "id": 128069, + "content": "<|reserved_special_token_61|>", + "single_word": false, + "lstrip": false, + "rstrip": false, + "normalized": false, + "special": true + }, + { + "id": 128070, + "content": "<|reserved_special_token_62|>", + "single_word": false, + "lstrip": false, + "rstrip": false, + "normalized": false, + "special": true + }, + { + "id": 128071, + "content": "<|reserved_special_token_63|>", + "single_word": false, + "lstrip": false, + "rstrip": false, + "normalized": false, + "special": true + }, + { + "id": 128072, + "content": "<|reserved_special_token_64|>", + "single_word": false, + "lstrip": false, + "rstrip": false, + "normalized": false, + "special": true + }, + { + "id": 128073, + "content": "<|reserved_special_token_65|>", + "single_word": false, + "lstrip": false, + "rstrip": false, + "normalized": false, + "special": true + }, + { + "id": 128074, + "content": "<|reserved_special_token_66|>", + "single_word": false, + "lstrip": false, + "rstrip": false, + "normalized": false, + "special": true + }, + { + "id": 128075, + "content": "<|reserved_special_token_67|>", + "single_word": false, + "lstrip": false, + "rstrip": false, + "normalized": false, + "special": true + }, + { + "id": 128076, + "content": "<|reserved_special_token_68|>", + "single_word": false, + "lstrip": false, + "rstrip": false, + "normalized": false, + "special": true + }, + { + "id": 128077, + "content": "<|reserved_special_token_69|>", + "single_word": false, + "lstrip": false, + "rstrip": false, + "normalized": false, + "special": true + }, + { + "id": 128078, + "content": "<|reserved_special_token_70|>", + "single_word": false, + "lstrip": false, + "rstrip": false, + "normalized": false, + "special": true + }, + { + "id": 128079, + "content": "<|reserved_special_token_71|>", + "single_word": false, + "lstrip": false, + "rstrip": false, + "normalized": false, + "special": true + }, + { + "id": 128080, + "content": "<|reserved_special_token_72|>", + "single_word": false, + "lstrip": false, + "rstrip": false, + "normalized": false, + "special": true + }, + { + "id": 128081, + "content": "<|reserved_special_token_73|>", + "single_word": false, + "lstrip": false, + "rstrip": false, + "normalized": false, + "special": true + }, + { + "id": 128082, + "content": "<|reserved_special_token_74|>", + "single_word": false, + "lstrip": false, + "rstrip": false, + "normalized": false, + "special": true + }, + { + "id": 128083, + "content": "<|reserved_special_token_75|>", + "single_word": false, + "lstrip": false, + "rstrip": false, + "normalized": false, + "special": true + }, + { + "id": 128084, + "content": "<|reserved_special_token_76|>", + "single_word": false, + "lstrip": false, + "rstrip": false, + "normalized": false, + "special": true + }, + { + "id": 128085, + "content": "<|reserved_special_token_77|>", + "single_word": false, + "lstrip": false, + "rstrip": false, + "normalized": false, + "special": true + }, + { + "id": 128086, + "content": "<|reserved_special_token_78|>", + "single_word": false, + "lstrip": false, + "rstrip": false, + "normalized": false, + "special": true + }, + { + "id": 128087, + "content": "<|reserved_special_token_79|>", + "single_word": false, + "lstrip": false, + "rstrip": false, + "normalized": false, + "special": true + }, + { + "id": 128088, + "content": "<|reserved_special_token_80|>", + "single_word": false, + "lstrip": false, + "rstrip": false, + "normalized": false, + "special": true + }, + { + "id": 128089, + "content": "<|reserved_special_token_81|>", + "single_word": false, + "lstrip": false, + "rstrip": false, + "normalized": false, + "special": true + }, + { + "id": 128090, + "content": "<|reserved_special_token_82|>", + "single_word": false, + "lstrip": false, + "rstrip": false, + "normalized": false, + "special": true + }, + { + "id": 128091, + "content": "<|reserved_special_token_83|>", + "single_word": false, + "lstrip": false, + "rstrip": false, + "normalized": false, + "special": true + }, + { + "id": 128092, + "content": "<|reserved_special_token_84|>", + "single_word": false, + "lstrip": false, + "rstrip": false, + "normalized": false, + "special": true + }, + { + "id": 128093, + "content": "<|reserved_special_token_85|>", + "single_word": false, + "lstrip": false, + "rstrip": false, + "normalized": false, + "special": true + }, + { + "id": 128094, + "content": "<|reserved_special_token_86|>", + "single_word": false, + "lstrip": false, + "rstrip": false, + "normalized": false, + "special": true + }, + { + "id": 128095, + "content": "<|reserved_special_token_87|>", + "single_word": false, + "lstrip": false, + "rstrip": false, + "normalized": false, + "special": true + }, + { + "id": 128096, + "content": "<|reserved_special_token_88|>", + "single_word": false, + "lstrip": false, + "rstrip": false, + "normalized": false, + "special": true + }, + { + "id": 128097, + "content": "<|reserved_special_token_89|>", + "single_word": false, + "lstrip": false, + "rstrip": false, + "normalized": false, + "special": true + }, + { + "id": 128098, + "content": "<|reserved_special_token_90|>", + "single_word": false, + "lstrip": false, + "rstrip": false, + "normalized": false, + "special": true + }, + { + "id": 128099, + "content": "<|reserved_special_token_91|>", + "single_word": false, + "lstrip": false, + "rstrip": false, + "normalized": false, + "special": true + }, + { + "id": 128100, + "content": "<|reserved_special_token_92|>", + "single_word": false, + "lstrip": false, + "rstrip": false, + "normalized": false, + "special": true + }, + { + "id": 128101, + "content": "<|reserved_special_token_93|>", + "single_word": false, + "lstrip": false, + "rstrip": false, + "normalized": false, + "special": true + }, + { + "id": 128102, + "content": "<|reserved_special_token_94|>", + "single_word": false, + "lstrip": false, + "rstrip": false, + "normalized": false, + "special": true + }, + { + "id": 128103, + "content": "<|reserved_special_token_95|>", + "single_word": false, + "lstrip": false, + "rstrip": false, + "normalized": false, + "special": true + }, + { + "id": 128104, + "content": "<|reserved_special_token_96|>", + "single_word": false, + "lstrip": false, + "rstrip": false, + "normalized": false, + "special": true + }, + { + "id": 128105, + "content": "<|reserved_special_token_97|>", + "single_word": false, + "lstrip": false, + "rstrip": false, + "normalized": false, + "special": true + }, + { + "id": 128106, + "content": "<|reserved_special_token_98|>", + "single_word": false, + "lstrip": false, + "rstrip": false, + "normalized": false, + "special": true + }, + { + "id": 128107, + "content": "<|reserved_special_token_99|>", + "single_word": false, + "lstrip": false, + "rstrip": false, + "normalized": false, + "special": true + }, + { + "id": 128108, + "content": "<|reserved_special_token_100|>", + "single_word": false, + "lstrip": false, + "rstrip": false, + "normalized": false, + "special": true + }, + { + "id": 128109, + "content": "<|reserved_special_token_101|>", + "single_word": false, + "lstrip": false, + "rstrip": false, + "normalized": false, + "special": true + }, + { + "id": 128110, + "content": "<|reserved_special_token_102|>", + "single_word": false, + "lstrip": false, + "rstrip": false, + "normalized": false, + "special": true + }, + { + "id": 128111, + "content": "<|reserved_special_token_103|>", + "single_word": false, + "lstrip": false, + "rstrip": false, + "normalized": false, + "special": true + }, + { + "id": 128112, + "content": "<|reserved_special_token_104|>", + "single_word": false, + "lstrip": false, + "rstrip": false, + "normalized": false, + "special": true + }, + { + "id": 128113, + "content": "<|reserved_special_token_105|>", + "single_word": false, + "lstrip": false, + "rstrip": false, + "normalized": false, + "special": true + }, + { + "id": 128114, + "content": "<|reserved_special_token_106|>", + "single_word": false, + "lstrip": false, + "rstrip": false, + "normalized": false, + "special": true + }, + { + "id": 128115, + "content": "<|reserved_special_token_107|>", + "single_word": false, + "lstrip": false, + "rstrip": false, + "normalized": false, + "special": true + }, + { + "id": 128116, + "content": "<|reserved_special_token_108|>", + "single_word": false, + "lstrip": false, + "rstrip": false, + "normalized": false, + "special": true + }, + { + "id": 128117, + "content": "<|reserved_special_token_109|>", + "single_word": false, + "lstrip": false, + "rstrip": false, + "normalized": false, + "special": true + }, + { + "id": 128118, + "content": "<|reserved_special_token_110|>", + "single_word": false, + "lstrip": false, + "rstrip": false, + "normalized": false, + "special": true + }, + { + "id": 128119, + "content": "<|reserved_special_token_111|>", + "single_word": false, + "lstrip": false, + "rstrip": false, + "normalized": false, + "special": true + }, + { + "id": 128120, + "content": "<|reserved_special_token_112|>", + "single_word": false, + "lstrip": false, + "rstrip": false, + "normalized": false, + "special": true + }, + { + "id": 128121, + "content": "<|reserved_special_token_113|>", + "single_word": false, + "lstrip": false, + "rstrip": false, + "normalized": false, + "special": true + }, + { + "id": 128122, + "content": "<|reserved_special_token_114|>", + "single_word": false, + "lstrip": false, + "rstrip": false, + "normalized": false, + "special": true + }, + { + "id": 128123, + "content": "<|reserved_special_token_115|>", + "single_word": false, + "lstrip": false, + "rstrip": false, + "normalized": false, + "special": true + }, + { + "id": 128124, + "content": "<|reserved_special_token_116|>", + "single_word": false, + "lstrip": false, + "rstrip": false, + "normalized": false, + "special": true + }, + { + "id": 128125, + "content": "<|reserved_special_token_117|>", + "single_word": false, + "lstrip": false, + "rstrip": false, + "normalized": false, + "special": true + }, + { + "id": 128126, + "content": "<|reserved_special_token_118|>", + "single_word": false, + "lstrip": false, + "rstrip": false, + "normalized": false, + "special": true + }, + { + "id": 128127, + "content": "<|reserved_special_token_119|>", + "single_word": false, + "lstrip": false, + "rstrip": false, + "normalized": false, + "special": true + }, + { + "id": 128128, + "content": "<|reserved_special_token_120|>", + "single_word": false, + "lstrip": false, + "rstrip": false, + "normalized": false, + "special": true + }, + { + "id": 128129, + "content": "<|reserved_special_token_121|>", + "single_word": false, + "lstrip": false, + "rstrip": false, + "normalized": false, + "special": true + }, + { + "id": 128130, + "content": "<|reserved_special_token_122|>", + "single_word": false, + "lstrip": false, + "rstrip": false, + "normalized": false, + "special": true + }, + { + "id": 128131, + "content": "<|reserved_special_token_123|>", + "single_word": false, + "lstrip": false, + "rstrip": false, + "normalized": false, + "special": true + }, + { + "id": 128132, + "content": "<|reserved_special_token_124|>", + "single_word": false, + "lstrip": false, + "rstrip": false, + "normalized": false, + "special": true + }, + { + "id": 128133, + "content": "<|reserved_special_token_125|>", + "single_word": false, + "lstrip": false, + "rstrip": false, + "normalized": false, + "special": true + }, + { + "id": 128134, + "content": "<|reserved_special_token_126|>", + "single_word": false, + "lstrip": false, + "rstrip": false, + "normalized": false, + "special": true + }, + { + "id": 128135, + "content": "<|reserved_special_token_127|>", + "single_word": false, + "lstrip": false, + "rstrip": false, + "normalized": false, + "special": true + }, + { + "id": 128136, + "content": "<|reserved_special_token_128|>", + "single_word": false, + "lstrip": false, + "rstrip": false, + "normalized": false, + "special": true + }, + { + "id": 128137, + "content": "<|reserved_special_token_129|>", + "single_word": false, + "lstrip": false, + "rstrip": false, + "normalized": false, + "special": true + }, + { + "id": 128138, + "content": "<|reserved_special_token_130|>", + "single_word": false, + "lstrip": false, + "rstrip": false, + "normalized": false, + "special": true + }, + { + "id": 128139, + "content": "<|reserved_special_token_131|>", + "single_word": false, + "lstrip": false, + "rstrip": false, + "normalized": false, + "special": true + }, + { + "id": 128140, + "content": "<|reserved_special_token_132|>", + "single_word": false, + "lstrip": false, + "rstrip": false, + "normalized": false, + "special": true + }, + { + "id": 128141, + "content": "<|reserved_special_token_133|>", + "single_word": false, + "lstrip": false, + "rstrip": false, + "normalized": false, + "special": true + }, + { + "id": 128142, + "content": "<|reserved_special_token_134|>", + "single_word": false, + "lstrip": false, + "rstrip": false, + "normalized": false, + "special": true + }, + { + "id": 128143, + "content": "<|reserved_special_token_135|>", + "single_word": false, + "lstrip": false, + "rstrip": false, + "normalized": false, + "special": true + }, + { + "id": 128144, + "content": "<|reserved_special_token_136|>", + "single_word": false, + "lstrip": false, + "rstrip": false, + "normalized": false, + "special": true + }, + { + "id": 128145, + "content": "<|reserved_special_token_137|>", + "single_word": false, + "lstrip": false, + "rstrip": false, + "normalized": false, + "special": true + }, + { + "id": 128146, + "content": "<|reserved_special_token_138|>", + "single_word": false, + "lstrip": false, + "rstrip": false, + "normalized": false, + "special": true + }, + { + "id": 128147, + "content": "<|reserved_special_token_139|>", + "single_word": false, + "lstrip": false, + "rstrip": false, + "normalized": false, + "special": true + }, + { + "id": 128148, + "content": "<|reserved_special_token_140|>", + "single_word": false, + "lstrip": false, + "rstrip": false, + "normalized": false, + "special": true + }, + { + "id": 128149, + "content": "<|reserved_special_token_141|>", + "single_word": false, + "lstrip": false, + "rstrip": false, + "normalized": false, + "special": true + }, + { + "id": 128150, + "content": "<|reserved_special_token_142|>", + "single_word": false, + "lstrip": false, + "rstrip": false, + "normalized": false, + "special": true + }, + { + "id": 128151, + "content": "<|reserved_special_token_143|>", + "single_word": false, + "lstrip": false, + "rstrip": false, + "normalized": false, + "special": true + }, + { + "id": 128152, + "content": "<|reserved_special_token_144|>", + "single_word": false, + "lstrip": false, + "rstrip": false, + "normalized": false, + "special": true + }, + { + "id": 128153, + "content": "<|reserved_special_token_145|>", + "single_word": false, + "lstrip": false, + "rstrip": false, + "normalized": false, + "special": true + }, + { + "id": 128154, + "content": "<|reserved_special_token_146|>", + "single_word": false, + "lstrip": false, + "rstrip": false, + "normalized": false, + "special": true + }, + { + "id": 128155, + "content": "<|reserved_special_token_147|>", + "single_word": false, + "lstrip": false, + "rstrip": false, + "normalized": false, + "special": true + }, + { + "id": 128156, + "content": "<|reserved_special_token_148|>", + "single_word": false, + "lstrip": false, + "rstrip": false, + "normalized": false, + "special": true + }, + { + "id": 128157, + "content": "<|reserved_special_token_149|>", + "single_word": false, + "lstrip": false, + "rstrip": false, + "normalized": false, + "special": true + }, + { + "id": 128158, + "content": "<|reserved_special_token_150|>", + "single_word": false, + "lstrip": false, + "rstrip": false, + "normalized": false, + "special": true + }, + { + "id": 128159, + "content": "<|reserved_special_token_151|>", + "single_word": false, + "lstrip": false, + "rstrip": false, + "normalized": false, + "special": true + }, + { + "id": 128160, + "content": "<|reserved_special_token_152|>", + "single_word": false, + "lstrip": false, + "rstrip": false, + "normalized": false, + "special": true + }, + { + "id": 128161, + "content": "<|reserved_special_token_153|>", + "single_word": false, + "lstrip": false, + "rstrip": false, + "normalized": false, + "special": true + }, + { + "id": 128162, + "content": "<|reserved_special_token_154|>", + "single_word": false, + "lstrip": false, + "rstrip": false, + "normalized": false, + "special": true + }, + { + "id": 128163, + "content": "<|reserved_special_token_155|>", + "single_word": false, + "lstrip": false, + "rstrip": false, + "normalized": false, + "special": true + }, + { + "id": 128164, + "content": "<|reserved_special_token_156|>", + "single_word": false, + "lstrip": false, + "rstrip": false, + "normalized": false, + "special": true + }, + { + "id": 128165, + "content": "<|reserved_special_token_157|>", + "single_word": false, + "lstrip": false, + "rstrip": false, + "normalized": false, + "special": true + }, + { + "id": 128166, + "content": "<|reserved_special_token_158|>", + "single_word": false, + "lstrip": false, + "rstrip": false, + "normalized": false, + "special": true + }, + { + "id": 128167, + "content": "<|reserved_special_token_159|>", + "single_word": false, + "lstrip": false, + "rstrip": false, + "normalized": false, + "special": true + }, + { + "id": 128168, + "content": "<|reserved_special_token_160|>", + "single_word": false, + "lstrip": false, + "rstrip": false, + "normalized": false, + "special": true + }, + { + "id": 128169, + "content": "<|reserved_special_token_161|>", + "single_word": false, + "lstrip": false, + "rstrip": false, + "normalized": false, + "special": true + }, + { + "id": 128170, + "content": "<|reserved_special_token_162|>", + "single_word": false, + "lstrip": false, + "rstrip": false, + "normalized": false, + "special": true + }, + { + "id": 128171, + "content": "<|reserved_special_token_163|>", + "single_word": false, + "lstrip": false, + "rstrip": false, + "normalized": false, + "special": true + }, + { + "id": 128172, + "content": "<|reserved_special_token_164|>", + "single_word": false, + "lstrip": false, + "rstrip": false, + "normalized": false, + "special": true + }, + { + "id": 128173, + "content": "<|reserved_special_token_165|>", + "single_word": false, + "lstrip": false, + "rstrip": false, + "normalized": false, + "special": true + }, + { + "id": 128174, + "content": "<|reserved_special_token_166|>", + "single_word": false, + "lstrip": false, + "rstrip": false, + "normalized": false, + "special": true + }, + { + "id": 128175, + "content": "<|reserved_special_token_167|>", + "single_word": false, + "lstrip": false, + "rstrip": false, + "normalized": false, + "special": true + }, + { + "id": 128176, + "content": "<|reserved_special_token_168|>", + "single_word": false, + "lstrip": false, + "rstrip": false, + "normalized": false, + "special": true + }, + { + "id": 128177, + "content": "<|reserved_special_token_169|>", + "single_word": false, + "lstrip": false, + "rstrip": false, + "normalized": false, + "special": true + }, + { + "id": 128178, + "content": "<|reserved_special_token_170|>", + "single_word": false, + "lstrip": false, + "rstrip": false, + "normalized": false, + "special": true + }, + { + "id": 128179, + "content": "<|reserved_special_token_171|>", + "single_word": false, + "lstrip": false, + "rstrip": false, + "normalized": false, + "special": true + }, + { + "id": 128180, + "content": "<|reserved_special_token_172|>", + "single_word": false, + "lstrip": false, + "rstrip": false, + "normalized": false, + "special": true + }, + { + "id": 128181, + "content": "<|reserved_special_token_173|>", + "single_word": false, + "lstrip": false, + "rstrip": false, + "normalized": false, + "special": true + }, + { + "id": 128182, + "content": "<|reserved_special_token_174|>", + "single_word": false, + "lstrip": false, + "rstrip": false, + "normalized": false, + "special": true + }, + { + "id": 128183, + "content": "<|reserved_special_token_175|>", + "single_word": false, + "lstrip": false, + "rstrip": false, + "normalized": false, + "special": true + }, + { + "id": 128184, + "content": "<|reserved_special_token_176|>", + "single_word": false, + "lstrip": false, + "rstrip": false, + "normalized": false, + "special": true + }, + { + "id": 128185, + "content": "<|reserved_special_token_177|>", + "single_word": false, + "lstrip": false, + "rstrip": false, + "normalized": false, + "special": true + }, + { + "id": 128186, + "content": "<|reserved_special_token_178|>", + "single_word": false, + "lstrip": false, + "rstrip": false, + "normalized": false, + "special": true + }, + { + "id": 128187, + "content": "<|reserved_special_token_179|>", + "single_word": false, + "lstrip": false, + "rstrip": false, + "normalized": false, + "special": true + }, + { + "id": 128188, + "content": "<|reserved_special_token_180|>", + "single_word": false, + "lstrip": false, + "rstrip": false, + "normalized": false, + "special": true + }, + { + "id": 128189, + "content": "<|reserved_special_token_181|>", + "single_word": false, + "lstrip": false, + "rstrip": false, + "normalized": false, + "special": true + }, + { + "id": 128190, + "content": "<|reserved_special_token_182|>", + "single_word": false, + "lstrip": false, + "rstrip": false, + "normalized": false, + "special": true + }, + { + "id": 128191, + "content": "<|reserved_special_token_183|>", + "single_word": false, + "lstrip": false, + "rstrip": false, + "normalized": false, + "special": true + }, + { + "id": 128192, + "content": "<|reserved_special_token_184|>", + "single_word": false, + "lstrip": false, + "rstrip": false, + "normalized": false, + "special": true + }, + { + "id": 128193, + "content": "<|reserved_special_token_185|>", + "single_word": false, + "lstrip": false, + "rstrip": false, + "normalized": false, + "special": true + }, + { + "id": 128194, + "content": "<|reserved_special_token_186|>", + "single_word": false, + "lstrip": false, + "rstrip": false, + "normalized": false, + "special": true + }, + { + "id": 128195, + "content": "<|reserved_special_token_187|>", + "single_word": false, + "lstrip": false, + "rstrip": false, + "normalized": false, + "special": true + }, + { + "id": 128196, + "content": "<|reserved_special_token_188|>", + "single_word": false, + "lstrip": false, + "rstrip": false, + "normalized": false, + "special": true + }, + { + "id": 128197, + "content": "<|reserved_special_token_189|>", + "single_word": false, + "lstrip": false, + "rstrip": false, + "normalized": false, + "special": true + }, + { + "id": 128198, + "content": "<|reserved_special_token_190|>", + "single_word": false, + "lstrip": false, + "rstrip": false, + "normalized": false, + "special": true + }, + { + "id": 128199, + "content": "<|reserved_special_token_191|>", + "single_word": false, + "lstrip": false, + "rstrip": false, + "normalized": false, + "special": true + }, + { + "id": 128200, + "content": "<|reserved_special_token_192|>", + "single_word": false, + "lstrip": false, + "rstrip": false, + "normalized": false, + "special": true + }, + { + "id": 128201, + "content": "<|reserved_special_token_193|>", + "single_word": false, + "lstrip": false, + "rstrip": false, + "normalized": false, + "special": true + }, + { + "id": 128202, + "content": "<|reserved_special_token_194|>", + "single_word": false, + "lstrip": false, + "rstrip": false, + "normalized": false, + "special": true + }, + { + "id": 128203, + "content": "<|reserved_special_token_195|>", + "single_word": false, + "lstrip": false, + "rstrip": false, + "normalized": false, + "special": true + }, + { + "id": 128204, + "content": "<|reserved_special_token_196|>", + "single_word": false, + "lstrip": false, + "rstrip": false, + "normalized": false, + "special": true + }, + { + "id": 128205, + "content": "<|reserved_special_token_197|>", + "single_word": false, + "lstrip": false, + "rstrip": false, + "normalized": false, + "special": true + }, + { + "id": 128206, + "content": "<|reserved_special_token_198|>", + "single_word": false, + "lstrip": false, + "rstrip": false, + "normalized": false, + "special": true + }, + { + "id": 128207, + "content": "<|reserved_special_token_199|>", + "single_word": false, + "lstrip": false, + "rstrip": false, + "normalized": false, + "special": true + }, + { + "id": 128208, + "content": "<|reserved_special_token_200|>", + "single_word": false, + "lstrip": false, + "rstrip": false, + "normalized": false, + "special": true + }, + { + "id": 128209, + "content": "<|reserved_special_token_201|>", + "single_word": false, + "lstrip": false, + "rstrip": false, + "normalized": false, + "special": true + }, + { + "id": 128210, + "content": "<|reserved_special_token_202|>", + "single_word": false, + "lstrip": false, + "rstrip": false, + "normalized": false, + "special": true + }, + { + "id": 128211, + "content": "<|reserved_special_token_203|>", + "single_word": false, + "lstrip": false, + "rstrip": false, + "normalized": false, + "special": true + }, + { + "id": 128212, + "content": "<|reserved_special_token_204|>", + "single_word": false, + "lstrip": false, + "rstrip": false, + "normalized": false, + "special": true + }, + { + "id": 128213, + "content": "<|reserved_special_token_205|>", + "single_word": false, + "lstrip": false, + "rstrip": false, + "normalized": false, + "special": true + }, + { + "id": 128214, + "content": "<|reserved_special_token_206|>", + "single_word": false, + "lstrip": false, + "rstrip": false, + "normalized": false, + "special": true + }, + { + "id": 128215, + "content": "<|reserved_special_token_207|>", + "single_word": false, + "lstrip": false, + "rstrip": false, + "normalized": false, + "special": true + }, + { + "id": 128216, + "content": "<|reserved_special_token_208|>", + "single_word": false, + "lstrip": false, + "rstrip": false, + "normalized": false, + "special": true + }, + { + "id": 128217, + "content": "<|reserved_special_token_209|>", + "single_word": false, + "lstrip": false, + "rstrip": false, + "normalized": false, + "special": true + }, + { + "id": 128218, + "content": "<|reserved_special_token_210|>", + "single_word": false, + "lstrip": false, + "rstrip": false, + "normalized": false, + "special": true + }, + { + "id": 128219, + "content": "<|reserved_special_token_211|>", + "single_word": false, + "lstrip": false, + "rstrip": false, + "normalized": false, + "special": true + }, + { + "id": 128220, + "content": "<|reserved_special_token_212|>", + "single_word": false, + "lstrip": false, + "rstrip": false, + "normalized": false, + "special": true + }, + { + "id": 128221, + "content": "<|reserved_special_token_213|>", + "single_word": false, + "lstrip": false, + "rstrip": false, + "normalized": false, + "special": true + }, + { + "id": 128222, + "content": "<|reserved_special_token_214|>", + "single_word": false, + "lstrip": false, + "rstrip": false, + "normalized": false, + "special": true + }, + { + "id": 128223, + "content": "<|reserved_special_token_215|>", + "single_word": false, + "lstrip": false, + "rstrip": false, + "normalized": false, + "special": true + }, + { + "id": 128224, + "content": "<|reserved_special_token_216|>", + "single_word": false, + "lstrip": false, + "rstrip": false, + "normalized": false, + "special": true + }, + { + "id": 128225, + "content": "<|reserved_special_token_217|>", + "single_word": false, + "lstrip": false, + "rstrip": false, + "normalized": false, + "special": true + }, + { + "id": 128226, + "content": "<|reserved_special_token_218|>", + "single_word": false, + "lstrip": false, + "rstrip": false, + "normalized": false, + "special": true + }, + { + "id": 128227, + "content": "<|reserved_special_token_219|>", + "single_word": false, + "lstrip": false, + "rstrip": false, + "normalized": false, + "special": true + }, + { + "id": 128228, + "content": "<|reserved_special_token_220|>", + "single_word": false, + "lstrip": false, + "rstrip": false, + "normalized": false, + "special": true + }, + { + "id": 128229, + "content": "<|reserved_special_token_221|>", + "single_word": false, + "lstrip": false, + "rstrip": false, + "normalized": false, + "special": true + }, + { + "id": 128230, + "content": "<|reserved_special_token_222|>", + "single_word": false, + "lstrip": false, + "rstrip": false, + "normalized": false, + "special": true + }, + { + "id": 128231, + "content": "<|reserved_special_token_223|>", + "single_word": false, + "lstrip": false, + "rstrip": false, + "normalized": false, + "special": true + }, + { + "id": 128232, + "content": "<|reserved_special_token_224|>", + "single_word": false, + "lstrip": false, + "rstrip": false, + "normalized": false, + "special": true + }, + { + "id": 128233, + "content": "<|reserved_special_token_225|>", + "single_word": false, + "lstrip": false, + "rstrip": false, + "normalized": false, + "special": true + }, + { + "id": 128234, + "content": "<|reserved_special_token_226|>", + "single_word": false, + "lstrip": false, + "rstrip": false, + "normalized": false, + "special": true + }, + { + "id": 128235, + "content": "<|reserved_special_token_227|>", + "single_word": false, + "lstrip": false, + "rstrip": false, + "normalized": false, + "special": true + }, + { + "id": 128236, + "content": "<|reserved_special_token_228|>", + "single_word": false, + "lstrip": false, + "rstrip": false, + "normalized": false, + "special": true + }, + { + "id": 128237, + "content": "<|reserved_special_token_229|>", + "single_word": false, + "lstrip": false, + "rstrip": false, + "normalized": false, + "special": true + }, + { + "id": 128238, + "content": "<|reserved_special_token_230|>", + "single_word": false, + "lstrip": false, + "rstrip": false, + "normalized": false, + "special": true + }, + { + "id": 128239, + "content": "<|reserved_special_token_231|>", + "single_word": false, + "lstrip": false, + "rstrip": false, + "normalized": false, + "special": true + }, + { + "id": 128240, + "content": "<|reserved_special_token_232|>", + "single_word": false, + "lstrip": false, + "rstrip": false, + "normalized": false, + "special": true + }, + { + "id": 128241, + "content": "<|reserved_special_token_233|>", + "single_word": false, + "lstrip": false, + "rstrip": false, + "normalized": false, + "special": true + }, + { + "id": 128242, + "content": "<|reserved_special_token_234|>", + "single_word": false, + "lstrip": false, + "rstrip": false, + "normalized": false, + "special": true + }, + { + "id": 128243, + "content": "<|reserved_special_token_235|>", + "single_word": false, + "lstrip": false, + "rstrip": false, + "normalized": false, + "special": true + }, + { + "id": 128244, + "content": "<|reserved_special_token_236|>", + "single_word": false, + "lstrip": false, + "rstrip": false, + "normalized": false, + "special": true + }, + { + "id": 128245, + "content": "<|reserved_special_token_237|>", + "single_word": false, + "lstrip": false, + "rstrip": false, + "normalized": false, + "special": true + }, + { + "id": 128246, + "content": "<|reserved_special_token_238|>", + "single_word": false, + "lstrip": false, + "rstrip": false, + "normalized": false, + "special": true + }, + { + "id": 128247, + "content": "<|reserved_special_token_239|>", + "single_word": false, + "lstrip": false, + "rstrip": false, + "normalized": false, + "special": true + }, + { + "id": 128248, + "content": "<|reserved_special_token_240|>", + "single_word": false, + "lstrip": false, + "rstrip": false, + "normalized": false, + "special": true + }, + { + "id": 128249, + "content": "<|reserved_special_token_241|>", + "single_word": false, + "lstrip": false, + "rstrip": false, + "normalized": false, + "special": true + }, + { + "id": 128250, + "content": "<|reserved_special_token_242|>", + "single_word": false, + "lstrip": false, + "rstrip": false, + "normalized": false, + "special": true + }, + { + "id": 128251, + "content": "<|reserved_special_token_243|>", + "single_word": false, + "lstrip": false, + "rstrip": false, + "normalized": false, + "special": true + }, + { + "id": 128252, + "content": "<|reserved_special_token_244|>", + "single_word": false, + "lstrip": false, + "rstrip": false, + "normalized": false, + "special": true + }, + { + "id": 128253, + "content": "<|reserved_special_token_245|>", + "single_word": false, + "lstrip": false, + "rstrip": false, + "normalized": false, + "special": true + }, + { + "id": 128254, + "content": "<|reserved_special_token_246|>", + "single_word": false, + "lstrip": false, + "rstrip": false, + "normalized": false, + "special": true + }, + { + "id": 128255, + "content": "<|reserved_special_token_247|>", + "single_word": false, + "lstrip": false, + "rstrip": false, + "normalized": false, + "special": true + } + ], + "normalizer": null, + "pre_tokenizer": { + "type": "Sequence", + "pretokenizers": [ + { + "type": "Split", + "pattern": { + "Regex": "(?i:'s|'t|'re|'ve|'m|'ll|'d)|[^\\r\\n\\p{L}\\p{N}]?\\p{L}+|\\p{N}{1,3}| ?[^\\s\\p{L}\\p{N}]+[\\r\\n]*|\\s*[\\r\\n]+|\\s+(?!\\S)|\\s+" + }, + "behavior": "Isolated", + "invert": false + }, + { + "type": "ByteLevel", + "add_prefix_space": false, + "trim_offsets": true, + "use_regex": false + } + ] + }, + "post_processor": { + "type": "Sequence", + "processors": [ + { + "type": "ByteLevel", + "add_prefix_space": true, + "trim_offsets": false, + "use_regex": true + }, + { + "type": "TemplateProcessing", + "single": [ + { + "SpecialToken": { + "id": "<|begin_of_text|>", + "type_id": 0 + } + }, + { + "Sequence": { + "id": "A", + "type_id": 0 + } + } + ], + "pair": [ + { + "SpecialToken": { + "id": "<|begin_of_text|>", + "type_id": 0 + } + }, + { + "Sequence": { + "id": "A", + "type_id": 0 + } + }, + { + "SpecialToken": { + "id": "<|begin_of_text|>", + "type_id": 1 + } + }, + { + "Sequence": { + "id": "B", + "type_id": 1 + } + } + ], + "special_tokens": { + "<|begin_of_text|>": { + "id": "<|begin_of_text|>", + "ids": [ + 128000 + ], + "tokens": [ + "<|begin_of_text|>" + ] + } + } + } + ] + }, + "decoder": { + "type": "ByteLevel", + "add_prefix_space": true, + "trim_offsets": true, + "use_regex": true + }, + "model": { + "type": "BPE", + "dropout": null, + "unk_token": null, + "continuing_subword_prefix": null, + "end_of_word_suffix": null, + "fuse_unk": false, + "byte_fallback": false, + "ignore_merges": true, + "vocab": { + "!": 0, + "\"": 1, + "#": 2, + "$": 3, + "%": 4, + "&": 5, + "'": 6, + "(": 7, + ")": 8, + "*": 9, + "+": 10, + ",": 11, + "-": 12, + ".": 13, + "/": 14, + "0": 15, + "1": 16, + "2": 17, + "3": 18, + "4": 19, + "5": 20, + "6": 21, + "7": 22, + "8": 23, + "9": 24, + ":": 25, + ";": 26, + "<": 27, + "=": 28, + ">": 29, + "?": 30, + "@": 31, + "A": 32, + "B": 33, + "C": 34, + "D": 35, + "E": 36, + "F": 37, + "G": 38, + "H": 39, + "I": 40, + "J": 41, + "K": 42, + "L": 43, + "M": 44, + "N": 45, + "O": 46, + "P": 47, + "Q": 48, + "R": 49, + "S": 50, + "T": 51, + "U": 52, + "V": 53, + "W": 54, + "X": 55, + "Y": 56, + "Z": 57, + "[": 58, + "\\": 59, + "]": 60, + "^": 61, + "_": 62, + "`": 63, + "a": 64, + "b": 65, + "c": 66, + "d": 67, + "e": 68, + "f": 69, + "g": 70, + "h": 71, + "i": 72, + "j": 73, + "k": 74, + "l": 75, + "m": 76, + "n": 77, + "o": 78, + "p": 79, + "q": 80, + "r": 81, + "s": 82, + "t": 83, + "u": 84, + "v": 85, + "w": 86, + "x": 87, + "y": 88, + "z": 89, + "{": 90, + "|": 91, + "}": 92, + "~": 93, + "¡": 94, + "¢": 95, + "£": 96, + "¤": 97, + "¥": 98, + "¦": 99, + "§": 100, + "¨": 101, + "©": 102, + "ª": 103, + "«": 104, + "¬": 105, + "®": 106, + "¯": 107, + "°": 108, + "±": 109, + "²": 110, + "³": 111, + "´": 112, + "µ": 113, + "¶": 114, + "·": 115, + "¸": 116, + "¹": 117, + "º": 118, + "»": 119, + "¼": 120, + "½": 121, + "¾": 122, + "¿": 123, + "À": 124, + "Á": 125, + "Â": 126, + "Ã": 127, + "Ä": 128, + "Å": 129, + "Æ": 130, + "Ç": 131, + "È": 132, + "É": 133, + "Ê": 134, + "Ë": 135, + "Ì": 136, + "Í": 137, + "Î": 138, + "Ï": 139, + "Ð": 140, + "Ñ": 141, + "Ò": 142, + "Ó": 143, + "Ô": 144, + "Õ": 145, + "Ö": 146, + "×": 147, + "Ø": 148, + "Ù": 149, + "Ú": 150, + "Û": 151, + "Ü": 152, + "Ý": 153, + "Þ": 154, + "ß": 155, + "à": 156, + "á": 157, + "â": 158, + "ã": 159, + "ä": 160, + "å": 161, + "æ": 162, + "ç": 163, + "è": 164, + "é": 165, + "ê": 166, + "ë": 167, + "ì": 168, + "í": 169, + "î": 170, + "ï": 171, + "ð": 172, + "ñ": 173, + "ò": 174, + "ó": 175, + "ô": 176, + "õ": 177, + "ö": 178, + "÷": 179, + "ø": 180, + "ù": 181, + "ú": 182, + "û": 183, + "ü": 184, + "ý": 185, + "þ": 186, + "ÿ": 187, + "Ā": 188, + "ā": 189, + "Ă": 190, + "ă": 191, + "Ą": 192, + "ą": 193, + "Ć": 194, + "ć": 195, + "Ĉ": 196, + "ĉ": 197, + "Ċ": 198, + "ċ": 199, + "Č": 200, + "č": 201, + "Ď": 202, + "ď": 203, + "Đ": 204, + "đ": 205, + "Ē": 206, + "ē": 207, + "Ĕ": 208, + "ĕ": 209, + "Ė": 210, + "ė": 211, + "Ę": 212, + "ę": 213, + "Ě": 214, + "ě": 215, + "Ĝ": 216, + "ĝ": 217, + "Ğ": 218, + "ğ": 219, + "Ġ": 220, + "ġ": 221, + "Ģ": 222, + "ģ": 223, + "Ĥ": 224, + "ĥ": 225, + "Ħ": 226, + "ħ": 227, + "Ĩ": 228, + "ĩ": 229, + "Ī": 230, + "ī": 231, + "Ĭ": 232, + "ĭ": 233, + "Į": 234, + "į": 235, + "İ": 236, + "ı": 237, + "IJ": 238, + "ij": 239, + "Ĵ": 240, + "ĵ": 241, + "Ķ": 242, + "ķ": 243, + "ĸ": 244, + "Ĺ": 245, + "ĺ": 246, + "Ļ": 247, + "ļ": 248, + "Ľ": 249, + "ľ": 250, + "Ŀ": 251, + "ŀ": 252, + "Ł": 253, + "ł": 254, + "Ń": 255, + "ĠĠ": 256, + "ĠĠĠĠ": 257, + "in": 258, + "Ġt": 259, + "ĠĠĠĠĠĠĠĠ": 260, + "er": 261, + "ĠĠĠ": 262, + "on": 263, + "Ġa": 264, + "re": 265, + "at": 266, + "st": 267, + "en": 268, + "or": 269, + "Ġth": 270, + "ĊĊ": 271, + "Ġc": 272, + "le": 273, + "Ġs": 274, + "it": 275, + "an": 276, + "ar": 277, + "al": 278, + "Ġthe": 279, + ";Ċ": 280, + "Ġp": 281, + "Ġf": 282, + "ou": 283, + "Ġ=": 284, + "is": 285, + "ĠĠĠĠĠĠĠ": 286, + "ing": 287, + "es": 288, + "Ġw": 289, + "ion": 290, + "ed": 291, + "ic": 292, + "Ġb": 293, + "Ġd": 294, + "et": 295, + "Ġm": 296, + "Ġo": 297, + "ĉĉ": 298, + "ro": 299, + "as": 300, + "el": 301, + "ct": 302, + "nd": 303, + "Ġin": 304, + "Ġh": 305, + "ent": 306, + "id": 307, + "Ġn": 308, + "am": 309, + "ĠĠĠĠĠĠĠĠĠĠĠ": 310, + "Ġto": 311, + "Ġre": 312, + "--": 313, + "Ġ{": 314, + "Ġof": 315, + "om": 316, + ");Ċ": 317, + "im": 318, + "čĊ": 319, + "Ġ(": 320, + "il": 321, + "//": 322, + "Ġand": 323, + "ur": 324, + "se": 325, + "Ġl": 326, + "ex": 327, + "ĠS": 328, + "ad": 329, + "Ġ\"": 330, + "ch": 331, + "ut": 332, + "if": 333, + "**": 334, + "Ġ}": 335, + "em": 336, + "ol": 337, + "ĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠ": 338, + "th": 339, + ")Ċ": 340, + "Ġ{Ċ": 341, + "Ġg": 342, + "ig": 343, + "iv": 344, + ",Ċ": 345, + "ce": 346, + "od": 347, + "Ġv": 348, + "ate": 349, + "ĠT": 350, + "ag": 351, + "ay": 352, + "Ġ*": 353, + "ot": 354, + "us": 355, + "ĠC": 356, + "Ġst": 357, + "ĠI": 358, + "un": 359, + "ul": 360, + "ue": 361, + "ĠA": 362, + "ow": 363, + "Ġ'": 364, + "ew": 365, + "Ġ<": 366, + "ation": 367, + "()": 368, + "Ġfor": 369, + "ab": 370, + "ort": 371, + "um": 372, + "ame": 373, + "Ġis": 374, + "pe": 375, + "tr": 376, + "ck": 377, + "âĢ": 378, + "Ġy": 379, + "ist": 380, + "----": 381, + ".ĊĊ": 382, + "he": 383, + "Ġe": 384, + "lo": 385, + "ĠM": 386, + "Ġbe": 387, + "ers": 388, + "Ġon": 389, + "Ġcon": 390, + "ap": 391, + "ub": 392, + "ĠP": 393, + "ĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠ": 394, + "ass": 395, + "int": 396, + ">Ċ": 397, + "ly": 398, + "urn": 399, + "Ġ$": 400, + ";ĊĊ": 401, + "av": 402, + "port": 403, + "ir": 404, + "->": 405, + "nt": 406, + "ction": 407, + "end": 408, + "Ġde": 409, + "00": 410, + "ith": 411, + "out": 412, + "turn": 413, + "our": 414, + "ĠĠĠĠĠ": 415, + "lic": 416, + "res": 417, + "pt": 418, + "==": 419, + "Ġthis": 420, + "Ġwh": 421, + "Ġif": 422, + "ĠD": 423, + "ver": 424, + "age": 425, + "ĠB": 426, + "ht": 427, + "ext": 428, + "=\"": 429, + "Ġthat": 430, + "****": 431, + "ĠR": 432, + "Ġit": 433, + "ess": 434, + "ĠF": 435, + "Ġr": 436, + "os": 437, + "and": 438, + "Ġas": 439, + "ect": 440, + "ke": 441, + "rom": 442, + "Ġ//": 443, + "con": 444, + "ĠL": 445, + "(\"": 446, + "qu": 447, + "lass": 448, + "Ġwith": 449, + "iz": 450, + "de": 451, + "ĠN": 452, + "Ġal": 453, + "op": 454, + "up": 455, + "get": 456, + "Ġ}Ċ": 457, + "ile": 458, + "Ġan": 459, + "ata": 460, + "ore": 461, + "ri": 462, + "Ġpro": 463, + ";čĊ": 464, + "ĉĉĉĉ": 465, + "ter": 466, + "ain": 467, + "ĠW": 468, + "ĠE": 469, + "Ġcom": 470, + "Ġreturn": 471, + "art": 472, + "ĠH": 473, + "ack": 474, + "import": 475, + "ublic": 476, + "Ġor": 477, + "est": 478, + "ment": 479, + "ĠG": 480, + "able": 481, + "Ġ-": 482, + "ine": 483, + "ill": 484, + "ind": 485, + "ere": 486, + "::": 487, + "ity": 488, + "Ġ+": 489, + "Ġtr": 490, + "elf": 491, + "ight": 492, + "('": 493, + "orm": 494, + "ult": 495, + "str": 496, + "..": 497, + "\",": 498, + "Ġyou": 499, + "ype": 500, + "pl": 501, + "Ġnew": 502, + "Ġj": 503, + "ĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠ": 504, + "Ġfrom": 505, + "Ġex": 506, + "ĠO": 507, + "20": 508, + "ld": 509, + "Ġ[": 510, + "oc": 511, + ":Ċ": 512, + "Ġse": 513, + "Ġle": 514, + "--------": 515, + ".s": 516, + "{Ċ": 517, + "',": 518, + "ant": 519, + "Ġat": 520, + "ase": 521, + ".c": 522, + "Ġch": 523, + "": 524, + "ave": 525, + "ang": 526, + "Ġare": 527, + "Ġint": 528, + "âĢĻ": 529, + "_t": 530, + "ert": 531, + "ial": 532, + "act": 533, + "}Ċ": 534, + "ive": 535, + "ode": 536, + "ost": 537, + "Ġclass": 538, + "Ġnot": 539, + "og": 540, + "ord": 541, + "alue": 542, + "all": 543, + "ff": 544, + "();Ċ": 545, + "ont": 546, + "ime": 547, + "are": 548, + "ĠU": 549, + "Ġpr": 550, + "Ġ:": 551, + "ies": 552, + "ize": 553, + "ure": 554, + "Ġby": 555, + "ire": 556, + "Ġ}ĊĊ": 557, + ".p": 558, + "Ġsh": 559, + "ice": 560, + "ast": 561, + "ption": 562, + "tring": 563, + "ok": 564, + "__": 565, + "cl": 566, + "##": 567, + "Ġhe": 568, + "ard": 569, + ").": 570, + "Ġ@": 571, + "iew": 572, + "ĉĉĉ": 573, + "Ġwas": 574, + "ip": 575, + "this": 576, + "Ġu": 577, + "ĠThe": 578, + "ide": 579, + "ace": 580, + "ib": 581, + "ac": 582, + "rou": 583, + "Ġwe": 584, + "ject": 585, + "Ġpublic": 586, + "ak": 587, + "ve": 588, + "ath": 589, + "oid": 590, + "Ġ=>": 591, + "ust": 592, + "que": 593, + "Ġres": 594, + "))": 595, + "'s": 596, + "Ġk": 597, + "ans": 598, + "yst": 599, + "unction": 600, + "********": 601, + "Ġi": 602, + "Ġus": 603, + "pp": 604, + "10": 605, + "one": 606, + "ail": 607, + "====": 608, + "name": 609, + "Ġstr": 610, + "Ġ/": 611, + "Ġ&": 612, + "ach": 613, + "div": 614, + "ystem": 615, + "ell": 616, + "Ġhave": 617, + "err": 618, + "ould": 619, + "ull": 620, + "pon": 621, + "ĠJ": 622, + "_p": 623, + "Ġ==": 624, + "ign": 625, + "St": 626, + ".Ċ": 627, + "Ġpl": 628, + ");ĊĊ": 629, + "form": 630, + "put": 631, + "ount": 632, + "}ĊĊ": 633, + "dd": 634, + "ite": 635, + "Ġget": 636, + "rr": 637, + "ome": 638, + "ĠâĢ": 639, + "aram": 640, + "cc": 641, + "Ġ*/": 642, + "ER": 643, + "In": 644, + "les": 645, + "_s": 646, + "ong": 647, + "ie": 648, + "Ġcan": 649, + "ĠV": 650, + "erv": 651, + "pr": 652, + "Ġun": 653, + "row": 654, + "ber": 655, + "Ġdo": 656, + "ll": 657, + "Ġel": 658, + "Ġself": 659, + "ated": 660, + "ary": 661, + "Ġ.": 662, + "']": 663, + "ud": 664, + "Ġen": 665, + "ĠTh": 666, + "ĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠ": 667, + "te": 668, + "_c": 669, + "uct": 670, + "Ġab": 671, + "ork": 672, + ".get": 673, + "Ġ#": 674, + "aw": 675, + "ress": 676, + "ob": 677, + "Name": 678, + "201": 679, + "app": 680, + "['": 681, + "Ġall": 682, + "ory": 683, + "ition": 684, + "ance": 685, + "ear": 686, + "Ġcont": 687, + "vent": 688, + "ia": 689, + "Ġwill": 690, + "IN": 691, + "ĠĠĠĠĠĠĠĠĠ": 692, + "return": 693, + "Ġ": 694, + "data": 695, + ")ĊĊ": 696, + "Re": 697, + "ple": 698, + "ild": 699, + "ther": 700, + "Ġyour": 701, + "\"Ċ": 702, + "($": 703, + "Ġout": 704, + "),": 705, + "Ġhas": 706, + "String": 707, + "so": 708, + "Ġup": 709, + "ax": 710, + "Ġdef": 711, + "Ġbo": 712, + "ge": 713, + "alse": 714, + "ON": 715, + "per": 716, + "12": 717, + "ich": 718, + "Ġbut": 719, + "ĠĊ": 720, + "Ġ_": 721, + "_m": 722, + "add": 723, + "quest": 724, + "odel": 725, + "self": 726, + "ery": 727, + "ft": 728, + "ens": 729, + "////": 730, + "ake": 731, + ".C": 732, + "Ġgo": 733, + "Ġfunction": 734, + "ĠK": 735, + "ivate": 736, + "Ġim": 737, + "Ġconst": 738, + ".t": 739, + "Ġ*/Ċ": 740, + ");čĊ": 741, + "Ġvoid": 742, + "Ġset": 743, + "ĠSystem": 744, + "cri": 745, + "()Ċ": 746, + "li": 747, + "ĉif": 748, + ".m": 749, + "ally": 750, + "set": 751, + "ep": 752, + "âĢĻs": 753, + "bo": 754, + "def": 755, + "',Ċ": 756, + "Ġme": 757, + "Ġ!": 758, + "atch": 759, + "\">": 760, + "\",Ċ": 761, + "ec": 762, + "ĠIn": 763, + "ph": 764, + "Ġ|": 765, + "_f": 766, + "Ġvar": 767, + "ence": 768, + "Id": 769, + "ree": 770, + "ink": 771, + "lect": 772, + "ug": 773, + "eth": 774, + "Ġelse": 775, + "----------------": 776, + "19": 777, + "cont": 778, + "Ġso": 779, + "atic": 780, + "Ġlo": 781, + "pro": 782, + "ton": 783, + "ss": 784, + "own": 785, + "abel": 786, + "oint": 787, + "ous": 788, + "eld": 789, + "ST": 790, + "The": 791, + "ĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠ": 792, + "RE": 793, + "\":": 794, + "olor": 795, + "tp": 796, + "eg": 797, + "key": 798, + "ude": 799, + "ĠSt": 800, + "ound": 801, + "Ġar": 802, + "\");Ċ": 803, + "ener": 804, + "ser": 805, + "11": 806, + "bject": 807, + "essage": 808, + "fer": 809, + "Ġmore": 810, + "ations": 811, + "ents": 812, + "Ġhis": 813, + "Ġthey": 814, + ".S": 815, + "ĠY": 816, + "use": 817, + "ne": 818, + "ish": 819, + "old": 820, + "_d": 821, + "io": 822, + "ield": 823, + "Ġper": 824, + "Cont": 825, + "ings": 826, + "####": 827, + "Ġdata": 828, + "Ġsa": 829, + "ef": 830, + "fo": 831, + "Ġone": 832, + "eng": 833, + "Ġdis": 834, + "AT": 835, + "Ġname": 836, + "Ġtrue": 837, + "val": 838, + "led": 839, + ".f": 840, + "Ġne": 841, + "Ġend": 842, + "32": 843, + ".T": 844, + "16": 845, + "cre": 846, + "ark": 847, + "log": 848, + "Ex": 849, + "error": 850, + "_id": 851, + "urre": 852, + "ange": 853, + "Ġnull": 854, + "rray": 855, + "Ġmy": 856, + "pan": 857, + "ict": 858, + "ator": 859, + "View": 860, + "List": 861, + "ĉreturn": 862, + "âĢĿ": 863, + "Ġpre": 864, + "Ġx": 865, + "clude": 866, + "arg": 867, + "15": 868, + "ov": 869, + ".h": 870, + "Ġ>": 871, + "Ġtheir": 872, + "')": 873, + "irst": 874, + "ick": 875, + "gh": 876, + "LE": 877, + "OR": 878, + "Ġprivate": 879, + "tem": 880, + "čĊčĊ": 881, + "user": 882, + "Ġ)": 883, + "com": 884, + ".A": 885, + "\";Ċ": 886, + "Ġid": 887, + "read": 888, + "Ġwho": 889, + "_b": 890, + "\">Ċ": 891, + "Ġtime": 892, + "Ġman": 893, + "ry": 894, + "========": 895, + "roup": 896, + "rop": 897, + "public": 898, + "vel": 899, + "umber": 900, + "ble": 901, + "Ġwhich": 902, + "****************": 903, + "Ġany": 904, + "Ġfalse": 905, + "we": 906, + "Ġvalue": 907, + "Ġli": 908, + "\")": 909, + "nder": 910, + "gr": 911, + "Ġno": 912, + "param": 913, + "25": 914, + "fig": 915, + ".com": 916, + "Ġapp": 917, + "_l": 918, + "ions": 919, + ".D": 920, + "ĠCh": 921, + "Ġabout": 922, + "Ġadd": 923, + "Ġsu": 924, + "Ġstring": 925, + "ID": 926, + "Ġover": 927, + "string": 928, + ".l": 929, + "ource": 930, + "000": 931, + "_C": 932, + "]Ċ": 933, + "Ġqu": 934, + "ĠString": 935, + "ca": 936, + "SE": 937, + "Ġro": 938, + "sh": 939, + "ual": 940, + "Type": 941, + "son": 942, + "new": 943, + "ern": 944, + "Ġag": 945, + "AR": 946, + "];Ċ": 947, + "].": 948, + "Ġ?": 949, + "ical": 950, + "Ġdes": 951, + "uth": 952, + "ix": 953, + "ays": 954, + "Ġtype": 955, + "'t": 956, + "ault": 957, + "Ġinter": 958, + "var": 959, + ".b": 960, + "Ġpart": 961, + ".d": 962, + "urrent": 963, + "IT": 964, + "EN": 965, + "30": 966, + "enc": 967, + "(f": 968, + "ra": 969, + "value": 970, + "cho": 971, + "18": 972, + "utton": 973, + "ose": 974, + "14": 975, + "Ġ!=": 976, + "ater": 977, + "é": 978, + "reate": 979, + "oll": 980, + "pos": 981, + "yle": 982, + "ng": 983, + "AL": 984, + "using": 985, + "ames": 986, + "Ġ{čĊ": 987, + "ates": 988, + "ely": 989, + "Ġwork": 990, + "Ġem": 991, + "inal": 992, + "Ġsp": 993, + "Ġwhen": 994, + ".set": 995, + "ĠĠĠĠĠĠ": 996, + "):Ċ": 997, + "to": 998, + "quire": 999, + "indow": 1000, + "lement": 1001, + "pect": 1002, + "ash": 1003, + "[i": 1004, + "Ġuse": 1005, + ".F": 1006, + "pec": 1007, + "Ġad": 1008, + "ove": 1009, + "ception": 1010, + "ength": 1011, + "include": 1012, + "ader": 1013, + "ĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠ": 1014, + "atus": 1015, + "Th": 1016, + "itle": 1017, + "rit": 1018, + "void": 1019, + "().": 1020, + "(Ċ": 1021, + "Ġoff": 1022, + "Ġother": 1023, + "Ġ&&": 1024, + "';Ċ": 1025, + "ms": 1026, + "Ġbeen": 1027, + "Ġte": 1028, + "ml": 1029, + "co": 1030, + "nc": 1031, + "13": 1032, + "ervice": 1033, + "Ġ%": 1034, + "**Ċ": 1035, + "ann": 1036, + "ade": 1037, + "ĊĊĊĊ": 1038, + "lock": 1039, + "const": 1040, + "100": 1041, + "ponse": 1042, + "Ġsup": 1043, + "++": 1044, + "date": 1045, + "Ġacc": 1046, + "Ġhad": 1047, + "Ġbu": 1048, + "200": 1049, + "ĠRe": 1050, + "Ġwere": 1051, + "Ġfile": 1052, + "Ġwould": 1053, + "ĠâĢľ": 1054, + "ven": 1055, + "iss": 1056, + "Ġour": 1057, + "class": 1058, + "raw": 1059, + "Ġyear": 1060, + "Data": 1061, + "Ġval": 1062, + "Ġsome": 1063, + "fter": 1064, + "ys": 1065, + "Ġ///": 1066, + "round": 1067, + "view": 1068, + "Ġpe": 1069, + "Ġthere": 1070, + "Ġsaid": 1071, + "du": 1072, + "of": 1073, + "line": 1074, + "/*": 1075, + "duct": 1076, + "Ġher": 1077, + "ĠĠĠĠĠĠĠĠĠĠĠĠĠ": 1078, + "Res": 1079, + "Ġco": 1080, + "Ġcomm": 1081, + "ise": 1082, + "min": 1083, + "ĠĠĠĠĊ": 1084, + "#include": 1085, + "ethod": 1086, + ".P": 1087, + "ute": 1088, + "Ġass": 1089, + "Int": 1090, + "ask": 1091, + "loc": 1092, + "Ġlike": 1093, + "ody": 1094, + "Ġlet": 1095, + "load": 1096, + "Ġam": 1097, + "rol": 1098, + "Ġgr": 1099, + "yp": 1100, + "Ġalso": 1101, + "ĠIt": 1102, + "url": 1103, + "ific": 1104, + "ors": 1105, + "_P": 1106, + "_n": 1107, + "igh": 1108, + "Ġthan": 1109, + "Com": 1110, + "AN": 1111, + "UL": 1112, + "ating": 1113, + "17": 1114, + "ĠThis": 1115, + "ref": 1116, + "_S": 1117, + "Ġstatic": 1118, + "roll": 1119, + "Ġjust": 1120, + "Ġresult": 1121, + "ian": 1122, + "idth": 1123, + "Ġthem": 1124, + "));Ċ": 1125, + "der": 1126, + "reak": 1127, + "Con": 1128, + "://": 1129, + "ule": 1130, + "...": 1131, + "arch": 1132, + "ement": 1133, + "Ġ<<": 1134, + "50": 1135, + "ush": 1136, + "ense": 1137, + "arr": 1138, + "Ġinto": 1139, + "cess": 1140, + "amp": 1141, + "ied": 1142, + "ument": 1143, + "Ġ\\": 1144, + "],": 1145, + "wo": 1146, + "als": 1147, + "Ġwhat": 1148, + "anc": 1149, + "Value": 1150, + "='": 1151, + "olum": 1152, + "Ġpos": 1153, + "ages": 1154, + "ayer": 1155, + "Ġsc": 1156, + "ues": 1157, + "\")Ċ": 1158, + "_T": 1159, + "Ġlist": 1160, + "(s": 1161, + "Ġcase": 1162, + "Ch": 1163, + "ĉĉĉĉĉ": 1164, + "////////": 1165, + "ponent": 1166, + "Ġz": 1167, + "Ġkn": 1168, + "let": 1169, + "DE": 1170, + "red": 1171, + "Ġfe": 1172, + "Ġ},Ċ": 1173, + "Ġ,": 1174, + "(t": 1175, + "Ġfirst": 1176, + "');Ċ": 1177, + "word": 1178, + "Ġimport": 1179, + "Ġact": 1180, + "Ġchar": 1181, + "CT": 1182, + "ĠTr": 1183, + "ople": 1184, + "={": 1185, + "ĉf": 1186, + "24": 1187, + "ient": 1188, + "cent": 1189, + ".j": 1190, + "lection": 1191, + "))Ċ": 1192, + "Ġonly": 1193, + "Ġprint": 1194, + "mer": 1195, + ".W": 1196, + "ock": 1197, + "Ġ--": 1198, + "Text": 1199, + "Ġop": 1200, + "ank": 1201, + "Ġits": 1202, + "Ġback": 1203, + "[\"": 1204, + "Ġneed": 1205, + "Ġcl": 1206, + "Ġsub": 1207, + "Ġla": 1208, + "((": 1209, + ".\"": 1210, + "Object": 1211, + "Ġstart": 1212, + "file": 1213, + "(self": 1214, + "ner": 1215, + "ey": 1216, + "Ġuser": 1217, + "Ġent": 1218, + "ĠCom": 1219, + "its": 1220, + "ĠCon": 1221, + "ouble": 1222, + "ower": 1223, + "item": 1224, + "very": 1225, + "ĠWe": 1226, + "64": 1227, + "lick": 1228, + "ĠQ": 1229, + "php": 1230, + "ttp": 1231, + "':": 1232, + "ics": 1233, + "Ġunder": 1234, + "Ġ*Ċ": 1235, + ".L": 1236, + ");": 1237, + "ices": 1238, + "Ġreg": 1239, + ")čĊ": 1240, + "ĉpublic": 1241, + "SS": 1242, + "Ġthen": 1243, + "reat": 1244, + "ious": 1245, + ".G": 1246, + "ek": 1247, + "irect": 1248, + "heck": 1249, + "cript": 1250, + "ning": 1251, + "ĠUn": 1252, + "Ġmay": 1253, + "ĠWh": 1254, + "Bo": 1255, + "Item": 1256, + "struct": 1257, + ".st": 1258, + "ream": 1259, + "ible": 1260, + "loat": 1261, + "Ġorg": 1262, + "und": 1263, + "sum": 1264, + "_in": 1265, + "../": 1266, + "_M": 1267, + "Ġhow": 1268, + "rite": 1269, + "'Ċ": 1270, + "To": 1271, + "40": 1272, + "ww": 1273, + "Ġpeople": 1274, + "index": 1275, + ".n": 1276, + "http": 1277, + "(m": 1278, + "ector": 1279, + "Ġind": 1280, + "Ġjav": 1281, + "],Ċ": 1282, + "ĠHe": 1283, + "_st": 1284, + "ful": 1285, + "ole": 1286, + "){Ċ": 1287, + "Ġshould": 1288, + "opy": 1289, + "elp": 1290, + "ier": 1291, + "_name": 1292, + "erson": 1293, + "ION": 1294, + "ote": 1295, + "Ġtest": 1296, + "Ġbet": 1297, + "rror": 1298, + "ular": 1299, + "ãĢ": 1300, + "ĠÐ": 1301, + "bs": 1302, + "ting": 1303, + "Ġmake": 1304, + "Tr": 1305, + "Ġafter": 1306, + "arget": 1307, + "RO": 1308, + "olumn": 1309, + "rc": 1310, + "_re": 1311, + "define": 1312, + "22": 1313, + "Ġright": 1314, + "right": 1315, + "day": 1316, + "Ġlong": 1317, + "[]": 1318, + "(p": 1319, + "td": 1320, + "cond": 1321, + "ĠPro": 1322, + "Ġrem": 1323, + "ptions": 1324, + "vid": 1325, + ".g": 1326, + "Ġext": 1327, + "Ġ__": 1328, + "')Ċ": 1329, + "pace": 1330, + "mp": 1331, + "Ġmin": 1332, + "stance": 1333, + "air": 1334, + "action": 1335, + "wh": 1336, + "type": 1337, + "util": 1338, + "ait": 1339, + "": 1340, + "IC": 1341, + "text": 1342, + "Ġph": 1343, + "Ġfl": 1344, + ".M": 1345, + "ccess": 1346, + "br": 1347, + "fore": 1348, + "ersion": 1349, + "),Ċ": 1350, + ".re": 1351, + "ateg": 1352, + "Ġloc": 1353, + "ins": 1354, + "-s": 1355, + "trib": 1356, + "ĠInt": 1357, + "Ġarray": 1358, + ",\"": 1359, + "Pro": 1360, + "(c": 1361, + "ession": 1362, + ">ĊĊ": 1363, + "Ġshe": 1364, + "\"]": 1365, + "aph": 1366, + "Ġexp": 1367, + "erty": 1368, + "ĠSe": 1369, + "Ġpar": 1370, + "unc": 1371, + "ET": 1372, + "Ġread": 1373, + "print": 1374, + "Ġrel": 1375, + "Ġform": 1376, + "Ġdr": 1377, + "Exception": 1378, + "input": 1379, + "Ġtrans": 1380, + "########": 1381, + "order": 1382, + "By": 1383, + "Ġaw": 1384, + "ities": 1385, + "uff": 1386, + "play": 1387, + ".add": 1388, + "ĠâĢĵ": 1389, + "Ġwant": 1390, + "Ġcomp": 1391, + "ments": 1392, + "Ġ||": 1393, + "az": 1394, + "be": 1395, + "Ġnumber": 1396, + "Ġrequire": 1397, + "ĠEx": 1398, + "60": 1399, + "Ġcol": 1400, + "Ġkey": 1401, + "ember": 1402, + "Ġtwo": 1403, + "Ġsize": 1404, + "Ġwhere": 1405, + "UT": 1406, + "result": 1407, + "ĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠ": 1408, + "ough": 1409, + "orld": 1410, + "ood": 1411, + "uch": 1412, + "ative": 1413, + "ger": 1414, + "arent": 1415, + "Ġ/*": 1416, + "Ġarg": 1417, + "Ġwhile": 1418, + "23": 1419, + "(this": 1420, + "Ġrec": 1421, + "Ġdif": 1422, + "State": 1423, + "Ġspec": 1424, + "ride": 1425, + "_F": 1426, + "Ġlook": 1427, + "AM": 1428, + "ility": 1429, + "eter": 1430, + "âĢĻt": 1431, + "ĊĊĊ": 1432, + "ayout": 1433, + "--------------------------------": 1434, + "ager": 1435, + "Ġcould": 1436, + "Ġbr": 1437, + "ends": 1438, + "ures": 1439, + "Ġknow": 1440, + "ets": 1441, + "ĠIf": 1442, + "ĠSh": 1443, + ".w": 1444, + "back": 1445, + "Ġser": 1446, + "Ġ+=": 1447, + "Ġfr": 1448, + "());Ċ": 1449, + "Ġhand": 1450, + "Ind": 1451, + "ULL": 1452, + "Im": 1453, + "();ĊĊ": 1454, + "Ġmost": 1455, + "Ġtry": 1456, + "Ġnow": 1457, + "rough": 1458, + ">čĊ": 1459, + "ackage": 1460, + "Ġhim": 1461, + "._": 1462, + "ify": 1463, + "Ġbreak": 1464, + "Ġ);Ċ": 1465, + "ren": 1466, + "#define": 1467, + "itt": 1468, + "Ġap": 1469, + "ĉc": 1470, + "(n": 1471, + "ĠYou": 1472, + ":ĊĊ": 1473, + "-m": 1474, + "Ġevery": 1475, + "ustom": 1476, + "lient": 1477, + "ocument": 1478, + "cription": 1479, + "Error": 1480, + "-b": 1481, + "о": 1482, + "][": 1483, + "99": 1484, + "trans": 1485, + "Ġpoint": 1486, + "Ġstd": 1487, + "Ġfil": 1488, + "Time": 1489, + "80": 1490, + "Ġmod": 1491, + "Ġ->": 1492, + "Ġerror": 1493, + "ah": 1494, + "Ġtext": 1495, + "roller": 1496, + "lose": 1497, + "ql": 1498, + "Ġpol": 1499, + ">": 1500, + "Ġshow": 1501, + "User": 1502, + "ased": 1503, + "Ġ{ĊĊ": 1504, + "Ġfind": 1505, + "а": 1506, + "ED": 1507, + "span": 1508, + "enu": 1509, + "Ġcurrent": 1510, + "Ġused": 1511, + "cept": 1512, + "clud": 1513, + "Ġplay": 1514, + "Ġlog": 1515, + "ution": 1516, + "fl": 1517, + "Ġsee": 1518, + "indows": 1519, + "Ġhelp": 1520, + "Ġthese": 1521, + "Ġpass": 1522, + "Ġdown": 1523, + "Ġeven": 1524, + "ason": 1525, + "uild": 1526, + "from": 1527, + "(d": 1528, + "Ġbl": 1529, + "label": 1530, + "else": 1531, + "е": 1532, + "Ġ(!": 1533, + "ized": 1534, + "(),": 1535, + "Ġob": 1536, + "Ġitem": 1537, + "ump": 1538, + "UR": 1539, + "orn": 1540, + "Ġdon": 1541, + "Se": 1542, + "man": 1543, + "27": 1544, + "ample": 1545, + "tn": 1546, + "================": 1547, + "He": 1548, + "gram": 1549, + "Ġdid": 1550, + "wn": 1551, + "_h": 1552, + "iver": 1553, + "Ġsm": 1554, + "Ġthrough": 1555, + "ĠAn": 1556, + "che": 1557, + "Ġinv": 1558, + "ouse": 1559, + "Ġes": 1560, + "ĠNew": 1561, + "export": 1562, + "mary": 1563, + "uto": 1564, + "ler": 1565, + "Ġlast": 1566, + "Ġevent": 1567, + "try": 1568, + "ï¼": 1569, + "ily": 1570, + "igned": 1571, + "ines": 1572, + "ollow": 1573, + "icense": 1574, + "sole": 1575, + "lear": 1576, + "(int": 1577, + "Ġagain": 1578, + "Ġhigh": 1579, + "html": 1580, + "Index": 1581, + "uthor": 1582, + "Ġ/**Ċ": 1583, + "Ġline": 1584, + "Event": 1585, + "_D": 1586, + "Ġdoes": 1587, + "itial": 1588, + "Ġcr": 1589, + "ars": 1590, + "28": 1591, + "Ġtem": 1592, + "cause": 1593, + "face": 1594, + "Ġ`": 1595, + "_A": 1596, + "Button": 1597, + "ature": 1598, + "ected": 1599, + "ES": 1600, + "ister": 1601, + "ĉĊ": 1602, + "Ġbefore": 1603, + "ale": 1604, + "other": 1605, + "Ġbecause": 1606, + "roid": 1607, + "Ġed": 1608, + "ik": 1609, + "reg": 1610, + "ĠDe": 1611, + "Ġdist": 1612, + "},Ċ": 1613, + "Ġstate": 1614, + "Ġcons": 1615, + "rint": 1616, + "att": 1617, + "Ġhere": 1618, + "ined": 1619, + "Ġfinal": 1620, + "Ġ\"\"": 1621, + "Key": 1622, + "LO": 1623, + "Ġdel": 1624, + "pty": 1625, + "thing": 1626, + "26": 1627, + "ĠAnd": 1628, + "Ġrun": 1629, + "ĠX": 1630, + "ym": 1631, + ".app": 1632, + "Ġvery": 1633, + "ces": 1634, + "_N": 1635, + "ared": 1636, + "ward": 1637, + "list": 1638, + "ited": 1639, + "olog": 1640, + "itch": 1641, + "Box": 1642, + "ife": 1643, + "33": 1644, + "Ġac": 1645, + "Ġmodel": 1646, + "Ġmon": 1647, + "Ġway": 1648, + "lete": 1649, + "Ġcall": 1650, + "Ġatt": 1651, + "Ġcal": 1652, + "vert": 1653, + "Ġdec": 1654, + "lease": 1655, + "oun": 1656, + "Ġ});Ċ": 1657, + "fr": 1658, + "formation": 1659, + "etail": 1660, + "Ġnum": 1661, + "aj": 1662, + "query": 1663, + "Ġwell": 1664, + "Ġobject": 1665, + "ĠAs": 1666, + "Ġyears": 1667, + "Color": 1668, + "IS": 1669, + "Ġdefault": 1670, + "Wh": 1671, + "Ġins": 1672, + "aint": 1673, + "Ġjava": 1674, + "Ġsim": 1675, + "ĠAr": 1676, + "mon": 1677, + "til": 1678, + "();čĊ": 1679, + "):": 1680, + "Set": 1681, + "29": 1682, + "atter": 1683, + "Ġview": 1684, + "Ġpres": 1685, + "array": 1686, + "We": 1687, + "At": 1688, + "Ġbel": 1689, + "Ġmany": 1690, + "21": 1691, + "Man": 1692, + "ender": 1693, + "Ġbeing": 1694, + "Ġgood": 1695, + "ĉĉĉĉĉĉ": 1696, + "ational": 1697, + "ware": 1698, + ".log": 1699, + "{čĊ": 1700, + "Ġusing": 1701, + "_B": 1702, + "Ġ:=": 1703, + "_w": 1704, + "ists": 1705, + "lish": 1706, + "Ġstud": 1707, + "ĠAl": 1708, + "Ġgu": 1709, + "config": 1710, + "uring": 1711, + "time": 1712, + "oken": 1713, + "amespace": 1714, + "Ġrequest": 1715, + "Ġchild": 1716, + "ĠÃ": 1717, + "lob": 1718, + "Ġparam": 1719, + "Ġ}čĊ": 1720, + "01": 1721, + "Ġecho": 1722, + "function": 1723, + "********************************": 1724, + "ps": 1725, + "Element": 1726, + "alk": 1727, + "lication": 1728, + "by": 1729, + "Size": 1730, + "rawing": 1731, + "Ġperson": 1732, + "ĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠ": 1733, + "\\n": 1734, + "object": 1735, + "ince": 1736, + "En": 1737, + "File": 1738, + "uf": 1739, + "ffect": 1740, + "AC": 1741, + "Ġstyle": 1742, + "summary": 1743, + "Ġque": 1744, + "_r": 1745, + "Ġ($": 1746, + "Model": 1747, + "ident": 1748, + "Ġmethod": 1749, + "IL": 1750, + "ott": 1751, + "less": 1752, + "ING": 1753, + "Ġ()": 1754, + "Ġexpect": 1755, + "ync": 1756, + "package": 1757, + "35": 1758, + "urs": 1759, + "Ġprot": 1760, + "./": 1761, + "pre": 1762, + "Ġ)Ċ": 1763, + "ma": 1764, + "Ġsur": 1765, + "Ġfound": 1766, + "Info": 1767, + "par": 1768, + "imes": 1769, + ".e": 1770, + "ains": 1771, + "Ġpost": 1772, + "-d": 1773, + "45": 1774, + "olean": 1775, + "Ġsl": 1776, + "PE": 1777, + "Ġsuch": 1778, + "select": 1779, + "ainer": 1780, + "Ġthink": 1781, + "Ġdiffer": 1782, + ".r": 1783, + "/**Ċ": 1784, + "FF": 1785, + "ool": 1786, + "plate": 1787, + "qual": 1788, + "ĠFor": 1789, + "Ġmuch": 1790, + "uc": 1791, + "(new": 1792, + "odule": 1793, + "Ġsom": 1794, + "Ġhttp": 1795, + "ĠList": 1796, + "Ġcount": 1797, + "Ġinst": 1798, + "char": 1799, + "mit": 1800, + ".id": 1801, + "aking": 1802, + "Ġgener": 1803, + "px": 1804, + "vice": 1805, + "37": 1806, + "_data": 1807, + "ĠNULL": 1808, + "}čĊ": 1809, + "idd": 1810, + "ãĢĤ": 1811, + "Ġmed": 1812, + "org": 1813, + "ider": 1814, + "ache": 1815, + "work": 1816, + "Ġcheck": 1817, + "ween": 1818, + "Ġ((": 1819, + "the": 1820, + "ants": 1821, + "><": 1822, + ".B": 1823, + "-c": 1824, + "Ġopen": 1825, + "Ġest": 1826, + "ĠĠĠĠĠĠĠĠĊ": 1827, + "Ġnext": 1828, + "IM": 1829, + "ÑĤ": 1830, + "OT": 1831, + "ó": 1832, + "Ġfollow": 1833, + "content": 1834, + "ĠĠĠĠĠĠĠĠĠĠĠĠ": 1835, + "Ġinclud": 1836, + "HE": 1837, + "ĠRes": 1838, + "Ġhref": 1839, + "и": 1840, + "Ġcar": 1841, + "ypes": 1842, + "image": 1843, + "Un": 1844, + "Ġbool": 1845, + "AD": 1846, + "Ġgame": 1847, + ".Form": 1848, + "rows": 1849, + "*/": 1850, + "velop": 1851, + ".Drawing": 1852, + "Ġpath": 1853, + "ision": 1854, + "Ġeach": 1855, + "ĠPl": 1856, + "_type": 1857, + "Path": 1858, + "nection": 1859, + "Ġav": 1860, + "').": 1861, + "Ġsupport": 1862, + "ENT": 1863, + "rem": 1864, + "\").": 1865, + "Ġown": 1866, + "Ġcor": 1867, + "count": 1868, + "miss": 1869, + "ually": 1870, + "Ġmem": 1871, + "std": 1872, + "ience": 1873, + "search": 1874, + "\"ĊĊ": 1875, + "Form": 1876, + "Ġsex": 1877, + "ename": 1878, + "Ġsign": 1879, + "Ġet": 1880, + "ĠĠĠĠĠĠĠĠĠĠ": 1881, + "','": 1882, + "ĠApp": 1883, + "Ġthose": 1884, + "off": 1885, + "Ġerr": 1886, + "Ġsystem": 1887, + "Ġbest": 1888, + "code": 1889, + "Ġsame": 1890, + "Ġdi": 1891, + "uss": 1892, + "Ġcreate": 1893, + "ather": 1894, + "Array": 1895, + ".in": 1896, + "fe": 1897, + "Service": 1898, + "UN": 1899, + "ats": 1900, + "ĠZ": 1901, + "alth": 1902, + "Ġmade": 1903, + "true": 1904, + "AB": 1905, + "Ġmark": 1906, + "rid": 1907, + "ified": 1908, + ",čĊ": 1909, + "yn": 1910, + "press": 1911, + "Ġgroup": 1912, + "Ġfin": 1913, + "ĠLicense": 1914, + "Field": 1915, + "eger": 1916, + "Ġworld": 1917, + "iness": 1918, + "ty": 1919, + "Ġprocess": 1920, + "(b": 1921, + "Ġcre": 1922, + "arn": 1923, + "ives": 1924, + "Ġmain": 1925, + "ideo": 1926, + "36": 1927, + "_g": 1928, + "AG": 1929, + "valid": 1930, + "img": 1931, + "PI": 1932, + "Ġcolor": 1933, + "Ġreport": 1934, + "Ġtake": 1935, + "rib": 1936, + "OM": 1937, + "Ġday": 1938, + "Request": 1939, + "Ġsk": 1940, + "bers": 1941, + "ĉs": 1942, + ".Add": 1943, + "oot": 1944, + "Image": 1945, + "Ġcomple": 1946, + "ollection": 1947, + "Ġtop": 1948, + "Ġfree": 1949, + "AS": 1950, + "De": 1951, + "ĠOn": 1952, + "IG": 1953, + "90": 1954, + "eta": 1955, + "Date": 1956, + "Ġaction": 1957, + "34": 1958, + "Over": 1959, + "itor": 1960, + "ĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠ": 1961, + "not": 1962, + "Ġindex": 1963, + "her": 1964, + "icon": 1965, + "On": 1966, + ";čĊčĊ": 1967, + "ivity": 1968, + "mand": 1969, + ".Windows": 1970, + "OL": 1971, + "Ġreal": 1972, + "Ġmax": 1973, + "land": 1974, + "....": 1975, + "raph": 1976, + "Ġbuild": 1977, + "leg": 1978, + "assword": 1979, + "?ĊĊ": 1980, + "âĢ¦": 1981, + "ook": 1982, + "uck": 1983, + "Ġmessage": 1984, + "test": 1985, + "ivers": 1986, + "38": 1987, + "Ġinput": 1988, + "Ġart": 1989, + "Ġbetween": 1990, + "Get": 1991, + "enter": 1992, + "ground": 1993, + "ene": 1994, + "á": 1995, + ".length": 1996, + "Node": 1997, + "(i": 1998, + "Class": 1999, + "for": 2000, + "ĠâĢĶ": 2001, + "ten": 2002, + "oin": 2003, + "Ġke": 2004, + "ui": 2005, + "ĠIN": 2006, + "Ġtable": 2007, + "sub": 2008, + "ĠLe": 2009, + "Ġhead": 2010, + "Ġmust": 2011, + "////////////////": 2012, + ".util": 2013, + "Context": 2014, + "Ġorder": 2015, + "Ġmov": 2016, + "over": 2017, + "Ġcontin": 2018, + "Ġsay": 2019, + "static": 2020, + ".Text": 2021, + "ĠclassName": 2022, + "pany": 2023, + "Ġter": 2024, + "head": 2025, + "rg": 2026, + "Ġproduct": 2027, + "This": 2028, + ".âĢĿ": 2029, + "ĠBut": 2030, + "70": 2031, + "loy": 2032, + "Ġdouble": 2033, + "sg": 2034, + "Ġplace": 2035, + ".x": 2036, + "message": 2037, + "Ġinformation": 2038, + "private": 2039, + "Ġoper": 2040, + "ced": 2041, + "db": 2042, + "\">": 2043, + "Param": 2044, + "icle": 2045, + "Ġweek": 2046, + "Ġprop": 2047, + "table": 2048, + "idget": 2049, + "place": 2050, + "Prop": 2051, + "ĠAll": 2052, + "els": 2053, + "box": 2054, + ".ĊĊĊĊ": 2055, + ".R": 2056, + "ĠTo": 2057, + "iter": 2058, + "Sh": 2059, + "uration": 2060, + "older": 2061, + "_list": 2062, + "come": 2063, + "Ġsw": 2064, + "ization": 2065, + "ĉfor": 2066, + "bl": 2067, + "Ġprogram": 2068, + "(e": 2069, + "ape": 2070, + "check": 2071, + ".Forms": 2072, + "Ġund": 2073, + "ategory": 2074, + "75": 2075, + "ags": 2076, + "Ġresponse": 2077, + "US": 2078, + "request": 2079, + "Ġstruct": 2080, + "escription": 2081, + "Ġcode": 2082, + "_H": 2083, + "uffer": 2084, + "Ġwithout": 2085, + "lobal": 2086, + "Manager": 2087, + "ilter": 2088, + "PO": 2089, + "ĉthis": 2090, + "option": 2091, + "Ġsol": 2092, + "Ġ===": 2093, + "akes": 2094, + "Controller": 2095, + "44": 2096, + "Message": 2097, + "Ġref": 2098, + "ever": 2099, + "ĠSo": 2100, + "aining": 2101, + ".append": 2102, + "Ġstill": 2103, + "Ġprovid": 2104, + "Ġassert": 2105, + "med": 2106, + "Ġcap": 2107, + "usiness": 2108, + "Ġrep": 2109, + "tings": 2110, + "ved": 2111, + ".N": 2112, + "api": 2113, + "OD": 2114, + "Ġfield": 2115, + "iven": 2116, + "oto": 2117, + "âĢľ": 2118, + "col": 2119, + "(x": 2120, + "ght": 2121, + "Result": 2122, + "Code": 2123, + ".is": 2124, + "link": 2125, + "Ġcour": 2126, + "An": 2127, + "Ġteam": 2128, + "ĉint": 2129, + "ift": 2130, + "55": 2131, + "Ġsecond": 2132, + "Ġgoing": 2133, + "Ġrange": 2134, + "_E": 2135, + "ness": 2136, + "39": 2137, + "Ġfam": 2138, + "Ġnil": 2139, + "ĠCont": 2140, + "ailable": 2141, + "utes": 2142, + "atab": 2143, + "Ġfact": 2144, + "Ġvis": 2145, + "(&": 2146, + "ĠAN": 2147, + "31": 2148, + "Al": 2149, + "title": 2150, + "Ġandroid": 2151, + "CE": 2152, + "\\\"": 2153, + "irt": 2154, + "Ġwrit": 2155, + "н": 2156, + "ĉm": 2157, + "ftware": 2158, + "ond": 2159, + "Ġret": 2160, + "osition": 2161, + "Ġhome": 2162, + "Ġleft": 2163, + "args": 2164, + "meric": 2165, + "48": 2166, + "Ġdirect": 2167, + "oci": 2168, + "Pl": 2169, + "As": 2170, + "ret": 2171, + "ado": 2172, + "Of": 2173, + "chn": 2174, + "ĠGet": 2175, + "ee": 2176, + "ross": 2177, + "();": 2178, + "____": 2179, + ".ph": 2180, + "It": 2181, + "oute": 2182, + "Ġexper": 2183, + "chool": 2184, + "www": 2185, + "},": 2186, + "Ġallow": 2187, + "ĠÂ": 2188, + "())": 2189, + "size": 2190, + "ism": 2191, + "ai": 2192, + "tract": 2193, + "ane": 2194, + "...ĊĊ": 2195, + "context": 2196, + "Ġbeg": 2197, + "CH": 2198, + "Ġpage": 2199, + "hip": 2200, + "no": 2201, + "core": 2202, + "sp": 2203, + "Ġdifferent": 2204, + "iable": 2205, + "ĠMe": 2206, + "_IN": 2207, + "button": 2208, + "ĠIs": 2209, + "ervices": 2210, + "Ġca": 2211, + "Ġaround": 2212, + "App": 2213, + "ration": 2214, + "Ġrece": 2215, + "Ġreally": 2216, + "Ġimage": 2217, + "Ġtarget": 2218, + "Ġdep": 2219, + "opyright": 2220, + "tra": 2221, + "ingle": 2222, + "ital": 2223, + "Layout": 2224, + "Ġboth": 2225, + "Override": 2226, + "arm": 2227, + "=>": 2228, + "aterial": 2229, + "iled": 2230, + "Ġput": 2231, + "Qu": 2232, + "ÑĢ": 2233, + "ung": 2234, + "map": 2235, + "ĉĉĉĉĉĉĉĉ": 2236, + "Ġlevel": 2237, + "Component": 2238, + "book": 2239, + "creen": 2240, + "_RE": 2241, + "Ġconfig": 2242, + "ãģ": 2243, + "Or": 2244, + ".data": 2245, + "Ġdocument": 2246, + "\",\"": 2247, + "tribute": 2248, + "ux": 2249, + "Log": 2250, + "ference": 2251, + "post": 2252, + "_e": 2253, + "Ġlocal": 2254, + "andom": 2255, + "assert": 2256, + "Val": 2257, + "lected": 2258, + "ina": 2259, + "atabase": 2260, + "Add": 2261, + "Ġcontent": 2262, + ".print": 2263, + "signed": 2264, + "ric": 2265, + ".\"ĊĊ": 2266, + "Ġfa": 2267, + "!ĊĊ": 2268, + "-f": 2269, + "ived": 2270, + "Ġquest": 2271, + ".ex": 2272, + "Ġfloat": 2273, + "Ġdevelop": 2274, + "оÐ": 2275, + "Map": 2276, + "ading": 2277, + "Ġposs": 2278, + "UE": 2279, + "namespace": 2280, + "_O": 2281, + "ĉb": 2282, + ".Get": 2283, + ">(": 2284, + "json": 2285, + "etails": 2286, + "66": 2287, + "Ġtoo": 2288, + "Ġextends": 2289, + "ĠNone": 2290, + "Ġfore": 2291, + "(String": 2292, + "format": 2293, + "Ġgreat": 2294, + "inter": 2295, + "cale": 2296, + "Ñģ": 2297, + "ron": 2298, + "iving": 2299, + "Ent": 2300, + "ency": 2301, + "xt": 2302, + "oy": 2303, + "05": 2304, + "Ġmonth": 2305, + "Ġhapp": 2306, + "Ġsuper": 2307, + "bar": 2308, + "default": 2309, + "_de": 2310, + "ords": 2311, + "ln": 2312, + "({Ċ": 2313, + "ĠInd": 2314, + "ases": 2315, + "Ġtitle": 2316, + "Ġcontext": 2317, + "08": 2318, + "oh": 2319, + "-p": 2320, + "Em": 2321, + "Ġmet": 2322, + "Test": 2323, + "Ġlife": 2324, + "_v": 2325, + "ĠUS": 2326, + "UI": 2327, + "ocation": 2328, + "md": 2329, + "Ġ[Ċ": 2330, + "Ġ]": 2331, + "sw": 2332, + "Ġincre": 2333, + "script": 2334, + "ential": 2335, + "ways": 2336, + ".de": 2337, + "Ġsrc": 2338, + "Ġcatch": 2339, + "ĠAmeric": 2340, + "//Ċ": 2341, + "ĠĠĠĠĠĠĠĠĠĠĠĠĠĠ": 2342, + "Ġpay": 2343, + "plit": 2344, + "âĢĶ": 2345, + "Ġcoun": 2346, + "obj": 2347, + ".php": 2348, + "Ġchange": 2349, + "ething": 2350, + "'re": 2351, + "aster": 2352, + "los": 2353, + "lation": 2354, + "ĠĠĊ": 2355, + "Le": 2356, + "ä": 2357, + "({": 2358, + "ready": 2359, + "ĠNo": 2360, + "Ġposition": 2361, + "Ġold": 2362, + "Ġbook": 2363, + "abled": 2364, + "bug": 2365, + "202": 2366, + "Hand": 2367, + "};ĊĊ": 2368, + "isplay": 2369, + "aving": 2370, + "04": 2371, + "Ġgover": 2372, + "Ġversion": 2373, + "System": 2374, + "nect": 2375, + "response": 2376, + "Style": 2377, + "Up": 2378, + "angu": 2379, + "Ġthree": 2380, + "init": 2381, + "ero": 2382, + "Ġlaw": 2383, + "endif": 2384, + "Ġbase": 2385, + "email": 2386, + "(l": 2387, + "_V": 2388, + "Ġconf": 2389, + "ATE": 2390, + "Ġduring": 2391, + "tes": 2392, + "Ġconsole": 2393, + "ĠPr": 2394, + "Ġspe": 2395, + "ves": 2396, + "65": 2397, + "path": 2398, + "ialog": 2399, + "dition": 2400, + "_to": 2401, + "ards": 2402, + "Ġagainst": 2403, + "etwork": 2404, + "ĠPh": 2405, + "_L": 2406, + "cur": 2407, + "imit": 2408, + "With": 2409, + "Ġpower": 2410, + "ium": 2411, + "';ĊĊ": 2412, + "Ġwom": 2413, + "left": 2414, + "ources": 2415, + "atri": 2416, + "ĠIm": 2417, + "ĠMan": 2418, + "orth": 2419, + "${": 2420, + "88": 2421, + "quals": 2422, + "ese": 2423, + "_size": 2424, + "Ġiss": 2425, + "otal": 2426, + "-g": 2427, + "ique": 2428, + "rame": 2429, + "Ġwidth": 2430, + "erg": 2431, + ")(": 2432, + "ittle": 2433, + "TR": 2434, + "ĠThey": 2435, + "ences": 2436, + "02": 2437, + "rl": 2438, + "ons": 2439, + "Ġlabel": 2440, + ".y": 2441, + "-t": 2442, + "update": 2443, + "anel": 2444, + "sc": 2445, + ".to": 2446, + "Ġproject": 2447, + "ü": 2448, + "Ġelement": 2449, + "Ġsuccess": 2450, + "ĉĉĊ": 2451, + ".sh": 2452, + "ram": 2453, + "ched": 2454, + "())Ċ": 2455, + "Ġ(Ċ": 2456, + "Ġdate": 2457, + "Ġtot": 2458, + "_ST": 2459, + "All": 2460, + "ification": 2461, + "ĉvar": 2462, + "Ġtri": 2463, + "chem": 2464, + "my": 2465, + "Ġbig": 2466, + "ĠAd": 2467, + "ĠAt": 2468, + "ots": 2469, + "num": 2470, + "Act": 2471, + "Ġmap": 2472, + "era": 2473, + "cope": 2474, + ".$": 2475, + ",âĢĿ": 2476, + "Ġpop": 2477, + "Ġfew": 2478, + "Ġlen": 2479, + "uid": 2480, + "eters": 2481, + "ules": 2482, + "ÃŃ": 2483, + "source": 2484, + "https": 2485, + "Ġdem": 2486, + "Ġear": 2487, + "################": 2488, + "Ġmatch": 2489, + "ories": 2490, + "49": 2491, + "aces": 2492, + "ĠCl": 2493, + "Ġnode": 2494, + "78": 2495, + "irc": 2496, + "local": 2497, + "unity": 2498, + "};Ċ": 2499, + "Ġanother": 2500, + "<<": 2501, + "ogle": 2502, + "Ġsit": 2503, + "ework": 2504, + "TE": 2505, + ".I": 2506, + "NS": 2507, + "ology": 2508, + "ought": 2509, + ".Cont": 2510, + ">>": 2511, + "Ġcare": 2512, + "state": 2513, + "ĉprivate": 2514, + "Ġeffect": 2515, + "++)": 2516, + "_file": 2517, + "ending": 2518, + "Line": 2519, + "For": 2520, + "ior": 2521, + "ĠSc": 2522, + "Ġfun": 2523, + ".Size": 2524, + "ĉelse": 2525, + "])": 2526, + "start": 2527, + "vious": 2528, + "Ġ},": 2529, + "ours": 2530, + "Ġleg": 2531, + "Ġservice": 2532, + "Ġsince": 2533, + "iron": 2534, + "Label": 2535, + "Ġnon": 2536, + "Ġlos": 2537, + "iction": 2538, + "Ġfull": 2539, + "acter": 2540, + "board": 2541, + "gress": 2542, + "Ġturn": 2543, + "ither": 2544, + "09": 2545, + ".size": 2546, + "Ġbody": 2547, + "resh": 2548, + "eturn": 2549, + "199": 2550, + "(_": 2551, + "yles": 2552, + "ormal": 2553, + "pi": 2554, + "Ġsomething": 2555, + "!--": 2556, + "uint": 2557, + "Ġprodu": 2558, + "Ġstand": 2559, + "Ġproble": 2560, + "Ġavailable": 2561, + "mt": 2562, + "ĠBl": 2563, + "Ġ...": 2564, + "Ġblock": 2565, + "Input": 2566, + "Ġkeep": 2567, + "Count": 2568, + "open": 2569, + "Ġ['": 2570, + "Ġthrow": 2571, + "uilder": 2572, + "Action": 2573, + "Ġthings": 2574, + "True": 2575, + "Ġurl": 2576, + "ĠBo": 2577, + "printf": 2578, + "Ġred": 2579, + "js": 2580, + ".create": 2581, + "ĠOr": 2582, + "Status": 2583, + "Instance": 2584, + "Ġcontrol": 2585, + "Ġcome": 2586, + "Ġcustom": 2587, + "location": 2588, + "07": 2589, + "model": 2590, + "ĠčĊ": 2591, + "Ġsource": 2592, + "Ġeas": 2593, + ".out": 2594, + "]ĊĊ": 2595, + "oney": 2596, + "Ġawait": 2597, + "Ġpartic": 2598, + "AP": 2599, + "ublish": 2600, + "odes": 2601, + "_pro": 2602, + "ply": 2603, + "riter": 2604, + "Ġprov": 2605, + "Ġmill": 2606, + "HT": 2607, + "])Ċ": 2608, + "Ġchang": 2609, + "Ġask": 2610, + "ĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠ": 2611, + "Ġoutput": 2612, + "Ġemail": 2613, + "68": 2614, + ".push": 2615, + "Ġ}čĊčĊ": 2616, + "ination": 2617, + "47": 2618, + "atrix": 2619, + "Table": 2620, + "uccess": 2621, + "]);Ċ": 2622, + "ĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠ": 2623, + "Ġdisc": 2624, + "([": 2625, + "Ġbusiness": 2626, + "height": 2627, + ".html": 2628, + "ta": 2629, + "field": 2630, + "Ġrequired": 2631, + "_R": 2632, + "Ġgovern": 2633, + "}čĊčĊ": 2634, + "lex": 2635, + "500": 2636, + ".,": 2637, + "ĠSet": 2638, + "urch": 2639, + "///": 2640, + "ts": 2641, + "af": 2642, + "Ġmight": 2643, + "istory": 2644, + "Str": 2645, + "Ġnever": 2646, + "Response": 2647, + "arse": 2648, + "ada": 2649, + "ĠHow": 2650, + "Ġ*)": 2651, + "Ġ;": 2652, + "Ġhard": 2653, + "Ad": 2654, + "Ġintern": 2655, + "used": 2656, + "(data": 2657, + "mod": 2658, + "annel": 2659, + "Ġnp": 2660, + "ugg": 2661, + "Ġ/>Ċ": 2662, + "Ġcalled": 2663, + "body": 2664, + "Ġcho": 2665, + "(r": 2666, + "_set": 2667, + "ird": 2668, + "Ġ>=": 2669, + "Ġ};Ċ": 2670, + "Ġoptions": 2671, + "ĠGener": 2672, + "Ġheight": 2673, + "Point": 2674, + "You": 2675, + "ety": 2676, + "Click": 2677, + "Ġsmall": 2678, + "Ġide": 2679, + "Ġaccess": 2680, + "anguage": 2681, + "Ġprotected": 2682, + "Ġjob": 2683, + "ĠThere": 2684, + "Def": 2685, + "Ġaddress": 2686, + "Ġuint": 2687, + "Not": 2688, + "oo": 2689, + "aps": 2690, + "
&": 5909, + "CON": 5910, + "Ġrepl": 5911, + "Ġregular": 5912, + "Storage": 5913, + "ramework": 5914, + "Ġgoal": 5915, + "Ġtouch": 5916, + ".widget": 5917, + "Ġbuilt": 5918, + "des": 5919, + "Part": 5920, + "(re": 5921, + "Ġworth": 5922, + "hib": 5923, + "game": 5924, + "91": 5925, + "192": 5926, + "Ġв": 5927, + "acion": 5928, + "ĠWhite": 5929, + "(type": 5930, + "(`": 5931, + "81": 5932, + "Ġnatural": 5933, + "Ġinj": 5934, + "Ġcalcul": 5935, + "ĠApril": 5936, + ".List": 5937, + "Ġassociated": 5938, + "ĉSystem": 5939, + "~~": 5940, + "=[": 5941, + "Ġstorage": 5942, + "Ġbytes": 5943, + "Ġtravel": 5944, + "Ġsou": 5945, + "Ġpassed": 5946, + "!=": 5947, + "ascript": 5948, + ".open": 5949, + "Ġgrid": 5950, + "Ġbus": 5951, + "Ġrecogn": 5952, + "Ab": 5953, + "Ġhon": 5954, + "ĠCenter": 5955, + "Ġprec": 5956, + "build": 5957, + "73": 5958, + "HTML": 5959, + "ĠSan": 5960, + "Ġcountries": 5961, + "aled": 5962, + "token": 5963, + "kt": 5964, + "Ġqual": 5965, + "Last": 5966, + "adow": 5967, + "Ġmanufact": 5968, + "idad": 5969, + "jango": 5970, + "Next": 5971, + "xf": 5972, + ".a": 5973, + "Ġporno": 5974, + "ĠPM": 5975, + "erve": 5976, + "iting": 5977, + "_th": 5978, + "ci": 5979, + "=None": 5980, + "gs": 5981, + "Ġlogin": 5982, + "atives": 5983, + "']);Ċ": 5984, + "Äħ": 5985, + "Ġill": 5986, + "IA": 5987, + "children": 5988, + "DO": 5989, + "Ġlevels": 5990, + "Ġ{{": 5991, + "Ġlooks": 5992, + "Ġ\"#": 5993, + "ToString": 5994, + "Ġnecessary": 5995, + "ĠĠĠĊ": 5996, + "cell": 5997, + "Entry": 5998, + "Ġ'#": 5999, + "Ġextrem": 6000, + "Selector": 6001, + "Ġplaceholder": 6002, + "Load": 6003, + "Ġreleased": 6004, + "ORE": 6005, + "Enumer": 6006, + "ĠTV": 6007, + "SET": 6008, + "inq": 6009, + "Press": 6010, + "ĠDepartment": 6011, + "Ġproperties": 6012, + "Ġrespond": 6013, + "Search": 6014, + "ael": 6015, + "Ġrequ": 6016, + "ĠBook": 6017, + "/Ċ": 6018, + "(st": 6019, + "Ġfinancial": 6020, + "icket": 6021, + "_input": 6022, + "Ġthreat": 6023, + "(in": 6024, + "Strip": 6025, + "ìĿ": 6026, + "ção": 6027, + "71": 6028, + "Ġevidence": 6029, + "));": 6030, + "ĠBro": 6031, + "Ġ[];Ċ": 6032, + "Ġou": 6033, + "buf": 6034, + "Script": 6035, + "dat": 6036, + "Ġrule": 6037, + "#import": 6038, + "=\"/": 6039, + "Serial": 6040, + "Ġstarting": 6041, + "[index": 6042, + "ae": 6043, + "Ġcontrib": 6044, + "session": 6045, + "_new": 6046, + "utable": 6047, + "ober": 6048, + "Ġ\"./": 6049, + "Ġlogger": 6050, + "Ġrecently": 6051, + "Ġreturned": 6052, + "ččĊ": 6053, + ")))Ċ": 6054, + "itions": 6055, + "Ġseek": 6056, + "Ġcommunic": 6057, + "Ġ\".": 6058, + "Ġusername": 6059, + "ECT": 6060, + "DS": 6061, + "Ġotherwise": 6062, + "ĠGerman": 6063, + ".aw": 6064, + "Adapter": 6065, + "ixel": 6066, + "Ġsystems": 6067, + "Ġdrop": 6068, + "83": 6069, + "Ġstructure": 6070, + "Ġ$(\"#": 6071, + "encies": 6072, + "anning": 6073, + "ĠLink": 6074, + "ĠResponse": 6075, + "Ġstri": 6076, + "ż": 6077, + "ĠDB": 6078, + "æĹ": 6079, + "android": 6080, + "submit": 6081, + "otion": 6082, + "92": 6083, + "(@": 6084, + ".test": 6085, + "82": 6086, + "ĊĊĊĊĊĊĊĊ": 6087, + "];čĊ": 6088, + "Ġdirectly": 6089, + "Ġ\"%": 6090, + "ris": 6091, + "elta": 6092, + "AIL": 6093, + "){čĊ": 6094, + "mine": 6095, + "ĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠ": 6096, + "(k": 6097, + "bon": 6098, + "asic": 6099, + "pite": 6100, + "___": 6101, + "Max": 6102, + "Ġerrors": 6103, + "ĠWhile": 6104, + "Ġarguments": 6105, + "Ġensure": 6106, + "Right": 6107, + "-based": 6108, + "Web": 6109, + "Ġ-=": 6110, + "Ġintrodu": 6111, + "ĠInst": 6112, + "ĠWash": 6113, + "ordin": 6114, + "join": 6115, + "Database": 6116, + "Ġgrad": 6117, + "Ġusually": 6118, + "ITE": 6119, + "Props": 6120, + "?>Ċ": 6121, + "ĠGo": 6122, + "@Override": 6123, + "REF": 6124, + "Ġip": 6125, + "ĠAustral": 6126, + "Ġist": 6127, + "ViewById": 6128, + "Ġserious": 6129, + "Ġcustomer": 6130, + ".prototype": 6131, + "odo": 6132, + "cor": 6133, + "Ġdoor": 6134, + "ĠWITHOUT": 6135, + "Ġplant": 6136, + "Ġbegan": 6137, + "Ġdistance": 6138, + "()).": 6139, + "Ġchance": 6140, + "Ġord": 6141, + "came": 6142, + "pragma": 6143, + "Ġprotect": 6144, + "ragment": 6145, + "ĠNode": 6146, + "ening": 6147, + "Ñĩ": 6148, + "Ġroute": 6149, + "ĠSchool": 6150, + "hi": 6151, + "Ġneighb": 6152, + "After": 6153, + "licit": 6154, + "Ġcontr": 6155, + "Ġprimary": 6156, + "AA": 6157, + ".WriteLine": 6158, + "utils": 6159, + "Ġbi": 6160, + "Red": 6161, + ".Linq": 6162, + ".object": 6163, + "Ġleaders": 6164, + "unities": 6165, + "Ġgun": 6166, + "onth": 6167, + "ĠDev": 6168, + "FILE": 6169, + "Ġcomments": 6170, + "_len": 6171, + "arrow": 6172, + "amount": 6173, + "Range": 6174, + "sert": 6175, + "GridView": 6176, + "Ġupdated": 6177, + "ĠMo": 6178, + "Ġinform": 6179, + "ociety": 6180, + "ala": 6181, + "Access": 6182, + "Ġhab": 6183, + "Ġcreat": 6184, + "_arg": 6185, + "ĠJanuary": 6186, + "ĠDay": 6187, + "\")čĊ": 6188, + "uple": 6189, + "document": 6190, + "gorith": 6191, + "menu": 6192, + "ĠOver": 6193, + "bb": 6194, + ".title": 6195, + "_out": 6196, + "Ġled": 6197, + "uri": 6198, + "Ġ?>": 6199, + "gl": 6200, + "Ġbank": 6201, + "ayment": 6202, + "ĉprintf": 6203, + "MD": 6204, + "Ġsample": 6205, + "Ġhands": 6206, + "ĠVersion": 6207, + "uario": 6208, + "Ġoffers": 6209, + "ityEngine": 6210, + "Ġshape": 6211, + "Ġsleep": 6212, + "_point": 6213, + "Settings": 6214, + "Ġachie": 6215, + "Ġsold": 6216, + "ota": 6217, + ".bind": 6218, + "Am": 6219, + "Ġsafe": 6220, + "Store": 6221, + "Ġshared": 6222, + "Ġpriv": 6223, + "_VAL": 6224, + "Ġsens": 6225, + "){": 6226, + "Ġremember": 6227, + "shared": 6228, + "element": 6229, + "Ġshoot": 6230, + "Vert": 6231, + "cout": 6232, + "Ġenv": 6233, + "_label": 6234, + "Ġ>Ċ": 6235, + "run": 6236, + "Ġscene": 6237, + "(array": 6238, + "device": 6239, + "_title": 6240, + "agon": 6241, + "]čĊ": 6242, + "aby": 6243, + "Ġbecame": 6244, + "boolean": 6245, + "Ġpark": 6246, + "ĠCode": 6247, + "upload": 6248, + "riday": 6249, + "ĠSeptember": 6250, + "Fe": 6251, + "Ġsen": 6252, + "cing": 6253, + "FL": 6254, + "Col": 6255, + "uts": 6256, + "_page": 6257, + "inn": 6258, + "Ġimplied": 6259, + "aling": 6260, + "Ġyourself": 6261, + ".Count": 6262, + "conf": 6263, + "Ġaud": 6264, + "_init": 6265, + ".)": 6266, + "Ġwrote": 6267, + "003": 6268, + "NG": 6269, + ".Error": 6270, + "ä»": 6271, + ".for": 6272, + "Ġequal": 6273, + "ĠRequest": 6274, + "Ġserial": 6275, + "Ġallows": 6276, + "XX": 6277, + "Ġmiddle": 6278, + "chor": 6279, + "195": 6280, + "94": 6281, + "ø": 6282, + "erval": 6283, + ".Column": 6284, + "reading": 6285, + "Ġescort": 6286, + "ĠAugust": 6287, + "Ġquickly": 6288, + "Ġweap": 6289, + "ĠCG": 6290, + "ropri": 6291, + "ho": 6292, + "Ġcop": 6293, + "(struct": 6294, + "ĠBig": 6295, + "Ġvs": 6296, + "Ġfrequ": 6297, + ".Value": 6298, + "Ġactions": 6299, + "Ġproper": 6300, + "Ġinn": 6301, + "Ġobjects": 6302, + "Ġmatrix": 6303, + "avascript": 6304, + "Ġones": 6305, + ".group": 6306, + "Ġgreen": 6307, + "Ġpaint": 6308, + "ools": 6309, + "ycl": 6310, + "encode": 6311, + "olt": 6312, + "comment": 6313, + ".api": 6314, + "Dir": 6315, + "Ġune": 6316, + "izont": 6317, + ".position": 6318, + "Ġdesigned": 6319, + "_val": 6320, + "avi": 6321, + "iring": 6322, + "tab": 6323, + "Ġlayer": 6324, + "Ġviews": 6325, + "Ġreve": 6326, + "rael": 6327, + "ĠON": 6328, + "rics": 6329, + "160": 6330, + "np": 6331, + "Ġcore": 6332, + "());čĊ": 6333, + "Main": 6334, + "Ġexpert": 6335, + "ĉĉčĊ": 6336, + "_en": 6337, + "Ġ/>": 6338, + "utter": 6339, + "IAL": 6340, + "ails": 6341, + "ĠKing": 6342, + "*/ĊĊ": 6343, + "ĠMet": 6344, + "_end": 6345, + "addr": 6346, + "ora": 6347, + "Ġir": 6348, + "Min": 6349, + "Ġsurpr": 6350, + "Ġrepe": 6351, + "Ġdirectory": 6352, + "PUT": 6353, + "-S": 6354, + "Ġelection": 6355, + "haps": 6356, + ".pre": 6357, + "cm": 6358, + "Values": 6359, + "Ġ\"Ċ": 6360, + "column": 6361, + "ivil": 6362, + "Login": 6363, + "inue": 6364, + "93": 6365, + "Ġbeautiful": 6366, + "Ġsecret": 6367, + "(event": 6368, + "Ġchat": 6369, + "ums": 6370, + "Ġorigin": 6371, + "Ġeffects": 6372, + "Ġmanagement": 6373, + "illa": 6374, + "tk": 6375, + "Ġsetting": 6376, + "ĠCour": 6377, + "Ġmassage": 6378, + "ĉend": 6379, + "Ġhappy": 6380, + "Ġfinish": 6381, + "Ġcamera": 6382, + "ĠVer": 6383, + "ĠDemocr": 6384, + "ĠHer": 6385, + "(Q": 6386, + "cons": 6387, + "ita": 6388, + "Ġ'.": 6389, + "{}": 6390, + "ĉC": 6391, + "Ġstuff": 6392, + "194": 6393, + "Ġ:Ċ": 6394, + "ĠAR": 6395, + "Task": 6396, + "hidden": 6397, + "eros": 6398, + "IGN": 6399, + "atio": 6400, + "ĠHealth": 6401, + "olute": 6402, + "Enter": 6403, + "'>": 6404, + "ĠTwitter": 6405, + "ĠCounty": 6406, + "scribe": 6407, + "Ġ=>Ċ": 6408, + "Ġhy": 6409, + "fit": 6410, + "Ġmilitary": 6411, + "Ġsale": 6412, + "required": 6413, + "non": 6414, + "bootstrap": 6415, + "hold": 6416, + "rim": 6417, + "-old": 6418, + "ĠDown": 6419, + "Ġmention": 6420, + "contact": 6421, + "_group": 6422, + "oday": 6423, + "Ġtown": 6424, + "Ġsolution": 6425, + "uate": 6426, + "elling": 6427, + "]->": 6428, + "otes": 6429, + "ental": 6430, + "omen": 6431, + "ospital": 6432, + "ĠSup": 6433, + "_EN": 6434, + "Ġslow": 6435, + "SESSION": 6436, + "Ġblue": 6437, + "ago": 6438, + "Ġlives": 6439, + "Ġ^": 6440, + ".un": 6441, + "inst": 6442, + "enge": 6443, + "Ġcustomers": 6444, + "Ġcast": 6445, + "udget": 6446, + "ï¼ģ": 6447, + "icens": 6448, + "Ġdetermin": 6449, + "Selected": 6450, + "_pl": 6451, + "ueue": 6452, + "Ġdark": 6453, + "//ĊĊ": 6454, + "si": 6455, + "thern": 6456, + "ĠJapan": 6457, + "/w": 6458, + "PU": 6459, + "ĠEast": 6460, + "ovie": 6461, + "Ġpackage": 6462, + "Ġnor": 6463, + "Ġapi": 6464, + "bot": 6465, + "\"];Ċ": 6466, + "_post": 6467, + "ulate": 6468, + "Ġclub": 6469, + "'));Ċ": 6470, + "Ġloop": 6471, + "PIO": 6472, + "ione": 6473, + "shot": 6474, + "Initial": 6475, + "Ġplayed": 6476, + "register": 6477, + "rought": 6478, + "_max": 6479, + "acement": 6480, + "match": 6481, + "raphics": 6482, + "AST": 6483, + "Ġexisting": 6484, + "Ġcomplex": 6485, + "DA": 6486, + ".Ch": 6487, + ".common": 6488, + "mo": 6489, + "Ġ'../../": 6490, + "ito": 6491, + "Ġanalysis": 6492, + "Ġdeliver": 6493, + "ĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĊ": 6494, + "idx": 6495, + "Ãł": 6496, + "ongo": 6497, + "ĠEnglish": 6498, + "Ċ": 10197, + "_default": 10198, + "ĠDatabase": 10199, + "rep": 10200, + "ESS": 10201, + "nergy": 10202, + ".Find": 10203, + "_mask": 10204, + "Ġrise": 10205, + "Ġkernel": 10206, + "::$": 10207, + ".Q": 10208, + "Ġoffering": 10209, + "decl": 10210, + "ĠCS": 10211, + "Ġlisted": 10212, + "Ġmostly": 10213, + "enger": 10214, + "Ġblocks": 10215, + "olo": 10216, + "Ġgoverning": 10217, + "\\F": 10218, + "Ġconcent": 10219, + ".getText": 10220, + "Ġmb": 10221, + "Ġoccurred": 10222, + "Ġchanging": 10223, + "Scene": 10224, + "_CODE": 10225, + "Beh": 10226, + "\"The": 10227, + "Ġtile": 10228, + "ĠAssociation": 10229, + "ĉP": 10230, + "alty": 10231, + "_ad": 10232, + "odies": 10233, + "iated": 10234, + "Ġprepared": 10235, + "possible": 10236, + "Ġmort": 10237, + "TEST": 10238, + "142": 10239, + "Ġignore": 10240, + "Ġcalc": 10241, + "Ġrs": 10242, + "ĠassertEquals": 10243, + "Ġsz": 10244, + "ĠTHIS": 10245, + ".\"Ċ": 10246, + "Ġcanvas": 10247, + "java": 10248, + "Ġdut": 10249, + "VALID": 10250, + ".sql": 10251, + ".input": 10252, + "Ġaux": 10253, + "Sup": 10254, + "Ġartist": 10255, + "Vec": 10256, + "_TIME": 10257, + ".stringify": 10258, + "etween": 10259, + "ĠCategory": 10260, + "Ġ[-": 10261, + "ĠDevExpress": 10262, + "ĠJul": 10263, + "Ġring": 10264, + ".ed": 10265, + "YY": 10266, + "Let": 10267, + "TextField": 10268, + "Ġflat": 10269, + "_print": 10270, + "ĠOTHER": 10271, + "adian": 10272, + "Ġchecked": 10273, + "ele": 10274, + "Align": 10275, + "standing": 10276, + "Ġ[],": 10277, + "Ġlab": 10278, + "ucky": 10279, + "ĠChristmas": 10280, + "(image": 10281, + ".module": 10282, + "Ġlots": 10283, + "Ġslightly": 10284, + "(final": 10285, + "erge": 10286, + "è¿": 10287, + "147": 10288, + "ĠPolice": 10289, + "143": 10290, + "ĠRight": 10291, + "Ġaward": 10292, + "ĠOS": 10293, + "Ġ{}ĊĊ": 10294, + "Ġptr": 10295, + "oves": 10296, + "icated": 10297, + "ем": 10298, + "Ġmanage": 10299, + "oliday": 10300, + "Amount": 10301, + "oolStrip": 10302, + "tbody": 10303, + "Nav": 10304, + "wrap": 10305, + "BB": 10306, + "Ġwatching": 10307, + "arios": 10308, + "Ġoptional": 10309, + "_K": 10310, + "ĠLicensed": 10311, + ".Map": 10312, + "Timer": 10313, + "ĠAP": 10314, + "ĠRev": 10315, + "(o": 10316, + ",c": 10317, + "umin": 10318, + "etailed": 10319, + "ĠHy": 10320, + "Ġblank": 10321, + "agger": 10322, + "ĠSelf": 10323, + "()[": 10324, + ".make": 10325, + "earn": 10326, + "channel": 10327, + ";Ċ": 10342, + "World": 10343, + "Ġpython": 10344, + "Ġlif": 10345, + "Ġtrav": 10346, + "Ġconven": 10347, + "company": 10348, + "ĠClub": 10349, + "138": 10350, + "Ver": 10351, + "Btn": 10352, + "Ġzone": 10353, + "products": 10354, + "ĠEduc": 10355, + "Ġverify": 10356, + "ĠMil": 10357, + "ono": 10358, + "]);ĊĊ": 10359, + "ENCE": 10360, + "Ġpacket": 10361, + "Ġcer": 10362, + "Ġenumer": 10363, + "Ġpars": 10364, + "formed": 10365, + "Ġoccup": 10366, + "tre": 10367, + "Ġexercise": 10368, + "Day": 10369, + "_sum": 10370, + "Ġasking": 10371, + "aption": 10372, + "Ġorders": 10373, + "Ġspending": 10374, + "ĠERR": 10375, + ".Dis": 10376, + "ĠUtil": 10377, + "âĢľI": 10378, + "\\'": 10379, + "?)": 10380, + "/>Ċ": 10381, + "Ġemot": 10382, + "Ġinfluence": 10383, + "ĠAfrica": 10384, + "atters": 10385, + "Ùħ": 10386, + ".session": 10387, + "Ġchief": 10388, + "ĉĉĉĉĉĉĉĉĉĉĉ": 10389, + "Ġtom": 10390, + "cluded": 10391, + "serial": 10392, + "_handler": 10393, + ".Type": 10394, + "aped": 10395, + "Ġpolicies": 10396, + "-ex": 10397, + "-tr": 10398, + "blank": 10399, + "merce": 10400, + "Ġcoverage": 10401, + "Ġrc": 10402, + "_matrix": 10403, + "_box": 10404, + "Ġcharges": 10405, + "ĠBoston": 10406, + "Pe": 10407, + "Ġcircum": 10408, + "Ġfilled": 10409, + "148": 10410, + "Ġnorth": 10411, + "ictureBox": 10412, + "ĉres": 10413, + "è®": 10414, + "Ġtermin": 10415, + "Ġ[âĢ¦": 10416, + "IRECT": 10417, + "Ġber": 10418, + "Ġ\"../../": 10419, + "retch": 10420, + ".code": 10421, + "_col": 10422, + "ĠGovernment": 10423, + "Ġargv": 10424, + "ĠLord": 10425, + "asi": 10426, + "Exec": 10427, + "ĉlet": 10428, + "vertis": 10429, + "Ġdiscussion": 10430, + "enance": 10431, + "outube": 10432, + "typeof": 10433, + "Ġserved": 10434, + "ĠPut": 10435, + "ĉx": 10436, + "Ġsweet": 10437, + "Before": 10438, + "ategy": 10439, + ".of": 10440, + "ĠMaterial": 10441, + "Sort": 10442, + "ONT": 10443, + "igital": 10444, + "Why": 10445, + "Ġsust": 10446, + "Ġç": 10447, + "abet": 10448, + "Ġsegment": 10449, + "Ġ[],Ċ": 10450, + "ĠMuslim": 10451, + "ĠfindViewById": 10452, + "cut": 10453, + "_TEXT": 10454, + "ĠMary": 10455, + "Ġloved": 10456, + "Ġlie": 10457, + "ĠJO": 10458, + "Ġisset": 10459, + "month": 10460, + "Ġprime": 10461, + "ti": 10462, + "ĠCarol": 10463, + "Use": 10464, + "146": 10465, + "ĠPop": 10466, + "ĠSave": 10467, + "Interval": 10468, + "execute": 10469, + "dy": 10470, + "ĠIran": 10471, + "_cont": 10472, + "ĉT": 10473, + "Ġphase": 10474, + "checkbox": 10475, + "week": 10476, + "Ġhide": 10477, + "Ġtil": 10478, + "Ġju": 10479, + "Custom": 10480, + "burg": 10481, + "/M": 10482, + "TON": 10483, + "Ġquant": 10484, + "Ġrub": 10485, + "ixels": 10486, + "Ġinstalled": 10487, + "Ġdump": 10488, + "Ġproperly": 10489, + "(List": 10490, + "Ġdecide": 10491, + "apply": 10492, + "Has": 10493, + "Ġkeeping": 10494, + "Ġcitizens": 10495, + "Ġjoint": 10496, + "pool": 10497, + "Socket": 10498, + "_op": 10499, + "Ġweapon": 10500, + "gnore": 10501, + "ĠExec": 10502, + "otten": 10503, + "ĠMS": 10504, + "Ġ(-": 10505, + "ĠReview": 10506, + "Ġexamples": 10507, + "Ġtight": 10508, + "!(": 10509, + "DP": 10510, + "ĠMessageBox": 10511, + "Ġphotograph": 10512, + "164": 10513, + "URI": 10514, + "ét": 10515, + "low": 10516, + "ĠGrand": 10517, + ".persistence": 10518, + "Ġmaintain": 10519, + "Ġnums": 10520, + "Ġzip": 10521, + "ials": 10522, + "ĠGets": 10523, + "peg": 10524, + "ĠBuffer": 10525, + "~~~~": 10526, + "rastructure": 10527, + "ĠPL": 10528, + "uen": 10529, + "obby": 10530, + "sizeof": 10531, + "Ġpic": 10532, + "Ġseed": 10533, + "Ġexperienced": 10534, + "Ġodd": 10535, + "Ġkick": 10536, + "Ġprocedure": 10537, + "avigator": 10538, + "-on": 10539, + ",j": 10540, + "ĠAlthough": 10541, + "ĠuserId": 10542, + "accept": 10543, + "Blue": 10544, + "IColor": 10545, + "layer": 10546, + "available": 10547, + "Ġends": 10548, + ".table": 10549, + "Ġdataset": 10550, + "bus": 10551, + "Ġexplain": 10552, + "(pro": 10553, + "ĠCommittee": 10554, + "Ġnoted": 10555, + "]:Ċ": 10556, + "Dim": 10557, + "stdio": 10558, + "154": 10559, + ".\",Ċ": 10560, + "_source": 10561, + "181": 10562, + "ĠWeek": 10563, + "ĠEdge": 10564, + "Ġoperating": 10565, + "Ġeste": 10566, + "ipl": 10567, + "330": 10568, + "agination": 10569, + "Ġproceed": 10570, + "Ġanimation": 10571, + ".Models": 10572, + "ĠWatch": 10573, + "iat": 10574, + "Ġoppon": 10575, + "/A": 10576, + "Report": 10577, + "Ġsounds": 10578, + "_buf": 10579, + "IELD": 10580, + "Ġbund": 10581, + "ĉget": 10582, + ".pr": 10583, + "(tmp": 10584, + "Ġkid": 10585, + ">ĊĊĊ": 10586, + "Ġyang": 10587, + "NotFound": 10588, + "ÑĨ": 10589, + "math": 10590, + "@gmail": 10591, + "ĠLIMIT": 10592, + "redients": 10593, + "Ġvent": 10594, + "avigate": 10595, + "Look": 10596, + "Ġreligious": 10597, + "Ġrand": 10598, + "rio": 10599, + "(GL": 10600, + "_ip": 10601, + "uan": 10602, + "iciency": 10603, + "ĠChange": 10604, + ">čĊčĊ": 10605, + "ĠEntity": 10606, + "Ġrencontre": 10607, + "ĠRet": 10608, + "plan": 10609, + "én": 10610, + "BOOL": 10611, + "uries": 10612, + "train": 10613, + "Definition": 10614, + "============": 10615, + "zz": 10616, + "450": 10617, + "Animation": 10618, + "ĠOK": 10619, + "_menu": 10620, + ".bl": 10621, + "_score": 10622, + "Ġacad": 10623, + "(System": 10624, + "Ġrefresh": 10625, + "'=>$": 10626, + ".Graphics": 10627, + "amento": 10628, + "pid": 10629, + "tc": 10630, + "Ġtips": 10631, + "Ġhomes": 10632, + "Ġfuel": 10633, + "âĸ": 10634, + "_helper": 10635, + "ĠĠčĊ": 10636, + "ĠRoom": 10637, + ".Close": 10638, + "_attr": 10639, + "ĠMount": 10640, + "ĠEv": 10641, + "arser": 10642, + "_top": 10643, + "eah": 10644, + "ĠDelete": 10645, + "ãĢį": 10646, + "uke": 10647, + "Ġusage": 10648, + "aria": 10649, + "_dev": 10650, + "Ġtexture": 10651, + "Ġconversation": 10652, + "eper": 10653, + "Bean": 10654, + "done": 10655, + "nonatomic": 10656, + "ĠSecond": 10657, + "Ġshooting": 10658, + "_pre": 10659, + "Components": 10660, + "Ġ]ĊĊ": 10661, + "__,": 10662, + "stitution": 10663, + ".Char": 10664, + ">();ĊĊ": 10665, + "Ġpresented": 10666, + "Ġwa": 10667, + "oker": 10668, + "-ĊĊ": 10669, + "iner": 10670, + "Ġbecoming": 10671, + "Ġincident": 10672, + "Att": 10673, + "162": 10674, + "Ġrevealed": 10675, + "forc": 10676, + "Ġboot": 10677, + ".page": 10678, + "Enumerator": 10679, + "165": 10680, + "_->": 10681, + "Photo": 10682, + "Ġspring": 10683, + ".\",": 10684, + "ĠDictionary": 10685, + "BJECT": 10686, + "Ġlocations": 10687, + "Ġsamples": 10688, + "InputStream": 10689, + "ĠBrown": 10690, + "Ġstats": 10691, + "quality": 10692, + "Ñħ": 10693, + "-dis": 10694, + "Ġhelping": 10695, + "Ġped": 10696, + "224": 10697, + "(se": 10698, + "ĠWho": 10699, + "alian": 10700, + "internal": 10701, + "Ġft": 10702, + ">().": 10703, + "->{": 10704, + "Ġmine": 10705, + "Ġsector": 10706, + "Ġgro": 10707, + "Ġopportunities": 10708, + "Ġü": 10709, + "Ġmp": 10710, + "Ġalleged": 10711, + "Ġdoubt": 10712, + "Mouse": 10713, + "About": 10714, + "_part": 10715, + "Ġchair": 10716, + "Ġstopped": 10717, + "161": 10718, + "loop": 10719, + "entities": 10720, + "Ġapps": 10721, + "ansion": 10722, + "Ġmental": 10723, + "ĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠ": 10724, + "FR": 10725, + "Ġdefend": 10726, + "care": 10727, + "Ġideal": 10728, + "/api": 10729, + "urface": 10730, + "011": 10731, + "Ġele": 10732, + "ulator": 10733, + "ĠRights": 10734, + "anguages": 10735, + "Ġfunds": 10736, + "Ġadapt": 10737, + "Attributes": 10738, + "Ġdeploy": 10739, + "opts": 10740, + "Ġvalidation": 10741, + "Ġconcerns": 10742, + "uce": 10743, + ".num": 10744, + "ulture": 10745, + "ila": 10746, + "Ġcup": 10747, + "Ġpure": 10748, + ".Fore": 10749, + "183": 10750, + "ĠHashMap": 10751, + ".valueOf": 10752, + "asm": 10753, + "MO": 10754, + "Ġcs": 10755, + "Ġstores": 10756, + "Ġ************************************************************************": 10757, + "Ġcommunication": 10758, + "mem": 10759, + ".EventHandler": 10760, + ".Status": 10761, + "_right": 10762, + ".setOn": 10763, + "Sheet": 10764, + "Ġidentify": 10765, + "enerated": 10766, + "ordered": 10767, + "Ġ\"[": 10768, + "Ġswe": 10769, + "Condition": 10770, + "ĠAccording": 10771, + "Ġprepare": 10772, + "Ġrob": 10773, + "Pool": 10774, + "Ġsport": 10775, + "rv": 10776, + "ĠRouter": 10777, + "Ġalternative": 10778, + "([]": 10779, + "ĠChicago": 10780, + "ipher": 10781, + "ische": 10782, + "ĠDirector": 10783, + "kl": 10784, + "ĠWil": 10785, + "keys": 10786, + "Ġmysql": 10787, + "Ġwelcome": 10788, + "king": 10789, + "ĠManager": 10790, + "Ġcaught": 10791, + ")}Ċ": 10792, + "Score": 10793, + "_PR": 10794, + "Ġsurvey": 10795, + "hab": 10796, + "Headers": 10797, + "ADER": 10798, + "Ġdecor": 10799, + "Ġturns": 10800, + "Ġradius": 10801, + "errupt": 10802, + "Cor": 10803, + "Ġmel": 10804, + "Ġintr": 10805, + "(q": 10806, + "ĠAC": 10807, + "amos": 10808, + "MAX": 10809, + "ĠGrid": 10810, + "ĠJesus": 10811, + "ĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠ": 10812, + ".DE": 10813, + "Ġts": 10814, + "Ġlinked": 10815, + "free": 10816, + "ĠQt": 10817, + "Ġ/**čĊ": 10818, + "Ġfaster": 10819, + "ctr": 10820, + "_J": 10821, + "DT": 10822, + ".Check": 10823, + "Ġcombination": 10824, + "Ġintended": 10825, + "-the": 10826, + "-type": 10827, + "182": 10828, + "ectors": 10829, + "ami": 10830, + "uting": 10831, + "Ġuma": 10832, + "XML": 10833, + "UCT": 10834, + "Ap": 10835, + "ĠRandom": 10836, + "Ġran": 10837, + ".sort": 10838, + "Ġsorted": 10839, + ".Un": 10840, + "401": 10841, + "_PER": 10842, + "itory": 10843, + "Ġpriority": 10844, + "ĠGal": 10845, + "ĠOld": 10846, + "hot": 10847, + "ĠDisplay": 10848, + "(sub": 10849, + "_TH": 10850, + "_Y": 10851, + "ĠCare": 10852, + "loading": 10853, + "Kind": 10854, + "_handle": 10855, + ",,": 10856, + "rase": 10857, + "_replace": 10858, + ".addEventListener": 10859, + "ĠRT": 10860, + "172": 10861, + "Ġentered": 10862, + "gers": 10863, + "Ġich": 10864, + "(start": 10865, + "205": 10866, + "/app": 10867, + "Ġbrother": 10868, + "Memory": 10869, + "Outlet": 10870, + "Ġutf": 10871, + "prec": 10872, + "Ġnavigation": 10873, + "ORK": 10874, + "Ġdst": 10875, + "Detail": 10876, + "Ġaudience": 10877, + "Ġdur": 10878, + "Ġcluster": 10879, + "unched": 10880, + "Ġ],": 10881, + "Ġcomfortable": 10882, + ".values": 10883, + "ĠTotal": 10884, + "Ġsnap": 10885, + "Ġstandards": 10886, + "Ġperformed": 10887, + "hand": 10888, + "(\"@": 10889, + "åŃ": 10890, + "Ġphil": 10891, + "ibr": 10892, + "trim": 10893, + "Ġforget": 10894, + "157": 10895, + "Ġdoctor": 10896, + ".TextBox": 10897, + "377": 10898, + "icons": 10899, + ",s": 10900, + "ĠOp": 10901, + "Sm": 10902, + "Stop": 10903, + "ĉList": 10904, + "ĉu": 10905, + "Comment": 10906, + "_VERSION": 10907, + ".Xtra": 10908, + "Person": 10909, + "rb": 10910, + "LOB": 10911, + "ĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĊ": 10912, + "ĠCentral": 10913, + "270": 10914, + "ICK": 10915, + "raq": 10916, + "Ġputting": 10917, + "Ġmd": 10918, + "ĠLove": 10919, + "Program": 10920, + "Border": 10921, + "oor": 10922, + "Ġallowing": 10923, + "after": 10924, + "Ġentries": 10925, + "ĠMaybe": 10926, + "]).": 10927, + "ĠShort": 10928, + ")\\": 10929, + ".now": 10930, + "friend": 10931, + "Ġprefer": 10932, + "ĠGPIO": 10933, + "osis": 10934, + "ĠGameObject": 10935, + "Ġskip": 10936, + "Ġcompetition": 10937, + "_match": 10938, + "lications": 10939, + "_CONT": 10940, + ".groupBox": 10941, + "Ġals": 10942, + "666": 10943, + "\"We": 10944, + "_eq": 10945, + "lan": 10946, + "_search": 10947, + "ĠMusic": 10948, + "asis": 10949, + "Ġbind": 10950, + "ĠIsland": 10951, + "rum": 10952, + "(E": 10953, + "Ġseat": 10954, + "Video": 10955, + "Ġack": 10956, + "reek": 10957, + "={()": 10958, + "Ġrating": 10959, + "Ġrestaurant": 10960, + "456": 10961, + "DEX": 10962, + "(buf": 10963, + "pping": 10964, + "uality": 10965, + "Ġleague": 10966, + "176": 10967, + "Ġfocused": 10968, + "apon": 10969, + "$data": 10970, + "CLUD": 10971, + "CLUDING": 10972, + "Ġabsolute": 10973, + "(query": 10974, + "Ġtells": 10975, + "Ang": 10976, + "Ġcommunities": 10977, + "Ġhonest": 10978, + "oking": 10979, + "Ġapart": 10980, + "arity": 10981, + "/$": 10982, + "_module": 10983, + "ĠEnc": 10984, + ".an": 10985, + ".Config": 10986, + "Cre": 10987, + "Ġshock": 10988, + "ĠArab": 10989, + "IENT": 10990, + "/re": 10991, + "Ġretrie": 10992, + "ycler": 10993, + "isa": 10994, + "ĠOrgan": 10995, + ".graph": 10996, + "Ġí": 10997, + "ĠBAS": 10998, + "Enum": 10999, + "Ġpossibly": 11000, + "ÑĢаÐ": 11001, + "ĠJapanese": 11002, + "Ġcraft": 11003, + "ĠPlace": 11004, + "Ġtalent": 11005, + "Ġfunding": 11006, + "Ġconfirmed": 11007, + "Ġcycle": 11008, + "/x": 11009, + "GE": 11010, + "Ġhearing": 11011, + "Ġplants": 11012, + "Ġmouth": 11013, + "pages": 11014, + "oria": 11015, + "ĠRemove": 11016, + "_total": 11017, + "Ġod": 11018, + "ollapse": 11019, + "door": 11020, + "Ġbought": 11021, + "Ġaddr": 11022, + "ARCH": 11023, + "_dim": 11024, + "dden": 11025, + "Ġdecades": 11026, + "REQUEST": 11027, + "Ġversions": 11028, + "fire": 11029, + "006": 11030, + "Ġmoves": 11031, + "fb": 11032, + "Ġcoffee": 11033, + ".connect": 11034, + "ĠRow": 11035, + "Ġschema": 11036, + "Scope": 11037, + "-Type": 11038, + "Ġfighting": 11039, + "Ġretail": 11040, + "Ġmodified": 11041, + "TF": 11042, + "Files": 11043, + "nie": 11044, + "_command": 11045, + "stone": 11046, + "ĠÑĤ": 11047, + "_thread": 11048, + "Ġbond": 11049, + "ĠDevelopment": 11050, + "Ġpt": 11051, + "FORM": 11052, + "plet": 11053, + "Ġidentified": 11054, + "cpp": 11055, + "206": 11056, + "225": 11057, + "Ġcoding": 11058, + "oked": 11059, + "ĠMaster": 11060, + "IDTH": 11061, + "Ġresidents": 11062, + "redit": 11063, + "ĠPhoto": 11064, + "=-": 11065, + "unte": 11066, + "ateur": 11067, + "159": 11068, + "_STATE": 11069, + "ĠSing": 11070, + "Ġsheet": 11071, + ".val": 11072, + "orse": 11073, + "Ġhers": 11074, + "Ġdetermined": 11075, + "Common": 11076, + "Ġwed": 11077, + "_queue": 11078, + "PH": 11079, + "ĠAtl": 11080, + "cred": 11081, + "/LICENSE": 11082, + "Ġmes": 11083, + "Ġadvanced": 11084, + ".java": 11085, + ".Sh": 11086, + "Go": 11087, + "kill": 11088, + "fp": 11089, + "_settings": 11090, + "Ġpal": 11091, + "Ġtruck": 11092, + "Ġcombined": 11093, + "Ġ\"${": 11094, + "ĠCorpor": 11095, + "Ġjoined": 11096, + "ĠJose": 11097, + "ĠCup": 11098, + "uns": 11099, + "estival": 11100, + "levision": 11101, + "Ġbroken": 11102, + "Ġmarriage": 11103, + "ĠWestern": 11104, + "Ġrepresents": 11105, + "ĠTitle": 11106, + "Ġss": 11107, + ".Ass": 11108, + "ongoose": 11109, + "iento": 11110, + "<>();Ċ": 11111, + "Ġabsolutely": 11112, + "Ġsmooth": 11113, + "TERN": 11114, + "ĠUnless": 11115, + "Word": 11116, + "Ġmerge": 11117, + "igan": 11118, + "ĠVol": 11119, + "Ġnn": 11120, + ".getId": 11121, + "Ġз": 11122, + "171": 11123, + "Ġsexy": 11124, + "Ġseeking": 11125, + "Single": 11126, + ".this": 11127, + "179": 11128, + "Ġkom": 11129, + "bound": 11130, + ";\"": 11131, + "ĠfontSize": 11132, + "_df": 11133, + "Ġinjury": 11134, + "(H": 11135, + "Ġissued": 11136, + "_END": 11137, + ":self": 11138, + "020": 11139, + "Ġpatch": 11140, + "Ġleaves": 11141, + "Ġadopt": 11142, + "FileName": 11143, + "ãĢIJ": 11144, + "Ġexecutive": 11145, + "ĠByte": 11146, + "]))Ċ": 11147, + "Ġnu": 11148, + "outing": 11149, + "cluding": 11150, + "-R": 11151, + ".options": 11152, + "Ġsubstant": 11153, + "avax": 11154, + "ĠBUT": 11155, + "Ġtechnical": 11156, + "Ġtwice": 11157, + "Ġmás": 11158, + "Ġunivers": 11159, + "yr": 11160, + "Ġdrag": 11161, + "ĠDC": 11162, + "Ġsed": 11163, + "Ġbot": 11164, + "ĠPal": 11165, + "ĠHall": 11166, + "forcement": 11167, + "Ġauch": 11168, + ".mod": 11169, + "notation": 11170, + "_files": 11171, + ".line": 11172, + "_flag": 11173, + "[name": 11174, + "Ġresolution": 11175, + "Ġbott": 11176, + "(\"[": 11177, + "ende": 11178, + "(arr": 11179, + "Free": 11180, + "(@\"": 11181, + "ĠDistrict": 11182, + "PEC": 11183, + ":-": 11184, + "Picker": 11185, + "ĠJo": 11186, + "ĠĠĠĠĠĊ": 11187, + "ĠRiver": 11188, + "_rows": 11189, + "Ġhelpful": 11190, + "Ġmassive": 11191, + "---Ċ": 11192, + "Ġmeasures": 11193, + "007": 11194, + "ĠRuntime": 11195, + "Ġworry": 11196, + "ĠSpec": 11197, + "ĉD": 11198, + "ãĢij": 11199, + "Ġ){Ċ": 11200, + "Ġworse": 11201, + "(filename": 11202, + "Ġlay": 11203, + "Ġmagic": 11204, + "ĠTheir": 11205, + "oul": 11206, + "stroy": 11207, + "ĠWhere": 11208, + "280": 11209, + "Ġsudden": 11210, + "Ġdefe": 11211, + "Ġbinding": 11212, + "Ġflight": 11213, + "ĠOnInit": 11214, + "ĠWomen": 11215, + "ĠPolicy": 11216, + "Ġdrugs": 11217, + "ishing": 11218, + "('../": 11219, + "ĠMel": 11220, + "peat": 11221, + "tor": 11222, + "Ġproposed": 11223, + "Ġstated": 11224, + "_RES": 11225, + "Ġeast": 11226, + "212": 11227, + "ĠCONDITION": 11228, + "_desc": 11229, + "Ġwinning": 11230, + "folio": 11231, + "Mapper": 11232, + "ĠPan": 11233, + "ĠAnge": 11234, + ".servlet": 11235, + "Ġcopies": 11236, + "LM": 11237, + "Ġvm": 11238, + "åį": 11239, + "Ġdictionary": 11240, + "Seg": 11241, + "177": 11242, + "elines": 11243, + "ĠSend": 11244, + "Ġiron": 11245, + "ĠFort": 11246, + "166": 11247, + ".domain": 11248, + "Ġdebate": 11249, + "NotNull": 11250, + "eq": 11251, + "acher": 11252, + "lf": 11253, + "ĉfmt": 11254, + "Ġlawy": 11255, + "178": 11256, + "ÄŁ": 11257, + "ĠMen": 11258, + "Ġtrim": 11259, + "(NULL": 11260, + "Ġ!!": 11261, + "Ġpad": 11262, + "Ġfollows": 11263, + "\"][\"": 11264, + "requ": 11265, + "ĠEp": 11266, + ".github": 11267, + "(img": 11268, + "eto": 11269, + "('\\": 11270, + "Services": 11271, + "umbnail": 11272, + "_main": 11273, + "pleted": 11274, + "fortunately": 11275, + "Ġwindows": 11276, + "Ġplane": 11277, + "ĠConnection": 11278, + ".local": 11279, + "uard": 11280, + "}\\": 11281, + "==\"": 11282, + "andon": 11283, + "ĠRoy": 11284, + "west": 11285, + "158": 11286, + "iginal": 11287, + "emies": 11288, + "itz": 11289, + "'):Ċ": 11290, + "ĠPeter": 11291, + "Ġtough": 11292, + "Ġreduced": 11293, + "Ġcalculate": 11294, + "Ġrapid": 11295, + "customer": 11296, + "Ġefficient": 11297, + "Ġmedium": 11298, + "Ġfell": 11299, + ".ref": 11300, + "ĠCas": 11301, + "Ġfeedback": 11302, + "Speed": 11303, + "(output": 11304, + "aje": 11305, + "Ġcategories": 11306, + "Ġfee": 11307, + "};": 11308, + "Ġdeleted": 11309, + "reh": 11310, + "Ġproof": 11311, + "Desc": 11312, + "Build": 11313, + "Ġsides": 11314, + ".ArrayList": 11315, + "-%": 11316, + "ĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠ": 11317, + "ر": 11318, + ".match": 11319, + "ли": 11320, + "Ġfeels": 11321, + "Ġachieve": 11322, + "Ġclim": 11323, + "_ON": 11324, + "ĠCD": 11325, + "Ġteacher": 11326, + "_current": 11327, + "bn": 11328, + "_PL": 11329, + "isting": 11330, + "Enable": 11331, + "GEN": 11332, + "Ġtv": 11333, + "Ġsock": 11334, + "Ġplays": 11335, + "Ġdiscount": 11336, + "ĠKE": 11337, + "ĠDebug": 11338, + "Fore": 11339, + "ĠIraq": 11340, + "Ġappearance": 11341, + "Mon": 11342, + "Ġstyled": 11343, + "ĠHuman": 11344, + "iot": 11345, + "ĠHistory": 11346, + "Ġsac": 11347, + "ĠCollection": 11348, + "Ġrecommended": 11349, + ".Selected": 11350, + "Ġorganizations": 11351, + "Ġdiscovered": 11352, + "cohol": 11353, + "adas": 11354, + "ĠThomas": 11355, + "May": 11356, + "Ġconserv": 11357, + "Ġdomin": 11358, + "ĠFollow": 11359, + "ĠSection": 11360, + "ĠThanks": 11361, + "Username": 11362, + "Ġrecipe": 11363, + "Ġwonderful": 11364, + ".sleep": 11365, + "_if": 11366, + "ĉĊĉĊ": 11367, + "orno": 11368, + "Ġru": 11369, + "_target": 11370, + ".\"\"": 11371, + "à¦": 11372, + "EventArgs": 11373, + "Ġinputs": 11374, + "Ġfif": 11375, + "Ġvision": 11376, + "cy": 11377, + "ĠSeries": 11378, + ")(((": 11379, + "Ġtrading": 11380, + "Ġmarker": 11381, + "Begin": 11382, + "Ġtypically": 11383, + "Ġcauses": 11384, + "dropdown": 11385, + "_DEBUG": 11386, + "260": 11387, + "Ġdetect": 11388, + "country": 11389, + "!\");Ċ": 11390, + "ĉR": 11391, + "appy": 11392, + "Ġcref": 11393, + "('<": 11394, + "\"=>": 11395, + "ĠLE": 11396, + "reader": 11397, + "Ġadministr": 11398, + "õ": 11399, + "ucket": 11400, + "Ġfashion": 11401, + ".char": 11402, + "izar": 11403, + "Ġdisable": 11404, + "Ġsuc": 11405, + "ĠLive": 11406, + "issue": 11407, + "Ġmetadata": 11408, + "flags": 11409, + "ĠðŁ": 11410, + "Ġcommitted": 11411, + "Ġva": 11412, + "Ġrough": 11413, + "Ġ'''Ċ": 11414, + "Ġhighlight": 11415, + "_vars": 11416, + "VO": 11417, + "Ġencoding": 11418, + "-Z": 11419, + "_sign": 11420, + "$(\"#": 11421, + "Ġrain": 11422, + "reatest": 11423, + "ĠEND": 11424, + "Selection": 11425, + "Ġcandidates": 11426, + "Ġsav": 11427, + ".Empty": 11428, + "Ġdecisions": 11429, + "Ġcollabor": 11430, + "ridge": 11431, + "feed": 11432, + "ression": 11433, + "Ġpersons": 11434, + "VM": 11435, + "008": 11436, + "ega": 11437, + "_BIT": 11438, + "According": 11439, + "acked": 11440, + "Ġdollars": 11441, + "_loss": 11442, + "ĠCost": 11443, + "}\"Ċ": 11444, + "Notification": 11445, + "Ġprostit": 11446, + "Ġauthority": 11447, + ".rec": 11448, + "Ġspokes": 11449, + "ĠToday": 11450, + "istant": 11451, + "ĠHead": 11452, + "âĢĿ.": 11453, + "ertainment": 11454, + "cean": 11455, + "culate": 11456, + "Ġven": 11457, + "However": 11458, + "_arr": 11459, + "Ġtokens": 11460, + "Graph": 11461, + "ĠJud": 11462, + "ĠVirgin": 11463, + "ĠSerial": 11464, + "unning": 11465, + "Mutable": 11466, + "agers": 11467, + ".csv": 11468, + "Ġdeveloping": 11469, + "Ġinstructions": 11470, + "Ġpromise": 11471, + "Ġrequested": 11472, + "_encode": 11473, + "/\"": 11474, + "ĠIcon": 11475, + "uilt": 11476, + "-day": 11477, + "Ġintelligence": 11478, + ".IS": 11479, + "ĠObservable": 11480, + "ĠHard": 11481, + "Bool": 11482, + "211": 11483, + "idential": 11484, + ".Anchor": 11485, + "Ġselling": 11486, + "CI": 11487, + "AGES": 11488, + "tle": 11489, + "bur": 11490, + "UFFER": 11491, + "RY": 11492, + "Ġbigger": 11493, + "Ġrat": 11494, + "Ġfamous": 11495, + "Ġtypename": 11496, + "Ġexplained": 11497, + "}}Ċ": 11498, + "Ġnuclear": 11499, + "-N": 11500, + "Ġcrisis": 11501, + "ĠEnter": 11502, + "Ġanswers": 11503, + "/${": 11504, + "/pl": 11505, + "Ġsequ": 11506, + "_next": 11507, + "mask": 11508, + "Ġstanding": 11509, + "Ġplenty": 11510, + "ĠCross": 11511, + "ĉret": 11512, + "dro": 11513, + "ĠCast": 11514, + "167": 11515, + "=true": 11516, + "ĠChris": 11517, + "icio": 11518, + "ĠMike": 11519, + "Decimal": 11520, + "addComponent": 11521, + "Len": 11522, + "Ġcock": 11523, + "Ġ#{": 11524, + "URN": 11525, + "": 11657, + "Ġ*=": 11658, + "ĠPS": 11659, + "Ġdangerous": 11660, + "[p": 11661, + "OME": 11662, + "Other": 11663, + "ĠStringBuilder": 11664, + "Points": 11665, + "heading": 11666, + "Ġcurrency": 11667, + "Ġpercentage": 11668, + "_API": 11669, + "Ġclassic": 11670, + "thead": 11671, + "ĠMO": 11672, + "FE": 11673, + "Idx": 11674, + "await": 11675, + "Ġè": 11676, + "Ġaccident": 11677, + "Ġvariant": 11678, + "Ġmyst": 11679, + "ĠLand": 11680, + "ĠBre": 11681, + "Ġharm": 11682, + "ĠAcc": 11683, + "Ġcharged": 11684, + "iones": 11685, + "Visibility": 11686, + "arry": 11687, + "ĠLanguage": 11688, + "Ġwalking": 11689, + "\".ĊĊ": 11690, + "ifer": 11691, + "Ġleadership": 11692, + ".From": 11693, + "ynam": 11694, + "Ġtimestamp": 11695, + "ipt": 11696, + "ĠHas": 11697, + "REFER": 11698, + "ĠIts": 11699, + "Ġlistener": 11700, + "UTE": 11701, + "213": 11702, + "_description": 11703, + "Ġexperiences": 11704, + "Ġcreates": 11705, + "RS": 11706, + "cart": 11707, + "black": 11708, + "Ġchoices": 11709, + "war": 11710, + "750": 11711, + "Ġ'''": 11712, + "Ġordered": 11713, + "Ġevening": 11714, + "Ġpil": 11715, + "Ġtun": 11716, + "ĠBad": 11717, + "(app": 11718, + "random": 11719, + "Ġexplicit": 11720, + "Ġarrived": 11721, + "Ġfly": 11722, + "Ġeconom": 11723, + "-mail": 11724, + "Ġlists": 11725, + "Ġarchitect": 11726, + "234": 11727, + "ĠPay": 11728, + "Ġds": 11729, + "ĠSol": 11730, + "Ġvehicles": 11731, + "Hz": 11732, + "-com": 11733, + "Ġking": 11734, + "_equal": 11735, + "ĠHelp": 11736, + "Ġabuse": 11737, + "480": 11738, + "169": 11739, + "--;Ċ": 11740, + "Ġextr": 11741, + "Ġchemical": 11742, + "ä¿": 11743, + "Ġorient": 11744, + "Ġbreath": 11745, + "ĠSpace": 11746, + "(element": 11747, + "wait": 11748, + "DED": 11749, + "igma": 11750, + "Ġentr": 11751, + "Ġsob": 11752, + "-name": 11753, + "Ġaffected": 11754, + "ika": 11755, + "Ġcoal": 11756, + "_work": 11757, + "Ġhundreds": 11758, + "Ġpolitics": 11759, + "subject": 11760, + "Ġconsumer": 11761, + "ANGE": 11762, + "Ġrepeated": 11763, + "Send": 11764, + "Ġ#[": 11765, + "Ġprotocol": 11766, + "Ġleads": 11767, + "useum": 11768, + "Every": 11769, + "808": 11770, + "174": 11771, + "Import": 11772, + "(count": 11773, + "Ġchallenges": 11774, + "Ġnovel": 11775, + "Ġdepart": 11776, + "bits": 11777, + ".Current": 11778, + "Ġ`${": 11779, + "oting": 11780, + "(\\": 11781, + "Ġcreative": 11782, + "Ġbuff": 11783, + "Ġintroduced": 11784, + "usic": 11785, + "modules": 11786, + "Are": 11787, + "-doc": 11788, + "language": 11789, + "_cache": 11790, + "Ġtod": 11791, + "?>": 11792, + "omething": 11793, + "Ġhun": 11794, + "åº": 11795, + "aters": 11796, + "Intent": 11797, + "Ġimplemented": 11798, + "ĠCase": 11799, + "Children": 11800, + "Ġnotification": 11801, + "Renderer": 11802, + "Wrapper": 11803, + "Objects": 11804, + "tl": 11805, + ".Contains": 11806, + "Plugin": 11807, + ".row": 11808, + "Ġforg": 11809, + "Ġpermit": 11810, + "Ġtargets": 11811, + "ĠIF": 11812, + "Ġtip": 11813, + "sex": 11814, + "Ġsupports": 11815, + "Ġfold": 11816, + "photo": 11817, + "},čĊ": 11818, + "Ġgoogle": 11819, + "$('#": 11820, + "Ġsharing": 11821, + "Ġgoods": 11822, + "vs": 11823, + "ĠDan": 11824, + "Rate": 11825, + "ĠMartin": 11826, + "Ġmanner": 11827, + "lie": 11828, + ".The": 11829, + "Internal": 11830, + "ĠCONTR": 11831, + "Mock": 11832, + "RIGHT": 11833, + "Ġ'{": 11834, + "Ġcontrols": 11835, + "Mat": 11836, + "Ġmand": 11837, + "Ġextended": 11838, + "Ok": 11839, + "Ġembed": 11840, + "Ġplanet": 11841, + "ĠNon": 11842, + "-ch": 11843, + ")\",": 11844, + "epar": 11845, + "Ġbelieved": 11846, + "ĠEnvironment": 11847, + "ĠFriend": 11848, + "-res": 11849, + "Ġhandling": 11850, + "nic": 11851, + "-level": 11852, + "scri": 11853, + "Xml": 11854, + "BE": 11855, + "ungen": 11856, + "Ġalter": 11857, + "[idx": 11858, + "Pop": 11859, + "cam": 11860, + "Ġ(((": 11861, + "Ġshipping": 11862, + "Ġbattery": 11863, + "iddleware": 11864, + "MC": 11865, + "Ġimpl": 11866, + "otation": 11867, + "ĠLab": 11868, + "