Synchronizing local compiler cache.
Browse filesThis view is limited to 50 files because it contains too many changes.
See raw diff
- .gitattributes +83 -0
- neuronxcc-2.13.66.0+6dfecc895/0_REGISTRY/0.0.21.dev0/inference/bloom/hf-internal-testing/tiny-random-BloomForCausalLM/4eccd2e66477ff5ef5b3.json +1 -0
- neuronxcc-2.13.66.0+6dfecc895/0_REGISTRY/0.0.21.dev0/inference/bloom/hf-internal-testing/tiny-random-BloomForCausalLM/d635309efcd921a3a3f6.json +1 -0
- neuronxcc-2.13.66.0+6dfecc895/0_REGISTRY/0.0.21.dev0/inference/gpt2/gpt2/2c3a47e6fca2fd6ac3cf.json +1 -0
- neuronxcc-2.13.66.0+6dfecc895/0_REGISTRY/0.0.21.dev0/inference/gpt2/hf-internal-testing/tiny-random-gpt2/2037d1be0146ceb8d639.json +1 -0
- neuronxcc-2.13.66.0+6dfecc895/0_REGISTRY/0.0.21.dev0/inference/gpt2/hf-internal-testing/tiny-random-gpt2/3edc0caf8d3805a471bd.json +1 -0
- neuronxcc-2.13.66.0+6dfecc895/0_REGISTRY/0.0.21.dev0/inference/llama/NousResearch/Llama-2-7b-chat-hf/6c7cfec071642e5fecee.json +1 -0
- neuronxcc-2.13.66.0+6dfecc895/0_REGISTRY/0.0.21.dev0/inference/llama/NousResearch/Llama-2-7b-chat-hf/b9904e0c46d338b623d2.json +1 -0
- neuronxcc-2.13.66.0+6dfecc895/0_REGISTRY/0.0.21.dev0/inference/llama/dacorvo/tiny-random-llama/7c595f241fa311d7b227.json +1 -0
- neuronxcc-2.13.66.0+6dfecc895/0_REGISTRY/0.0.21.dev0/inference/llama/dacorvo/tiny-random-llama/b650b9d41c4386d36a89.json +1 -0
- neuronxcc-2.13.66.0+6dfecc895/0_REGISTRY/0.0.21.dev0/inference/llama/meta-llama/Llama-2-7b-chat-hf/3a328b55c6445b520b8e.json +1 -0
- neuronxcc-2.13.66.0+6dfecc895/0_REGISTRY/0.0.21.dev0/inference/llama/meta-llama/Llama-2-7b-chat-hf/44c81784d74cd4713969.json +1 -0
- neuronxcc-2.13.66.0+6dfecc895/0_REGISTRY/0.0.21.dev0/inference/llama/meta-llama/Llama-2-7b-chat-hf/58eec5854a9f564f8a27.json +1 -0
- neuronxcc-2.13.66.0+6dfecc895/0_REGISTRY/0.0.21.dev0/inference/llama/meta-llama/Llama-2-7b-chat-hf/6a314b3d69ca329d2711.json +1 -0
- neuronxcc-2.13.66.0+6dfecc895/0_REGISTRY/0.0.21.dev0/inference/llama/meta-llama/Llama-2-7b-chat-hf/7a0a7caf91982c3c67b0.json +1 -0
- neuronxcc-2.13.66.0+6dfecc895/0_REGISTRY/0.0.21.dev0/inference/llama/meta-llama/Llama-2-7b-chat-hf/ad1f7e0ca3ab366f91d4.json +1 -0
- neuronxcc-2.13.66.0+6dfecc895/0_REGISTRY/0.0.21.dev0/inference/llama/meta-llama/Llama-2-7b-chat-hf/dd0ae8e9d982a2ed293b.json +1 -0
- neuronxcc-2.13.66.0+6dfecc895/0_REGISTRY/0.0.21.dev0/inference/llama/meta-llama/Llama-2-7b-chat-hf/efa4119dabd9c1f8277f.json +1 -0
- neuronxcc-2.13.66.0+6dfecc895/0_REGISTRY/0.0.21.dev0/inference/mistral/dacorvo/tiny-random-MistralForCausalLM/43c8baaa3d89767fe7f2.json +1 -0
- neuronxcc-2.13.66.0+6dfecc895/0_REGISTRY/0.0.21.dev0/inference/mistral/dacorvo/tiny-random-MistralForCausalLM/f9b607c62a5393e0b42f.json +1 -0
- neuronxcc-2.13.66.0+6dfecc895/0_REGISTRY/0.0.21.dev0/inference/mistral/mistralai/Mistral-7B-Instruct-v0.1/7a99245071db1d4e48e6.json +1 -0
- neuronxcc-2.13.66.0+6dfecc895/0_REGISTRY/0.0.21.dev0/inference/mistral/mistralai/Mistral-7B-Instruct-v0.1/b0a5a2e1e8dd02dbe923.json +1 -0
- neuronxcc-2.13.66.0+6dfecc895/0_REGISTRY/0.0.21.dev0/inference/mistral/mistralai/Mistral-7B-Instruct-v0.2/1c4540cc29dcb3b99ff8.json +1 -0
- neuronxcc-2.13.66.0+6dfecc895/0_REGISTRY/0.0.21.dev0/inference/mistral/mistralai/Mistral-7B-Instruct-v0.2/2d6c1d9513debbcc5f7f.json +1 -0
- neuronxcc-2.13.66.0+6dfecc895/0_REGISTRY/0.0.21.dev0/inference/mistral/mistralai/Mistral-7B-Instruct-v0.2/41284a607429208347c0.json +1 -0
- neuronxcc-2.13.66.0+6dfecc895/0_REGISTRY/0.0.21.dev0/inference/mistral/mistralai/Mistral-7B-Instruct-v0.2/4620b76c4f39e71a8459.json +1 -0
- neuronxcc-2.13.66.0+6dfecc895/0_REGISTRY/0.0.21.dev0/inference/mistral/mistralai/Mistral-7B-Instruct-v0.2/4fcf5530fe99f6f4f3e6.json +1 -0
- neuronxcc-2.13.66.0+6dfecc895/0_REGISTRY/0.0.21.dev0/inference/mistral/mistralai/Mistral-7B-Instruct-v0.2/5c9bfea69bb260a59b0c.json +1 -0
- neuronxcc-2.13.66.0+6dfecc895/0_REGISTRY/0.0.21.dev0/inference/mistral/mistralai/Mistral-7B-Instruct-v0.2/698b844f07e07829c78e.json +1 -0
- neuronxcc-2.13.66.0+6dfecc895/0_REGISTRY/0.0.21.dev0/inference/mistral/mistralai/Mistral-7B-Instruct-v0.2/a2a8d3c93367a33b250c.json +1 -0
- neuronxcc-2.13.66.0+6dfecc895/0_REGISTRY/0.0.21.dev0/inference/mistral/mistralai/Mistral-7B-Instruct-v0.2/a61d8247a227bdcc16f2.json +1 -0
- neuronxcc-2.13.66.0+6dfecc895/0_REGISTRY/0.0.21.dev0/inference/mistral/mistralai/Mistral-7B-Instruct-v0.2/ca9ec6b16e1a62dbd649.json +1 -0
- neuronxcc-2.13.66.0+6dfecc895/0_REGISTRY/0.0.21.dev0/inference/mistral/mistralai/Mistral-7B-Instruct-v0.2/d75fbb50baa6a3b15792.json +1 -0
- neuronxcc-2.13.66.0+6dfecc895/0_REGISTRY/0.0.21.dev0/inference/mistral/mistralai/Mistral-7B-Instruct-v0.2/de7f8e3eb7e911ee8559.json +1 -0
- neuronxcc-2.13.66.0+6dfecc895/0_REGISTRY/0.0.21.dev0/inference/mistral/mistralai/Mistral-7B-v0.1/12212376258a9fed88b2.json +1 -0
- neuronxcc-2.13.66.0+6dfecc895/0_REGISTRY/0.0.21.dev0/inference/mistral/mistralai/Mistral-7B-v0.1/15f566c81d1f67f6fd2e.json +1 -0
- neuronxcc-2.13.66.0+6dfecc895/0_REGISTRY/0.0.21.dev0/inference/mistral/mistralai/Mistral-7B-v0.1/192bd7f0468f78103585.json +1 -0
- neuronxcc-2.13.66.0+6dfecc895/0_REGISTRY/0.0.21.dev0/inference/mistral/mistralai/Mistral-7B-v0.1/1dd1302a33c9e69730d2.json +1 -0
- neuronxcc-2.13.66.0+6dfecc895/0_REGISTRY/0.0.21.dev0/inference/mistral/mistralai/Mistral-7B-v0.1/21bbf61928931f72378f.json +1 -0
- neuronxcc-2.13.66.0+6dfecc895/0_REGISTRY/0.0.21.dev0/inference/mistral/mistralai/Mistral-7B-v0.1/5141bf1fc4c434ce1ea7.json +1 -0
- neuronxcc-2.13.66.0+6dfecc895/0_REGISTRY/0.0.21.dev0/inference/mistral/mistralai/Mistral-7B-v0.1/67dd8bb21d625e22cd5b.json +1 -0
- neuronxcc-2.13.66.0+6dfecc895/0_REGISTRY/0.0.21.dev0/inference/mistral/mistralai/Mistral-7B-v0.1/81be8abab2484836ea97.json +1 -0
- neuronxcc-2.13.66.0+6dfecc895/0_REGISTRY/0.0.21.dev0/inference/mistral/mistralai/Mistral-7B-v0.1/9e73f279a5220e46dcf1.json +1 -0
- neuronxcc-2.13.66.0+6dfecc895/0_REGISTRY/0.0.21.dev0/inference/mistral/mistralai/Mistral-7B-v0.1/9eb34c5c51dae5739571.json +1 -0
- neuronxcc-2.13.66.0+6dfecc895/0_REGISTRY/0.0.21.dev0/inference/opt/hf-internal-testing/tiny-random-OPTForCausalLM/0eb6d808c895dd98f1d9.json +1 -0
- neuronxcc-2.13.66.0+6dfecc895/0_REGISTRY/0.0.21.dev0/inference/opt/hf-internal-testing/tiny-random-OPTForCausalLM/244d78257ade535a7f03.json +1 -0
- neuronxcc-2.13.66.0+6dfecc895/MODULE_03e6f14805ce902580b5+2c2d707e/compile_flags.txt +1 -0
- neuronxcc-2.13.66.0+6dfecc895/MODULE_03e6f14805ce902580b5+2c2d707e/model.hlo.pb +3 -0
- neuronxcc-2.13.66.0+6dfecc895/MODULE_03e6f14805ce902580b5+2c2d707e/model.neff +3 -0
- neuronxcc-2.13.66.0+6dfecc895/MODULE_05115dff8003eb0253a4+2c2d707e/compile_flags.txt +1 -0
.gitattributes
CHANGED
@@ -1670,3 +1670,86 @@ neuronxcc-2.12.68.0+4480452af/MODULE_b81e85f01e4cafae9777/unet/model.neuron filt
|
|
1670 |
neuronxcc-2.12.68.0+4480452af/MODULE_b81e85f01e4cafae9777/vae_decoder/model.neuron filter=lfs diff=lfs merge=lfs -text
|
1671 |
neuronxcc-2.12.68.0+4480452af/MODULE_b81e85f01e4cafae9777/vae_encoder/model.neuron filter=lfs diff=lfs merge=lfs -text
|
1672 |
neuronxcc-2.12.68.0+4480452af/MODULE_0cf5036cc6234d94717d/vae_decoder/model.neuron filter=lfs diff=lfs merge=lfs -text
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1670 |
neuronxcc-2.12.68.0+4480452af/MODULE_b81e85f01e4cafae9777/vae_decoder/model.neuron filter=lfs diff=lfs merge=lfs -text
|
1671 |
neuronxcc-2.12.68.0+4480452af/MODULE_b81e85f01e4cafae9777/vae_encoder/model.neuron filter=lfs diff=lfs merge=lfs -text
|
1672 |
neuronxcc-2.12.68.0+4480452af/MODULE_0cf5036cc6234d94717d/vae_decoder/model.neuron filter=lfs diff=lfs merge=lfs -text
|
1673 |
+
neuronxcc-2.13.66.0+6dfecc895/MODULE_03e6f14805ce902580b5+2c2d707e/model.neff filter=lfs diff=lfs merge=lfs -text
|
1674 |
+
neuronxcc-2.13.66.0+6dfecc895/MODULE_05115dff8003eb0253a4+2c2d707e/model.neff filter=lfs diff=lfs merge=lfs -text
|
1675 |
+
neuronxcc-2.13.66.0+6dfecc895/MODULE_07094468d7551fab9fdc+2c2d707e/model.neff filter=lfs diff=lfs merge=lfs -text
|
1676 |
+
neuronxcc-2.13.66.0+6dfecc895/MODULE_095d0a89b9edc7553e7d+2c2d707e/model.neff filter=lfs diff=lfs merge=lfs -text
|
1677 |
+
neuronxcc-2.13.66.0+6dfecc895/MODULE_0996f2b96198b08c3286+2c2d707e/model.neff filter=lfs diff=lfs merge=lfs -text
|
1678 |
+
neuronxcc-2.13.66.0+6dfecc895/MODULE_0eefbe1ea116943f78db+2c2d707e/model.neff filter=lfs diff=lfs merge=lfs -text
|
1679 |
+
neuronxcc-2.13.66.0+6dfecc895/MODULE_12168d57509f6fd83d62+2c2d707e/model.neff filter=lfs diff=lfs merge=lfs -text
|
1680 |
+
neuronxcc-2.13.66.0+6dfecc895/MODULE_1392608e4e2384009161+2c2d707e/model.neff filter=lfs diff=lfs merge=lfs -text
|
1681 |
+
neuronxcc-2.13.66.0+6dfecc895/MODULE_187e41e4e83672ee0ab2+2c2d707e/model.neff filter=lfs diff=lfs merge=lfs -text
|
1682 |
+
neuronxcc-2.13.66.0+6dfecc895/MODULE_18ce21b41f081127069b+2c2d707e/model.neff filter=lfs diff=lfs merge=lfs -text
|
1683 |
+
neuronxcc-2.13.66.0+6dfecc895/MODULE_1b2a75fa86a836669b6e+2c2d707e/model.neff filter=lfs diff=lfs merge=lfs -text
|
1684 |
+
neuronxcc-2.13.66.0+6dfecc895/MODULE_1f42d4e3599f2c568dff+2c2d707e/model.neff filter=lfs diff=lfs merge=lfs -text
|
1685 |
+
neuronxcc-2.13.66.0+6dfecc895/MODULE_21d7ca83fa40eaf22249+2c2d707e/model.neff filter=lfs diff=lfs merge=lfs -text
|
1686 |
+
neuronxcc-2.13.66.0+6dfecc895/MODULE_21ed6cb6eb66db9977ec+2c2d707e/model.neff filter=lfs diff=lfs merge=lfs -text
|
1687 |
+
neuronxcc-2.13.66.0+6dfecc895/MODULE_22512a4c8b1c5b2889da+2c2d707e/model.neff filter=lfs diff=lfs merge=lfs -text
|
1688 |
+
neuronxcc-2.13.66.0+6dfecc895/MODULE_2dd8cab3c2b7a2a2ec97+2c2d707e/model.neff filter=lfs diff=lfs merge=lfs -text
|
1689 |
+
neuronxcc-2.13.66.0+6dfecc895/MODULE_2ed40b6357fdea887a02+2c2d707e/model.neff filter=lfs diff=lfs merge=lfs -text
|
1690 |
+
neuronxcc-2.13.66.0+6dfecc895/MODULE_2f4064fb71e6a101298d+2c2d707e/model.neff filter=lfs diff=lfs merge=lfs -text
|
1691 |
+
neuronxcc-2.13.66.0+6dfecc895/MODULE_308d19f6ceb008fe81c6+2c2d707e/model.neff filter=lfs diff=lfs merge=lfs -text
|
1692 |
+
neuronxcc-2.13.66.0+6dfecc895/MODULE_360af0ca410b154b1ace+2c2d707e/model.neff filter=lfs diff=lfs merge=lfs -text
|
1693 |
+
neuronxcc-2.13.66.0+6dfecc895/MODULE_3c1747dd4e7a2cf4a941+2c2d707e/model.neff filter=lfs diff=lfs merge=lfs -text
|
1694 |
+
neuronxcc-2.13.66.0+6dfecc895/MODULE_3cfe8bed0e2b690bb440+2c2d707e/model.neff filter=lfs diff=lfs merge=lfs -text
|
1695 |
+
neuronxcc-2.13.66.0+6dfecc895/MODULE_3d29e09e1a40ac65135e+2c2d707e/model.neff filter=lfs diff=lfs merge=lfs -text
|
1696 |
+
neuronxcc-2.13.66.0+6dfecc895/MODULE_41a4e96d479110ad05e9+2c2d707e/model.neff filter=lfs diff=lfs merge=lfs -text
|
1697 |
+
neuronxcc-2.13.66.0+6dfecc895/MODULE_42adf97202125e822891+2c2d707e/model.neff filter=lfs diff=lfs merge=lfs -text
|
1698 |
+
neuronxcc-2.13.66.0+6dfecc895/MODULE_468fdaac8c4e04ccff09+2c2d707e/model.neff filter=lfs diff=lfs merge=lfs -text
|
1699 |
+
neuronxcc-2.13.66.0+6dfecc895/MODULE_46ad015ca5c214c77ae7+2c2d707e/model.neff filter=lfs diff=lfs merge=lfs -text
|
1700 |
+
neuronxcc-2.13.66.0+6dfecc895/MODULE_4a5008dd1f169231c8ef+2c2d707e/model.neff filter=lfs diff=lfs merge=lfs -text
|
1701 |
+
neuronxcc-2.13.66.0+6dfecc895/MODULE_4dee62f3941c8a0c1522+2c2d707e/model.neff filter=lfs diff=lfs merge=lfs -text
|
1702 |
+
neuronxcc-2.13.66.0+6dfecc895/MODULE_4e02f6a89a1b9523cdb6+2c2d707e/model.neff filter=lfs diff=lfs merge=lfs -text
|
1703 |
+
neuronxcc-2.13.66.0+6dfecc895/MODULE_4f35658764c67c288bfa+2c2d707e/model.neff filter=lfs diff=lfs merge=lfs -text
|
1704 |
+
neuronxcc-2.13.66.0+6dfecc895/MODULE_5392086576db6170dc5d+2c2d707e/model.neff filter=lfs diff=lfs merge=lfs -text
|
1705 |
+
neuronxcc-2.13.66.0+6dfecc895/MODULE_53f9eec678d92377113f+2c2d707e/model.neff filter=lfs diff=lfs merge=lfs -text
|
1706 |
+
neuronxcc-2.13.66.0+6dfecc895/MODULE_55bf7d5c44a1887eee30+2c2d707e/model.neff filter=lfs diff=lfs merge=lfs -text
|
1707 |
+
neuronxcc-2.13.66.0+6dfecc895/MODULE_57a455f0da18be1bfc7e+2c2d707e/model.neff filter=lfs diff=lfs merge=lfs -text
|
1708 |
+
neuronxcc-2.13.66.0+6dfecc895/MODULE_5d3278fc5d1f8c7be231+2c2d707e/model.neff filter=lfs diff=lfs merge=lfs -text
|
1709 |
+
neuronxcc-2.13.66.0+6dfecc895/MODULE_654bf58515e816add109+2c2d707e/model.neff filter=lfs diff=lfs merge=lfs -text
|
1710 |
+
neuronxcc-2.13.66.0+6dfecc895/MODULE_66500b17d83bf57f9047+2c2d707e/model.neff filter=lfs diff=lfs merge=lfs -text
|
1711 |
+
neuronxcc-2.13.66.0+6dfecc895/MODULE_67263bb0ab07a92a9298+2c2d707e/model.neff filter=lfs diff=lfs merge=lfs -text
|
1712 |
+
neuronxcc-2.13.66.0+6dfecc895/MODULE_6731115cadf25f9d2615+2c2d707e/model.neff filter=lfs diff=lfs merge=lfs -text
|
1713 |
+
neuronxcc-2.13.66.0+6dfecc895/MODULE_685aea3a3722d7e811aa+2c2d707e/model.neff filter=lfs diff=lfs merge=lfs -text
|
1714 |
+
neuronxcc-2.13.66.0+6dfecc895/MODULE_68da807c3de1c79a2480+2c2d707e/model.neff filter=lfs diff=lfs merge=lfs -text
|
1715 |
+
neuronxcc-2.13.66.0+6dfecc895/MODULE_73ef6824f1be65053232+2c2d707e/model.neff filter=lfs diff=lfs merge=lfs -text
|
1716 |
+
neuronxcc-2.13.66.0+6dfecc895/MODULE_761ebeb9760c5013a1ee+2c2d707e/model.neff filter=lfs diff=lfs merge=lfs -text
|
1717 |
+
neuronxcc-2.13.66.0+6dfecc895/MODULE_79510c16b2a69226bf74+2c2d707e/model.neff filter=lfs diff=lfs merge=lfs -text
|
1718 |
+
neuronxcc-2.13.66.0+6dfecc895/MODULE_82c12499f2f9a0ad38cc+2c2d707e/model.neff filter=lfs diff=lfs merge=lfs -text
|
1719 |
+
neuronxcc-2.13.66.0+6dfecc895/MODULE_8c2f4d9e3c1b0b5977e6+2c2d707e/model.neff filter=lfs diff=lfs merge=lfs -text
|
1720 |
+
neuronxcc-2.13.66.0+6dfecc895/MODULE_8cb29fff32cf48a38ac9+2c2d707e/model.neff filter=lfs diff=lfs merge=lfs -text
|
1721 |
+
neuronxcc-2.13.66.0+6dfecc895/MODULE_92503e1fc7e9cbf8e9e0+2c2d707e/model.neff filter=lfs diff=lfs merge=lfs -text
|
1722 |
+
neuronxcc-2.13.66.0+6dfecc895/MODULE_938709e4ce13e9bd463f+2c2d707e/model.neff filter=lfs diff=lfs merge=lfs -text
|
1723 |
+
neuronxcc-2.13.66.0+6dfecc895/MODULE_95a2e81174583451888c+2c2d707e/model.neff filter=lfs diff=lfs merge=lfs -text
|
1724 |
+
neuronxcc-2.13.66.0+6dfecc895/MODULE_9b4e781d84e3ad27ec93+2c2d707e/model.neff filter=lfs diff=lfs merge=lfs -text
|
1725 |
+
neuronxcc-2.13.66.0+6dfecc895/MODULE_9fb4f17da3ced15711f2+2c2d707e/model.neff filter=lfs diff=lfs merge=lfs -text
|
1726 |
+
neuronxcc-2.13.66.0+6dfecc895/MODULE_ab853fab33da28c1092e+2c2d707e/model.neff filter=lfs diff=lfs merge=lfs -text
|
1727 |
+
neuronxcc-2.13.66.0+6dfecc895/MODULE_ace8e33ed5e9bedef6f1+2c2d707e/model.neff filter=lfs diff=lfs merge=lfs -text
|
1728 |
+
neuronxcc-2.13.66.0+6dfecc895/MODULE_ae5d6772a3d1f4a67fe4+2c2d707e/model.neff filter=lfs diff=lfs merge=lfs -text
|
1729 |
+
neuronxcc-2.13.66.0+6dfecc895/MODULE_b202a8494e6fa17d96f6+2c2d707e/model.neff filter=lfs diff=lfs merge=lfs -text
|
1730 |
+
neuronxcc-2.13.66.0+6dfecc895/MODULE_bd3143c0ee5eecf7a709+2c2d707e/model.neff filter=lfs diff=lfs merge=lfs -text
|
1731 |
+
neuronxcc-2.13.66.0+6dfecc895/MODULE_bf473fa2692e665f714a+2c2d707e/model.neff filter=lfs diff=lfs merge=lfs -text
|
1732 |
+
neuronxcc-2.13.66.0+6dfecc895/MODULE_c1ba00cedaa0eb6aea8b+2c2d707e/model.neff filter=lfs diff=lfs merge=lfs -text
|
1733 |
+
neuronxcc-2.13.66.0+6dfecc895/MODULE_c231b1612b7e21a13adc+2c2d707e/model.neff filter=lfs diff=lfs merge=lfs -text
|
1734 |
+
neuronxcc-2.13.66.0+6dfecc895/MODULE_c6a7accddaedb5e55b4d+2c2d707e/model.neff filter=lfs diff=lfs merge=lfs -text
|
1735 |
+
neuronxcc-2.13.66.0+6dfecc895/MODULE_c87736fc8d652bb9115d+2c2d707e/model.neff filter=lfs diff=lfs merge=lfs -text
|
1736 |
+
neuronxcc-2.13.66.0+6dfecc895/MODULE_d0d1a0d9ca4d7cc149a1+2c2d707e/model.neff filter=lfs diff=lfs merge=lfs -text
|
1737 |
+
neuronxcc-2.13.66.0+6dfecc895/MODULE_d3c63fb2d0c46ee11f9b+2c2d707e/model.neff filter=lfs diff=lfs merge=lfs -text
|
1738 |
+
neuronxcc-2.13.66.0+6dfecc895/MODULE_d752ec325d750ef42acb+2c2d707e/model.neff filter=lfs diff=lfs merge=lfs -text
|
1739 |
+
neuronxcc-2.13.66.0+6dfecc895/MODULE_d9a0688e74add49a38ab+2c2d707e/model.neff filter=lfs diff=lfs merge=lfs -text
|
1740 |
+
neuronxcc-2.13.66.0+6dfecc895/MODULE_d9b09088fcdcbb3215b7+2c2d707e/model.neff filter=lfs diff=lfs merge=lfs -text
|
1741 |
+
neuronxcc-2.13.66.0+6dfecc895/MODULE_dae64c7e2fea06ef54cc+2c2d707e/model.neff filter=lfs diff=lfs merge=lfs -text
|
1742 |
+
neuronxcc-2.13.66.0+6dfecc895/MODULE_ddc0d5d420d12c3bc7e3+2c2d707e/model.neff filter=lfs diff=lfs merge=lfs -text
|
1743 |
+
neuronxcc-2.13.66.0+6dfecc895/MODULE_de41be3e18f84c316d01+2c2d707e/model.neff filter=lfs diff=lfs merge=lfs -text
|
1744 |
+
neuronxcc-2.13.66.0+6dfecc895/MODULE_de5dda9f9f101cdf43f2+2c2d707e/model.neff filter=lfs diff=lfs merge=lfs -text
|
1745 |
+
neuronxcc-2.13.66.0+6dfecc895/MODULE_df2340fad5eed1527e92+2c2d707e/model.neff filter=lfs diff=lfs merge=lfs -text
|
1746 |
+
neuronxcc-2.13.66.0+6dfecc895/MODULE_e08070f3a58b9fa8575f+2c2d707e/model.neff filter=lfs diff=lfs merge=lfs -text
|
1747 |
+
neuronxcc-2.13.66.0+6dfecc895/MODULE_e9a9ed0a0d3bbe461a97+2c2d707e/model.neff filter=lfs diff=lfs merge=lfs -text
|
1748 |
+
neuronxcc-2.13.66.0+6dfecc895/MODULE_ea5c01ba138b0c3501c2+2c2d707e/model.neff filter=lfs diff=lfs merge=lfs -text
|
1749 |
+
neuronxcc-2.13.66.0+6dfecc895/MODULE_eea3cf695e9cf086ac91+2c2d707e/model.neff filter=lfs diff=lfs merge=lfs -text
|
1750 |
+
neuronxcc-2.13.66.0+6dfecc895/MODULE_f1f512eea34f52b7f8d9+2c2d707e/model.neff filter=lfs diff=lfs merge=lfs -text
|
1751 |
+
neuronxcc-2.13.66.0+6dfecc895/MODULE_f38d6a7d178ff69e2bb3+2c2d707e/model.neff filter=lfs diff=lfs merge=lfs -text
|
1752 |
+
neuronxcc-2.13.66.0+6dfecc895/MODULE_f412a6bebf32d04bf191+2c2d707e/model.neff filter=lfs diff=lfs merge=lfs -text
|
1753 |
+
neuronxcc-2.13.66.0+6dfecc895/MODULE_f7d991bf6082f2b30284+2c2d707e/model.neff filter=lfs diff=lfs merge=lfs -text
|
1754 |
+
neuronxcc-2.13.66.0+6dfecc895/MODULE_fb6a058e29b833424101+2c2d707e/model.neff filter=lfs diff=lfs merge=lfs -text
|
1755 |
+
neuronxcc-2.13.66.0+6dfecc895/MODULE_fd411c71433f0e6756af+2c2d707e/model.neff filter=lfs diff=lfs merge=lfs -text
|
neuronxcc-2.13.66.0+6dfecc895/0_REGISTRY/0.0.21.dev0/inference/bloom/hf-internal-testing/tiny-random-BloomForCausalLM/4eccd2e66477ff5ef5b3.json
ADDED
@@ -0,0 +1 @@
|
|
|
|
|
1 |
+
{"vocab_size": 1024, "hidden_size": 32, "n_layer": 5, "n_head": 4, "layer_norm_epsilon": 1e-05, "initializer_range": 0.02, "use_cache": true, "pretraining_tp": 1, "apply_residual_connection_post_layernorm": false, "hidden_dropout": 0.1, "attention_dropout": 0.1, "bos_token_id": 1, "eos_token_id": 2, "slow_but_exact": true, "torch_dtype": "float32", "is_decoder": true, "architectures": ["BloomForCausalLM"], "id2label": {"0": "LABEL_0", "1": "LABEL_1", "2": "LABEL_2"}, "label2id": {"LABEL_0": 0, "LABEL_1": 1, "LABEL_2": 2}, "pad_token_id": 3, "dtype": "float32", "gradient_checkpointing": false, "model_type": "bloom", "n_positions": 512, "neuron": {"auto_cast_type": "fp32", "batch_size": 2, "checkpoint_id": "hf-internal-testing/tiny-random-BloomForCausalLM", "checkpoint_revision": "0f4f06f162cd67d34d03ee156484e4001d468500", "compiler_type": "neuronx-cc", "compiler_version": "2.13.66.0+6dfecc895", "num_cores": 2, "sequence_length": 100, "task": "text-generation"}, "seq_length": 7, "type_vocab_size": 16}
|
neuronxcc-2.13.66.0+6dfecc895/0_REGISTRY/0.0.21.dev0/inference/bloom/hf-internal-testing/tiny-random-BloomForCausalLM/d635309efcd921a3a3f6.json
ADDED
@@ -0,0 +1 @@
|
|
|
|
|
1 |
+
{"vocab_size": 1024, "hidden_size": 32, "n_layer": 5, "n_head": 4, "layer_norm_epsilon": 1e-05, "initializer_range": 0.02, "use_cache": true, "pretraining_tp": 1, "apply_residual_connection_post_layernorm": false, "hidden_dropout": 0.1, "attention_dropout": 0.1, "bos_token_id": 1, "eos_token_id": 2, "slow_but_exact": true, "torch_dtype": "float32", "is_decoder": true, "architectures": ["BloomForCausalLM"], "id2label": {"0": "LABEL_0", "1": "LABEL_1", "2": "LABEL_2"}, "label2id": {"LABEL_0": 0, "LABEL_1": 1, "LABEL_2": 2}, "pad_token_id": 3, "dtype": "float32", "gradient_checkpointing": false, "model_type": "bloom", "n_positions": 512, "seq_length": 7, "type_vocab_size": 16, "neuron": {"task": "text-generation", "batch_size": 2, "num_cores": 2, "auto_cast_type": "fp32", "sequence_length": 100, "compiler_type": "neuronx-cc", "compiler_version": "2.13.66.0+6dfecc895", "checkpoint_id": "hf-internal-testing/tiny-random-BloomForCausalLM", "checkpoint_revision": "0f4f06f162cd67d34d03ee156484e4001d468500"}}
|
neuronxcc-2.13.66.0+6dfecc895/0_REGISTRY/0.0.21.dev0/inference/gpt2/gpt2/2c3a47e6fca2fd6ac3cf.json
ADDED
@@ -0,0 +1 @@
|
|
|
|
|
1 |
+
{"vocab_size": 50257, "n_positions": 1024, "n_embd": 768, "n_layer": 12, "n_head": 12, "n_inner": null, "activation_function": "gelu_new", "resid_pdrop": 0.1, "embd_pdrop": 0.1, "attn_pdrop": 0.1, "layer_norm_epsilon": 1e-05, "initializer_range": 0.02, "summary_type": "cls_index", "summary_use_proj": true, "summary_activation": null, "summary_first_dropout": 0.1, "summary_proj_to_labels": true, "scale_attn_weights": true, "use_cache": true, "scale_attn_by_inverse_layer_idx": false, "reorder_and_upcast_attn": false, "bos_token_id": 50256, "eos_token_id": 50256, "architectures": ["GPT2LMHeadModel"], "task_specific_params": {"text-generation": {"do_sample": true, "max_length": 50}}, "model_type": "gpt2", "n_ctx": 1024, "neuron": {"task": "text-generation", "batch_size": 4, "num_cores": 2, "auto_cast_type": "fp32", "sequence_length": 1024, "compiler_type": "neuronx-cc", "compiler_version": "2.13.66.0+6dfecc895", "checkpoint_id": "gpt2", "checkpoint_revision": "607a30d783dfa663caf39e06633721c8d4cfcd7e"}}
|
neuronxcc-2.13.66.0+6dfecc895/0_REGISTRY/0.0.21.dev0/inference/gpt2/hf-internal-testing/tiny-random-gpt2/2037d1be0146ceb8d639.json
ADDED
@@ -0,0 +1 @@
|
|
|
|
|
1 |
+
{"vocab_size": 1000, "n_positions": 512, "n_embd": 32, "n_layer": 5, "n_head": 4, "n_inner": null, "activation_function": "gelu_new", "resid_pdrop": 0.1, "embd_pdrop": 0.1, "attn_pdrop": 0.1, "layer_norm_epsilon": 1e-05, "initializer_range": 0.02, "summary_type": "cls_index", "summary_use_proj": true, "summary_activation": null, "summary_first_dropout": 0.1, "summary_proj_to_labels": true, "scale_attn_weights": true, "use_cache": true, "scale_attn_by_inverse_layer_idx": false, "reorder_and_upcast_attn": false, "bos_token_id": 98, "eos_token_id": 98, "pad_token_id": 98, "attention_probs_dropout_prob": 0.1, "gradient_checkpointing": false, "hidden_act": "gelu", "hidden_dropout_prob": 0.1, "intermediate_size": 37, "model_type": "gpt2", "n_ctx": 512, "neuron": {"auto_cast_type": "fp32", "batch_size": 2, "checkpoint_id": "hf-internal-testing/tiny-random-gpt2", "checkpoint_revision": "91c0fe31d692dd8448d9bc06e8d1877345009e3b", "compiler_type": "neuronx-cc", "compiler_version": "2.13.66.0+6dfecc895", "num_cores": 2, "sequence_length": 100, "task": "text-generation"}, "type_vocab_size": 16}
|
neuronxcc-2.13.66.0+6dfecc895/0_REGISTRY/0.0.21.dev0/inference/gpt2/hf-internal-testing/tiny-random-gpt2/3edc0caf8d3805a471bd.json
ADDED
@@ -0,0 +1 @@
|
|
|
|
|
1 |
+
{"vocab_size": 1000, "n_positions": 512, "n_embd": 32, "n_layer": 5, "n_head": 4, "n_inner": null, "activation_function": "gelu_new", "resid_pdrop": 0.1, "embd_pdrop": 0.1, "attn_pdrop": 0.1, "layer_norm_epsilon": 1e-05, "initializer_range": 0.02, "summary_type": "cls_index", "summary_use_proj": true, "summary_activation": null, "summary_first_dropout": 0.1, "summary_proj_to_labels": true, "scale_attn_weights": true, "use_cache": true, "scale_attn_by_inverse_layer_idx": false, "reorder_and_upcast_attn": false, "bos_token_id": 98, "eos_token_id": 98, "pad_token_id": 98, "attention_probs_dropout_prob": 0.1, "gradient_checkpointing": false, "hidden_act": "gelu", "hidden_dropout_prob": 0.1, "intermediate_size": 37, "model_type": "gpt2", "n_ctx": 512, "type_vocab_size": 16, "neuron": {"task": "text-generation", "batch_size": 2, "num_cores": 2, "auto_cast_type": "fp32", "sequence_length": 100, "compiler_type": "neuronx-cc", "compiler_version": "2.13.66.0+6dfecc895", "checkpoint_id": "hf-internal-testing/tiny-random-gpt2", "checkpoint_revision": "91c0fe31d692dd8448d9bc06e8d1877345009e3b"}}
|
neuronxcc-2.13.66.0+6dfecc895/0_REGISTRY/0.0.21.dev0/inference/llama/NousResearch/Llama-2-7b-chat-hf/6c7cfec071642e5fecee.json
ADDED
@@ -0,0 +1 @@
|
|
|
|
|
1 |
+
{"vocab_size": 32000, "max_position_embeddings": 4096, "hidden_size": 4096, "intermediate_size": 11008, "num_hidden_layers": 32, "num_attention_heads": 32, "num_key_value_heads": 32, "hidden_act": "silu", "initializer_range": 0.02, "rms_norm_eps": 1e-05, "pretraining_tp": 1, "use_cache": true, "rope_theta": 10000.0, "rope_scaling": null, "attention_bias": false, "attention_dropout": 0.0, "torch_dtype": "float16", "tie_word_embeddings": false, "architectures": ["LlamaForCausalLM"], "bos_token_id": 1, "pad_token_id": 0, "eos_token_id": 2, "model_type": "llama", "neuron": {"auto_cast_type": "fp16", "batch_size": 4, "checkpoint_id": "NousResearch/Llama-2-7b-chat-hf", "checkpoint_revision": "37892f30c23786c0d5367d80481fa0d9fba93cf8", "compiler_type": "neuronx-cc", "compiler_version": "2.13.66.0+6dfecc895", "num_cores": 24, "sequence_length": 4096, "task": "text-generation"}}
|
neuronxcc-2.13.66.0+6dfecc895/0_REGISTRY/0.0.21.dev0/inference/llama/NousResearch/Llama-2-7b-chat-hf/b9904e0c46d338b623d2.json
ADDED
@@ -0,0 +1 @@
|
|
|
|
|
1 |
+
{"vocab_size": 32000, "max_position_embeddings": 4096, "hidden_size": 4096, "intermediate_size": 11008, "num_hidden_layers": 32, "num_attention_heads": 32, "num_key_value_heads": 32, "hidden_act": "silu", "initializer_range": 0.02, "rms_norm_eps": 1e-05, "pretraining_tp": 1, "use_cache": true, "rope_theta": 10000.0, "rope_scaling": null, "attention_bias": false, "attention_dropout": 0.0, "torch_dtype": "float16", "tie_word_embeddings": false, "architectures": ["LlamaForCausalLM"], "bos_token_id": 1, "pad_token_id": 0, "eos_token_id": 2, "model_type": "llama", "neuron": {"task": "text-generation", "batch_size": 4, "num_cores": 24, "auto_cast_type": "fp16", "sequence_length": 4096, "compiler_type": "neuronx-cc", "compiler_version": "2.13.66.0+6dfecc895", "checkpoint_id": "NousResearch/Llama-2-7b-chat-hf", "checkpoint_revision": "37892f30c23786c0d5367d80481fa0d9fba93cf8"}}
|
neuronxcc-2.13.66.0+6dfecc895/0_REGISTRY/0.0.21.dev0/inference/llama/dacorvo/tiny-random-llama/7c595f241fa311d7b227.json
ADDED
@@ -0,0 +1 @@
|
|
|
|
|
1 |
+
{"vocab_size": 32000, "max_position_embeddings": 512, "hidden_size": 128, "intermediate_size": 256, "num_hidden_layers": 1, "num_attention_heads": 1, "num_key_value_heads": 1, "hidden_act": "silu", "initializer_range": 0.02, "rms_norm_eps": 1e-06, "pretraining_tp": 1, "use_cache": true, "rope_theta": 10000.0, "rope_scaling": null, "attention_bias": false, "attention_dropout": 0.0, "torch_dtype": "float32", "tie_word_embeddings": false, "architectures": ["LlamaForCausalLM"], "bos_token_id": 1, "eos_token_id": 2, "model_type": "llama", "neuron": {"auto_cast_type": "fp32", "batch_size": 2, "checkpoint_id": "dacorvo/tiny-random-llama", "checkpoint_revision": "7fdafd2fe6a2d31c6abb72ae60db606d8bb23196", "compiler_type": "neuronx-cc", "compiler_version": "2.13.66.0+6dfecc895", "num_cores": 2, "sequence_length": 100, "task": "text-generation"}}
|
neuronxcc-2.13.66.0+6dfecc895/0_REGISTRY/0.0.21.dev0/inference/llama/dacorvo/tiny-random-llama/b650b9d41c4386d36a89.json
ADDED
@@ -0,0 +1 @@
|
|
|
|
|
1 |
+
{"vocab_size": 32000, "max_position_embeddings": 512, "hidden_size": 128, "intermediate_size": 256, "num_hidden_layers": 1, "num_attention_heads": 1, "num_key_value_heads": 1, "hidden_act": "silu", "initializer_range": 0.02, "rms_norm_eps": 1e-06, "pretraining_tp": 1, "use_cache": true, "rope_theta": 10000.0, "rope_scaling": null, "attention_bias": false, "attention_dropout": 0.0, "torch_dtype": "float32", "tie_word_embeddings": false, "architectures": ["LlamaForCausalLM"], "bos_token_id": 1, "eos_token_id": 2, "model_type": "llama", "neuron": {"task": "text-generation", "batch_size": 2, "num_cores": 2, "auto_cast_type": "fp32", "sequence_length": 100, "compiler_type": "neuronx-cc", "compiler_version": "2.13.66.0+6dfecc895", "checkpoint_id": "dacorvo/tiny-random-llama", "checkpoint_revision": "7fdafd2fe6a2d31c6abb72ae60db606d8bb23196"}}
|
neuronxcc-2.13.66.0+6dfecc895/0_REGISTRY/0.0.21.dev0/inference/llama/meta-llama/Llama-2-7b-chat-hf/3a328b55c6445b520b8e.json
ADDED
@@ -0,0 +1 @@
|
|
|
|
|
1 |
+
{"vocab_size": 32000, "max_position_embeddings": 4096, "hidden_size": 4096, "intermediate_size": 11008, "num_hidden_layers": 32, "num_attention_heads": 32, "num_key_value_heads": 32, "hidden_act": "silu", "initializer_range": 0.02, "rms_norm_eps": 1e-05, "pretraining_tp": 1, "use_cache": true, "rope_theta": 10000.0, "rope_scaling": null, "attention_bias": false, "attention_dropout": 0.0, "torch_dtype": "float16", "tie_word_embeddings": false, "architectures": ["LlamaForCausalLM"], "bos_token_id": 1, "eos_token_id": 2, "model_type": "llama", "neuron": {"task": "text-generation", "batch_size": 16, "num_cores": 24, "auto_cast_type": "fp16", "sequence_length": 4096, "compiler_type": "neuronx-cc", "compiler_version": "2.13.66.0+6dfecc895", "checkpoint_id": "meta-llama/Llama-2-7b-chat-hf", "checkpoint_revision": "92011f62d7604e261f748ec0cfe6329f31193e33"}}
|
neuronxcc-2.13.66.0+6dfecc895/0_REGISTRY/0.0.21.dev0/inference/llama/meta-llama/Llama-2-7b-chat-hf/44c81784d74cd4713969.json
ADDED
@@ -0,0 +1 @@
|
|
|
|
|
1 |
+
{"vocab_size": 32000, "max_position_embeddings": 4096, "hidden_size": 4096, "intermediate_size": 11008, "num_hidden_layers": 32, "num_attention_heads": 32, "num_key_value_heads": 32, "hidden_act": "silu", "initializer_range": 0.02, "rms_norm_eps": 1e-05, "pretraining_tp": 1, "use_cache": true, "rope_theta": 10000.0, "rope_scaling": null, "attention_bias": false, "attention_dropout": 0.0, "torch_dtype": "float16", "tie_word_embeddings": false, "architectures": ["LlamaForCausalLM"], "bos_token_id": 1, "eos_token_id": 2, "model_type": "llama", "neuron": {"task": "text-generation", "batch_size": 8, "num_cores": 24, "auto_cast_type": "fp16", "sequence_length": 4096, "compiler_type": "neuronx-cc", "compiler_version": "2.13.66.0+6dfecc895", "checkpoint_id": "meta-llama/Llama-2-7b-chat-hf", "checkpoint_revision": "92011f62d7604e261f748ec0cfe6329f31193e33"}}
|
neuronxcc-2.13.66.0+6dfecc895/0_REGISTRY/0.0.21.dev0/inference/llama/meta-llama/Llama-2-7b-chat-hf/58eec5854a9f564f8a27.json
ADDED
@@ -0,0 +1 @@
|
|
|
|
|
1 |
+
{"vocab_size": 32000, "max_position_embeddings": 4096, "hidden_size": 4096, "intermediate_size": 11008, "num_hidden_layers": 32, "num_attention_heads": 32, "num_key_value_heads": 32, "hidden_act": "silu", "initializer_range": 0.02, "rms_norm_eps": 1e-05, "pretraining_tp": 1, "use_cache": true, "rope_theta": 10000.0, "rope_scaling": null, "attention_bias": false, "attention_dropout": 0.0, "torch_dtype": "float16", "tie_word_embeddings": false, "architectures": ["LlamaForCausalLM"], "bos_token_id": 1, "eos_token_id": 2, "model_type": "llama", "neuron": {"auto_cast_type": "fp16", "batch_size": 16, "checkpoint_id": "meta-llama/Llama-2-7b-chat-hf", "checkpoint_revision": "92011f62d7604e261f748ec0cfe6329f31193e33", "compiler_type": "neuronx-cc", "compiler_version": "2.13.66.0+6dfecc895", "num_cores": 24, "sequence_length": 4096, "task": "text-generation"}}
|
neuronxcc-2.13.66.0+6dfecc895/0_REGISTRY/0.0.21.dev0/inference/llama/meta-llama/Llama-2-7b-chat-hf/6a314b3d69ca329d2711.json
ADDED
@@ -0,0 +1 @@
|
|
|
|
|
1 |
+
{"vocab_size": 32000, "max_position_embeddings": 4096, "hidden_size": 4096, "intermediate_size": 11008, "num_hidden_layers": 32, "num_attention_heads": 32, "num_key_value_heads": 32, "hidden_act": "silu", "initializer_range": 0.02, "rms_norm_eps": 1e-05, "pretraining_tp": 1, "use_cache": true, "rope_theta": 10000.0, "rope_scaling": null, "attention_bias": false, "attention_dropout": 0.0, "torch_dtype": "float16", "tie_word_embeddings": false, "architectures": ["LlamaForCausalLM"], "bos_token_id": 1, "eos_token_id": 2, "model_type": "llama", "neuron": {"task": "text-generation", "batch_size": 1, "num_cores": 24, "auto_cast_type": "fp16", "sequence_length": 4096, "compiler_type": "neuronx-cc", "compiler_version": "2.13.66.0+6dfecc895", "checkpoint_id": "meta-llama/Llama-2-7b-chat-hf", "checkpoint_revision": "92011f62d7604e261f748ec0cfe6329f31193e33"}}
|
neuronxcc-2.13.66.0+6dfecc895/0_REGISTRY/0.0.21.dev0/inference/llama/meta-llama/Llama-2-7b-chat-hf/7a0a7caf91982c3c67b0.json
ADDED
@@ -0,0 +1 @@
|
|
|
|
|
1 |
+
{"vocab_size": 32000, "max_position_embeddings": 4096, "hidden_size": 4096, "intermediate_size": 11008, "num_hidden_layers": 32, "num_attention_heads": 32, "num_key_value_heads": 32, "hidden_act": "silu", "initializer_range": 0.02, "rms_norm_eps": 1e-05, "pretraining_tp": 1, "use_cache": true, "rope_theta": 10000.0, "rope_scaling": null, "attention_bias": false, "attention_dropout": 0.0, "torch_dtype": "float16", "tie_word_embeddings": false, "architectures": ["LlamaForCausalLM"], "bos_token_id": 1, "eos_token_id": 2, "model_type": "llama", "neuron": {"auto_cast_type": "fp16", "batch_size": 4, "checkpoint_id": "meta-llama/Llama-2-7b-chat-hf", "checkpoint_revision": "92011f62d7604e261f748ec0cfe6329f31193e33", "compiler_type": "neuronx-cc", "compiler_version": "2.13.66.0+6dfecc895", "num_cores": 24, "sequence_length": 4096, "task": "text-generation"}}
|
neuronxcc-2.13.66.0+6dfecc895/0_REGISTRY/0.0.21.dev0/inference/llama/meta-llama/Llama-2-7b-chat-hf/ad1f7e0ca3ab366f91d4.json
ADDED
@@ -0,0 +1 @@
|
|
|
|
|
1 |
+
{"vocab_size": 32000, "max_position_embeddings": 4096, "hidden_size": 4096, "intermediate_size": 11008, "num_hidden_layers": 32, "num_attention_heads": 32, "num_key_value_heads": 32, "hidden_act": "silu", "initializer_range": 0.02, "rms_norm_eps": 1e-05, "pretraining_tp": 1, "use_cache": true, "rope_theta": 10000.0, "rope_scaling": null, "attention_bias": false, "attention_dropout": 0.0, "torch_dtype": "float16", "tie_word_embeddings": false, "architectures": ["LlamaForCausalLM"], "bos_token_id": 1, "eos_token_id": 2, "model_type": "llama", "neuron": {"auto_cast_type": "fp16", "batch_size": 8, "checkpoint_id": "meta-llama/Llama-2-7b-chat-hf", "checkpoint_revision": "92011f62d7604e261f748ec0cfe6329f31193e33", "compiler_type": "neuronx-cc", "compiler_version": "2.13.66.0+6dfecc895", "num_cores": 24, "sequence_length": 4096, "task": "text-generation"}}
|
neuronxcc-2.13.66.0+6dfecc895/0_REGISTRY/0.0.21.dev0/inference/llama/meta-llama/Llama-2-7b-chat-hf/dd0ae8e9d982a2ed293b.json
ADDED
@@ -0,0 +1 @@
|
|
|
|
|
1 |
+
{"vocab_size": 32000, "max_position_embeddings": 4096, "hidden_size": 4096, "intermediate_size": 11008, "num_hidden_layers": 32, "num_attention_heads": 32, "num_key_value_heads": 32, "hidden_act": "silu", "initializer_range": 0.02, "rms_norm_eps": 1e-05, "pretraining_tp": 1, "use_cache": true, "rope_theta": 10000.0, "rope_scaling": null, "attention_bias": false, "attention_dropout": 0.0, "torch_dtype": "float16", "tie_word_embeddings": false, "architectures": ["LlamaForCausalLM"], "bos_token_id": 1, "eos_token_id": 2, "model_type": "llama", "neuron": {"task": "text-generation", "batch_size": 4, "num_cores": 24, "auto_cast_type": "fp16", "sequence_length": 4096, "compiler_type": "neuronx-cc", "compiler_version": "2.13.66.0+6dfecc895", "checkpoint_id": "meta-llama/Llama-2-7b-chat-hf", "checkpoint_revision": "92011f62d7604e261f748ec0cfe6329f31193e33"}}
|
neuronxcc-2.13.66.0+6dfecc895/0_REGISTRY/0.0.21.dev0/inference/llama/meta-llama/Llama-2-7b-chat-hf/efa4119dabd9c1f8277f.json
ADDED
@@ -0,0 +1 @@
|
|
|
|
|
1 |
+
{"vocab_size": 32000, "max_position_embeddings": 4096, "hidden_size": 4096, "intermediate_size": 11008, "num_hidden_layers": 32, "num_attention_heads": 32, "num_key_value_heads": 32, "hidden_act": "silu", "initializer_range": 0.02, "rms_norm_eps": 1e-05, "pretraining_tp": 1, "use_cache": true, "rope_theta": 10000.0, "rope_scaling": null, "attention_bias": false, "attention_dropout": 0.0, "torch_dtype": "float16", "tie_word_embeddings": false, "architectures": ["LlamaForCausalLM"], "bos_token_id": 1, "eos_token_id": 2, "model_type": "llama", "neuron": {"auto_cast_type": "fp16", "batch_size": 1, "checkpoint_id": "meta-llama/Llama-2-7b-chat-hf", "checkpoint_revision": "92011f62d7604e261f748ec0cfe6329f31193e33", "compiler_type": "neuronx-cc", "compiler_version": "2.13.66.0+6dfecc895", "num_cores": 24, "sequence_length": 4096, "task": "text-generation"}}
|
neuronxcc-2.13.66.0+6dfecc895/0_REGISTRY/0.0.21.dev0/inference/mistral/dacorvo/tiny-random-MistralForCausalLM/43c8baaa3d89767fe7f2.json
ADDED
@@ -0,0 +1 @@
|
|
|
|
|
1 |
+
{"vocab_size": 32000, "max_position_embeddings": 512, "hidden_size": 32, "intermediate_size": 37, "num_hidden_layers": 2, "num_attention_heads": 4, "sliding_window": 4096, "num_key_value_heads": 2, "hidden_act": "gelu", "initializer_range": 0.02, "rms_norm_eps": 1e-06, "use_cache": true, "rope_theta": 10000.0, "attention_dropout": 0.0, "torch_dtype": "float32", "tie_word_embeddings": false, "is_decoder": true, "architectures": ["MistralForCausalLM"], "bos_token_id": 1, "pad_token_id": 0, "eos_token_id": 2, "attention_probs_dropout_prob": 0.1, "hidden_dropout_prob": 0.1, "model_type": "mistral", "neuron": {"auto_cast_type": "fp32", "batch_size": 2, "checkpoint_id": "dacorvo/tiny-random-MistralForCausalLM", "checkpoint_revision": "81d453e3c8985649e9ee3d4c9378461029d1c73a", "compiler_type": "neuronx-cc", "compiler_version": "2.13.66.0+6dfecc895", "num_cores": 2, "sequence_length": 100, "task": "text-generation"}, "type_vocab_size": 16}
|
neuronxcc-2.13.66.0+6dfecc895/0_REGISTRY/0.0.21.dev0/inference/mistral/dacorvo/tiny-random-MistralForCausalLM/f9b607c62a5393e0b42f.json
ADDED
@@ -0,0 +1 @@
|
|
|
|
|
1 |
+
{"vocab_size": 32000, "max_position_embeddings": 512, "hidden_size": 32, "intermediate_size": 37, "num_hidden_layers": 2, "num_attention_heads": 4, "sliding_window": 4096, "num_key_value_heads": 2, "hidden_act": "gelu", "initializer_range": 0.02, "rms_norm_eps": 1e-06, "use_cache": true, "rope_theta": 10000.0, "attention_dropout": 0.0, "torch_dtype": "float32", "tie_word_embeddings": false, "is_decoder": true, "architectures": ["MistralForCausalLM"], "bos_token_id": 1, "pad_token_id": 0, "eos_token_id": 2, "attention_probs_dropout_prob": 0.1, "hidden_dropout_prob": 0.1, "model_type": "mistral", "type_vocab_size": 16, "neuron": {"task": "text-generation", "batch_size": 2, "num_cores": 2, "auto_cast_type": "fp32", "sequence_length": 100, "compiler_type": "neuronx-cc", "compiler_version": "2.13.66.0+6dfecc895", "checkpoint_id": "dacorvo/tiny-random-MistralForCausalLM", "checkpoint_revision": "81d453e3c8985649e9ee3d4c9378461029d1c73a"}}
|
neuronxcc-2.13.66.0+6dfecc895/0_REGISTRY/0.0.21.dev0/inference/mistral/mistralai/Mistral-7B-Instruct-v0.1/7a99245071db1d4e48e6.json
ADDED
@@ -0,0 +1 @@
|
|
|
|
|
1 |
+
{"vocab_size": 32000, "max_position_embeddings": 32768, "hidden_size": 4096, "intermediate_size": 14336, "num_hidden_layers": 32, "num_attention_heads": 32, "sliding_window": 4096, "num_key_value_heads": 8, "hidden_act": "silu", "initializer_range": 0.02, "rms_norm_eps": 1e-05, "use_cache": true, "rope_theta": 10000.0, "attention_dropout": 0.0, "torch_dtype": "bfloat16", "tie_word_embeddings": false, "architectures": ["MistralForCausalLM"], "bos_token_id": 1, "eos_token_id": 2, "model_type": "mistral", "neuron": {"auto_cast_type": "fp16", "batch_size": 1, "checkpoint_id": "mistralai/Mistral-7B-Instruct-v0.1", "checkpoint_revision": "73068f3702d050a2fd5aa2ca1e612e5036429398", "compiler_type": "neuronx-cc", "compiler_version": "2.13.66.0+6dfecc895", "num_cores": 24, "sequence_length": 4096, "task": "text-generation"}}
|
neuronxcc-2.13.66.0+6dfecc895/0_REGISTRY/0.0.21.dev0/inference/mistral/mistralai/Mistral-7B-Instruct-v0.1/b0a5a2e1e8dd02dbe923.json
ADDED
@@ -0,0 +1 @@
|
|
|
|
|
1 |
+
{"vocab_size": 32000, "max_position_embeddings": 32768, "hidden_size": 4096, "intermediate_size": 14336, "num_hidden_layers": 32, "num_attention_heads": 32, "sliding_window": 4096, "num_key_value_heads": 8, "hidden_act": "silu", "initializer_range": 0.02, "rms_norm_eps": 1e-05, "use_cache": true, "rope_theta": 10000.0, "attention_dropout": 0.0, "torch_dtype": "bfloat16", "tie_word_embeddings": false, "architectures": ["MistralForCausalLM"], "bos_token_id": 1, "eos_token_id": 2, "model_type": "mistral", "neuron": {"task": "text-generation", "batch_size": 1, "num_cores": 24, "auto_cast_type": "fp16", "sequence_length": 4096, "compiler_type": "neuronx-cc", "compiler_version": "2.13.66.0+6dfecc895", "checkpoint_id": "mistralai/Mistral-7B-Instruct-v0.1", "checkpoint_revision": "73068f3702d050a2fd5aa2ca1e612e5036429398"}}
|
neuronxcc-2.13.66.0+6dfecc895/0_REGISTRY/0.0.21.dev0/inference/mistral/mistralai/Mistral-7B-Instruct-v0.2/1c4540cc29dcb3b99ff8.json
ADDED
@@ -0,0 +1 @@
|
|
|
|
|
1 |
+
{"vocab_size": 32000, "max_position_embeddings": 32768, "hidden_size": 4096, "intermediate_size": 14336, "num_hidden_layers": 32, "num_attention_heads": 32, "sliding_window": null, "num_key_value_heads": 8, "hidden_act": "silu", "initializer_range": 0.02, "rms_norm_eps": 1e-05, "use_cache": true, "rope_theta": 1000000.0, "attention_dropout": 0.0, "torch_dtype": "bfloat16", "tie_word_embeddings": false, "architectures": ["MistralForCausalLM"], "bos_token_id": 1, "eos_token_id": 2, "model_type": "mistral", "neuron": {"auto_cast_type": "bf16", "batch_size": 1, "checkpoint_id": "mistralai/Mistral-7B-Instruct-v0.2", "checkpoint_revision": "41b61a33a2483885c981aa79e0df6b32407ed873", "compiler_type": "neuronx-cc", "compiler_version": "2.13.66.0+6dfecc895", "num_cores": 8, "sequence_length": 4096, "task": "text-generation"}}
|
neuronxcc-2.13.66.0+6dfecc895/0_REGISTRY/0.0.21.dev0/inference/mistral/mistralai/Mistral-7B-Instruct-v0.2/2d6c1d9513debbcc5f7f.json
ADDED
@@ -0,0 +1 @@
|
|
|
|
|
1 |
+
{"vocab_size": 32000, "max_position_embeddings": 32768, "hidden_size": 4096, "intermediate_size": 14336, "num_hidden_layers": 32, "num_attention_heads": 32, "sliding_window": null, "num_key_value_heads": 8, "hidden_act": "silu", "initializer_range": 0.02, "rms_norm_eps": 1e-05, "use_cache": true, "rope_theta": 1000000.0, "attention_dropout": 0.0, "torch_dtype": "bfloat16", "tie_word_embeddings": false, "architectures": ["MistralForCausalLM"], "bos_token_id": 1, "eos_token_id": 2, "model_type": "mistral", "neuron": {"task": "text-generation", "batch_size": 1, "num_cores": 8, "auto_cast_type": "bf16", "sequence_length": 4096, "compiler_type": "neuronx-cc", "compiler_version": "2.13.66.0+6dfecc895", "checkpoint_id": "mistralai/Mistral-7B-Instruct-v0.2", "checkpoint_revision": "41b61a33a2483885c981aa79e0df6b32407ed873"}}
|
neuronxcc-2.13.66.0+6dfecc895/0_REGISTRY/0.0.21.dev0/inference/mistral/mistralai/Mistral-7B-Instruct-v0.2/41284a607429208347c0.json
ADDED
@@ -0,0 +1 @@
|
|
|
|
|
1 |
+
{"vocab_size": 32000, "max_position_embeddings": 32768, "hidden_size": 4096, "intermediate_size": 14336, "num_hidden_layers": 32, "num_attention_heads": 32, "sliding_window": null, "num_key_value_heads": 8, "hidden_act": "silu", "initializer_range": 0.02, "rms_norm_eps": 1e-05, "use_cache": true, "rope_theta": 1000000.0, "attention_dropout": 0.0, "torch_dtype": "bfloat16", "tie_word_embeddings": false, "architectures": ["MistralForCausalLM"], "bos_token_id": 1, "eos_token_id": 2, "model_type": "mistral", "neuron": {"task": "text-generation", "batch_size": 16, "num_cores": 8, "auto_cast_type": "bf16", "sequence_length": 4096, "compiler_type": "neuronx-cc", "compiler_version": "2.13.66.0+6dfecc895", "checkpoint_id": "mistralai/Mistral-7B-Instruct-v0.2", "checkpoint_revision": "41b61a33a2483885c981aa79e0df6b32407ed873"}}
|
neuronxcc-2.13.66.0+6dfecc895/0_REGISTRY/0.0.21.dev0/inference/mistral/mistralai/Mistral-7B-Instruct-v0.2/4620b76c4f39e71a8459.json
ADDED
@@ -0,0 +1 @@
|
|
|
|
|
1 |
+
{"vocab_size": 32000, "max_position_embeddings": 32768, "hidden_size": 4096, "intermediate_size": 14336, "num_hidden_layers": 32, "num_attention_heads": 32, "sliding_window": null, "num_key_value_heads": 8, "hidden_act": "silu", "initializer_range": 0.02, "rms_norm_eps": 1e-05, "use_cache": true, "rope_theta": 1000000.0, "attention_dropout": 0.0, "torch_dtype": "bfloat16", "tie_word_embeddings": false, "architectures": ["MistralForCausalLM"], "bos_token_id": 1, "eos_token_id": 2, "model_type": "mistral", "neuron": {"task": "text-generation", "batch_size": 1, "num_cores": 24, "auto_cast_type": "fp16", "sequence_length": 4096, "compiler_type": "neuronx-cc", "compiler_version": "2.13.66.0+6dfecc895", "checkpoint_id": "mistralai/Mistral-7B-Instruct-v0.2", "checkpoint_revision": "41b61a33a2483885c981aa79e0df6b32407ed873"}}
|
neuronxcc-2.13.66.0+6dfecc895/0_REGISTRY/0.0.21.dev0/inference/mistral/mistralai/Mistral-7B-Instruct-v0.2/4fcf5530fe99f6f4f3e6.json
ADDED
@@ -0,0 +1 @@
|
|
|
|
|
1 |
+
{"vocab_size": 32000, "max_position_embeddings": 32768, "hidden_size": 4096, "intermediate_size": 14336, "num_hidden_layers": 32, "num_attention_heads": 32, "sliding_window": null, "num_key_value_heads": 8, "hidden_act": "silu", "initializer_range": 0.02, "rms_norm_eps": 1e-05, "use_cache": true, "rope_theta": 1000000.0, "attention_dropout": 0.0, "torch_dtype": "bfloat16", "tie_word_embeddings": false, "architectures": ["MistralForCausalLM"], "bos_token_id": 1, "eos_token_id": 2, "model_type": "mistral", "neuron": {"auto_cast_type": "bf16", "batch_size": 8, "checkpoint_id": "mistralai/Mistral-7B-Instruct-v0.2", "checkpoint_revision": "41b61a33a2483885c981aa79e0df6b32407ed873", "compiler_type": "neuronx-cc", "compiler_version": "2.13.66.0+6dfecc895", "num_cores": 8, "sequence_length": 4096, "task": "text-generation"}}
|
neuronxcc-2.13.66.0+6dfecc895/0_REGISTRY/0.0.21.dev0/inference/mistral/mistralai/Mistral-7B-Instruct-v0.2/5c9bfea69bb260a59b0c.json
ADDED
@@ -0,0 +1 @@
|
|
|
|
|
1 |
+
{"vocab_size": 32000, "max_position_embeddings": 32768, "hidden_size": 4096, "intermediate_size": 14336, "num_hidden_layers": 32, "num_attention_heads": 32, "sliding_window": null, "num_key_value_heads": 8, "hidden_act": "silu", "initializer_range": 0.02, "rms_norm_eps": 1e-05, "use_cache": true, "rope_theta": 1000000.0, "attention_dropout": 0.0, "torch_dtype": "bfloat16", "tie_word_embeddings": false, "architectures": ["MistralForCausalLM"], "bos_token_id": 1, "eos_token_id": 2, "model_type": "mistral", "neuron": {"auto_cast_type": "bf16", "batch_size": 16, "checkpoint_id": "mistralai/Mistral-7B-Instruct-v0.2", "checkpoint_revision": "41b61a33a2483885c981aa79e0df6b32407ed873", "compiler_type": "neuronx-cc", "compiler_version": "2.13.66.0+6dfecc895", "num_cores": 8, "sequence_length": 4096, "task": "text-generation"}}
|
neuronxcc-2.13.66.0+6dfecc895/0_REGISTRY/0.0.21.dev0/inference/mistral/mistralai/Mistral-7B-Instruct-v0.2/698b844f07e07829c78e.json
ADDED
@@ -0,0 +1 @@
|
|
|
|
|
1 |
+
{"vocab_size": 32000, "max_position_embeddings": 32768, "hidden_size": 4096, "intermediate_size": 14336, "num_hidden_layers": 32, "num_attention_heads": 32, "sliding_window": null, "num_key_value_heads": 8, "hidden_act": "silu", "initializer_range": 0.02, "rms_norm_eps": 1e-05, "use_cache": true, "rope_theta": 1000000.0, "attention_dropout": 0.0, "torch_dtype": "bfloat16", "tie_word_embeddings": false, "architectures": ["MistralForCausalLM"], "bos_token_id": 1, "eos_token_id": 2, "model_type": "mistral", "neuron": {"auto_cast_type": "fp16", "batch_size": 1, "checkpoint_id": "mistralai/Mistral-7B-Instruct-v0.2", "checkpoint_revision": "41b61a33a2483885c981aa79e0df6b32407ed873", "compiler_type": "neuronx-cc", "compiler_version": "2.13.66.0+6dfecc895", "num_cores": 24, "sequence_length": 4096, "task": "text-generation"}}
|
neuronxcc-2.13.66.0+6dfecc895/0_REGISTRY/0.0.21.dev0/inference/mistral/mistralai/Mistral-7B-Instruct-v0.2/a2a8d3c93367a33b250c.json
ADDED
@@ -0,0 +1 @@
|
|
|
|
|
1 |
+
{"vocab_size": 32000, "max_position_embeddings": 32768, "hidden_size": 4096, "intermediate_size": 14336, "num_hidden_layers": 32, "num_attention_heads": 32, "sliding_window": null, "num_key_value_heads": 8, "hidden_act": "silu", "initializer_range": 0.02, "rms_norm_eps": 1e-05, "use_cache": true, "rope_theta": 1000000.0, "attention_dropout": 0.0, "torch_dtype": "bfloat16", "tie_word_embeddings": false, "architectures": ["MistralForCausalLM"], "bos_token_id": 1, "eos_token_id": 2, "model_type": "mistral", "neuron": {"task": "text-generation", "batch_size": 8, "num_cores": 8, "auto_cast_type": "bf16", "sequence_length": 4096, "compiler_type": "neuronx-cc", "compiler_version": "2.13.66.0+6dfecc895", "checkpoint_id": "mistralai/Mistral-7B-Instruct-v0.2", "checkpoint_revision": "41b61a33a2483885c981aa79e0df6b32407ed873"}}
|
neuronxcc-2.13.66.0+6dfecc895/0_REGISTRY/0.0.21.dev0/inference/mistral/mistralai/Mistral-7B-Instruct-v0.2/a61d8247a227bdcc16f2.json
ADDED
@@ -0,0 +1 @@
|
|
|
|
|
1 |
+
{"vocab_size": 32000, "max_position_embeddings": 32768, "hidden_size": 4096, "intermediate_size": 14336, "num_hidden_layers": 32, "num_attention_heads": 32, "sliding_window": null, "num_key_value_heads": 8, "hidden_act": "silu", "initializer_range": 0.02, "rms_norm_eps": 1e-05, "use_cache": true, "rope_theta": 1000000.0, "attention_dropout": 0.0, "torch_dtype": "bfloat16", "tie_word_embeddings": false, "architectures": ["MistralForCausalLM"], "bos_token_id": 1, "eos_token_id": 2, "model_type": "mistral", "neuron": {"task": "text-generation", "batch_size": 32, "num_cores": 8, "auto_cast_type": "bf16", "sequence_length": 4096, "compiler_type": "neuronx-cc", "compiler_version": "2.13.66.0+6dfecc895", "checkpoint_id": "mistralai/Mistral-7B-Instruct-v0.2", "checkpoint_revision": "41b61a33a2483885c981aa79e0df6b32407ed873"}}
|
neuronxcc-2.13.66.0+6dfecc895/0_REGISTRY/0.0.21.dev0/inference/mistral/mistralai/Mistral-7B-Instruct-v0.2/ca9ec6b16e1a62dbd649.json
ADDED
@@ -0,0 +1 @@
|
|
|
|
|
1 |
+
{"vocab_size": 32000, "max_position_embeddings": 32768, "hidden_size": 4096, "intermediate_size": 14336, "num_hidden_layers": 32, "num_attention_heads": 32, "sliding_window": null, "num_key_value_heads": 8, "hidden_act": "silu", "initializer_range": 0.02, "rms_norm_eps": 1e-05, "use_cache": true, "rope_theta": 1000000.0, "attention_dropout": 0.0, "torch_dtype": "bfloat16", "tie_word_embeddings": false, "architectures": ["MistralForCausalLM"], "bos_token_id": 1, "eos_token_id": 2, "model_type": "mistral", "neuron": {"task": "text-generation", "batch_size": 4, "num_cores": 8, "auto_cast_type": "bf16", "sequence_length": 4096, "compiler_type": "neuronx-cc", "compiler_version": "2.13.66.0+6dfecc895", "checkpoint_id": "mistralai/Mistral-7B-Instruct-v0.2", "checkpoint_revision": "41b61a33a2483885c981aa79e0df6b32407ed873"}}
|
neuronxcc-2.13.66.0+6dfecc895/0_REGISTRY/0.0.21.dev0/inference/mistral/mistralai/Mistral-7B-Instruct-v0.2/d75fbb50baa6a3b15792.json
ADDED
@@ -0,0 +1 @@
|
|
|
|
|
1 |
+
{"vocab_size": 32000, "max_position_embeddings": 32768, "hidden_size": 4096, "intermediate_size": 14336, "num_hidden_layers": 32, "num_attention_heads": 32, "sliding_window": null, "num_key_value_heads": 8, "hidden_act": "silu", "initializer_range": 0.02, "rms_norm_eps": 1e-05, "use_cache": true, "rope_theta": 1000000.0, "attention_dropout": 0.0, "torch_dtype": "bfloat16", "tie_word_embeddings": false, "architectures": ["MistralForCausalLM"], "bos_token_id": 1, "eos_token_id": 2, "model_type": "mistral", "neuron": {"auto_cast_type": "bf16", "batch_size": 4, "checkpoint_id": "mistralai/Mistral-7B-Instruct-v0.2", "checkpoint_revision": "41b61a33a2483885c981aa79e0df6b32407ed873", "compiler_type": "neuronx-cc", "compiler_version": "2.13.66.0+6dfecc895", "num_cores": 8, "sequence_length": 4096, "task": "text-generation"}}
|
neuronxcc-2.13.66.0+6dfecc895/0_REGISTRY/0.0.21.dev0/inference/mistral/mistralai/Mistral-7B-Instruct-v0.2/de7f8e3eb7e911ee8559.json
ADDED
@@ -0,0 +1 @@
|
|
|
|
|
1 |
+
{"vocab_size": 32000, "max_position_embeddings": 32768, "hidden_size": 4096, "intermediate_size": 14336, "num_hidden_layers": 32, "num_attention_heads": 32, "sliding_window": null, "num_key_value_heads": 8, "hidden_act": "silu", "initializer_range": 0.02, "rms_norm_eps": 1e-05, "use_cache": true, "rope_theta": 1000000.0, "attention_dropout": 0.0, "torch_dtype": "bfloat16", "tie_word_embeddings": false, "architectures": ["MistralForCausalLM"], "bos_token_id": 1, "eos_token_id": 2, "model_type": "mistral", "neuron": {"auto_cast_type": "bf16", "batch_size": 32, "checkpoint_id": "mistralai/Mistral-7B-Instruct-v0.2", "checkpoint_revision": "41b61a33a2483885c981aa79e0df6b32407ed873", "compiler_type": "neuronx-cc", "compiler_version": "2.13.66.0+6dfecc895", "num_cores": 8, "sequence_length": 4096, "task": "text-generation"}}
|
neuronxcc-2.13.66.0+6dfecc895/0_REGISTRY/0.0.21.dev0/inference/mistral/mistralai/Mistral-7B-v0.1/12212376258a9fed88b2.json
ADDED
@@ -0,0 +1 @@
|
|
|
|
|
1 |
+
{"vocab_size": 32000, "max_position_embeddings": 32768, "hidden_size": 4096, "intermediate_size": 14336, "num_hidden_layers": 32, "num_attention_heads": 32, "sliding_window": 4096, "num_key_value_heads": 8, "hidden_act": "silu", "initializer_range": 0.02, "rms_norm_eps": 1e-05, "use_cache": true, "rope_theta": 10000.0, "attention_dropout": 0.0, "torch_dtype": "bfloat16", "tie_word_embeddings": false, "architectures": ["MistralForCausalLM"], "bos_token_id": 1, "eos_token_id": 2, "model_type": "mistral", "neuron": {"auto_cast_type": "bf16", "batch_size": 32, "checkpoint_id": "mistralai/Mistral-7B-v0.1", "checkpoint_revision": "26bca36bde8333b5d7f72e9ed20ccda6a618af24", "compiler_type": "neuronx-cc", "compiler_version": "2.13.66.0+6dfecc895", "num_cores": 8, "sequence_length": 4096, "task": "text-generation"}}
|
neuronxcc-2.13.66.0+6dfecc895/0_REGISTRY/0.0.21.dev0/inference/mistral/mistralai/Mistral-7B-v0.1/15f566c81d1f67f6fd2e.json
ADDED
@@ -0,0 +1 @@
|
|
|
|
|
1 |
+
{"vocab_size": 32000, "max_position_embeddings": 32768, "hidden_size": 4096, "intermediate_size": 14336, "num_hidden_layers": 32, "num_attention_heads": 32, "sliding_window": 4096, "num_key_value_heads": 8, "hidden_act": "silu", "initializer_range": 0.02, "rms_norm_eps": 1e-05, "use_cache": true, "rope_theta": 10000.0, "attention_dropout": 0.0, "torch_dtype": "bfloat16", "tie_word_embeddings": false, "architectures": ["MistralForCausalLM"], "bos_token_id": 1, "eos_token_id": 2, "model_type": "mistral", "neuron": {"auto_cast_type": "bf16", "batch_size": 8, "checkpoint_id": "mistralai/Mistral-7B-v0.1", "checkpoint_revision": "26bca36bde8333b5d7f72e9ed20ccda6a618af24", "compiler_type": "neuronx-cc", "compiler_version": "2.13.66.0+6dfecc895", "num_cores": 8, "sequence_length": 4096, "task": "text-generation"}}
|
neuronxcc-2.13.66.0+6dfecc895/0_REGISTRY/0.0.21.dev0/inference/mistral/mistralai/Mistral-7B-v0.1/192bd7f0468f78103585.json
ADDED
@@ -0,0 +1 @@
|
|
|
|
|
1 |
+
{"vocab_size": 32000, "max_position_embeddings": 32768, "hidden_size": 4096, "intermediate_size": 14336, "num_hidden_layers": 32, "num_attention_heads": 32, "sliding_window": 4096, "num_key_value_heads": 8, "hidden_act": "silu", "initializer_range": 0.02, "rms_norm_eps": 1e-05, "use_cache": true, "rope_theta": 10000.0, "attention_dropout": 0.0, "torch_dtype": "bfloat16", "tie_word_embeddings": false, "architectures": ["MistralForCausalLM"], "bos_token_id": 1, "eos_token_id": 2, "model_type": "mistral", "neuron": {"task": "text-generation", "batch_size": 8, "num_cores": 8, "auto_cast_type": "bf16", "sequence_length": 4096, "compiler_type": "neuronx-cc", "compiler_version": "2.13.66.0+6dfecc895", "checkpoint_id": "mistralai/Mistral-7B-v0.1", "checkpoint_revision": "26bca36bde8333b5d7f72e9ed20ccda6a618af24"}}
|
neuronxcc-2.13.66.0+6dfecc895/0_REGISTRY/0.0.21.dev0/inference/mistral/mistralai/Mistral-7B-v0.1/1dd1302a33c9e69730d2.json
ADDED
@@ -0,0 +1 @@
|
|
|
|
|
1 |
+
{"vocab_size": 32000, "max_position_embeddings": 32768, "hidden_size": 4096, "intermediate_size": 14336, "num_hidden_layers": 32, "num_attention_heads": 32, "sliding_window": 4096, "num_key_value_heads": 8, "hidden_act": "silu", "initializer_range": 0.02, "rms_norm_eps": 1e-05, "use_cache": true, "rope_theta": 10000.0, "attention_dropout": 0.0, "torch_dtype": "bfloat16", "tie_word_embeddings": false, "architectures": ["MistralForCausalLM"], "bos_token_id": 1, "eos_token_id": 2, "model_type": "mistral", "neuron": {"auto_cast_type": "bf16", "batch_size": 16, "checkpoint_id": "mistralai/Mistral-7B-v0.1", "checkpoint_revision": "26bca36bde8333b5d7f72e9ed20ccda6a618af24", "compiler_type": "neuronx-cc", "compiler_version": "2.13.66.0+6dfecc895", "num_cores": 8, "sequence_length": 4096, "task": "text-generation"}}
|
neuronxcc-2.13.66.0+6dfecc895/0_REGISTRY/0.0.21.dev0/inference/mistral/mistralai/Mistral-7B-v0.1/21bbf61928931f72378f.json
ADDED
@@ -0,0 +1 @@
|
|
|
|
|
1 |
+
{"vocab_size": 32000, "max_position_embeddings": 32768, "hidden_size": 4096, "intermediate_size": 14336, "num_hidden_layers": 32, "num_attention_heads": 32, "sliding_window": 4096, "num_key_value_heads": 8, "hidden_act": "silu", "initializer_range": 0.02, "rms_norm_eps": 1e-05, "use_cache": true, "rope_theta": 10000.0, "attention_dropout": 0.0, "torch_dtype": "bfloat16", "tie_word_embeddings": false, "architectures": ["MistralForCausalLM"], "bos_token_id": 1, "eos_token_id": 2, "model_type": "mistral", "neuron": {"task": "text-generation", "batch_size": 1, "num_cores": 8, "auto_cast_type": "bf16", "sequence_length": 4096, "compiler_type": "neuronx-cc", "compiler_version": "2.13.66.0+6dfecc895", "checkpoint_id": "mistralai/Mistral-7B-v0.1", "checkpoint_revision": "26bca36bde8333b5d7f72e9ed20ccda6a618af24"}}
|
neuronxcc-2.13.66.0+6dfecc895/0_REGISTRY/0.0.21.dev0/inference/mistral/mistralai/Mistral-7B-v0.1/5141bf1fc4c434ce1ea7.json
ADDED
@@ -0,0 +1 @@
|
|
|
|
|
1 |
+
{"vocab_size": 32000, "max_position_embeddings": 32768, "hidden_size": 4096, "intermediate_size": 14336, "num_hidden_layers": 32, "num_attention_heads": 32, "sliding_window": 4096, "num_key_value_heads": 8, "hidden_act": "silu", "initializer_range": 0.02, "rms_norm_eps": 1e-05, "use_cache": true, "rope_theta": 10000.0, "attention_dropout": 0.0, "torch_dtype": "bfloat16", "tie_word_embeddings": false, "architectures": ["MistralForCausalLM"], "bos_token_id": 1, "eos_token_id": 2, "model_type": "mistral", "neuron": {"auto_cast_type": "bf16", "batch_size": 4, "checkpoint_id": "mistralai/Mistral-7B-v0.1", "checkpoint_revision": "26bca36bde8333b5d7f72e9ed20ccda6a618af24", "compiler_type": "neuronx-cc", "compiler_version": "2.13.66.0+6dfecc895", "num_cores": 8, "sequence_length": 4096, "task": "text-generation"}}
|
neuronxcc-2.13.66.0+6dfecc895/0_REGISTRY/0.0.21.dev0/inference/mistral/mistralai/Mistral-7B-v0.1/67dd8bb21d625e22cd5b.json
ADDED
@@ -0,0 +1 @@
|
|
|
|
|
1 |
+
{"vocab_size": 32000, "max_position_embeddings": 32768, "hidden_size": 4096, "intermediate_size": 14336, "num_hidden_layers": 32, "num_attention_heads": 32, "sliding_window": 4096, "num_key_value_heads": 8, "hidden_act": "silu", "initializer_range": 0.02, "rms_norm_eps": 1e-05, "use_cache": true, "rope_theta": 10000.0, "attention_dropout": 0.0, "torch_dtype": "bfloat16", "tie_word_embeddings": false, "architectures": ["MistralForCausalLM"], "bos_token_id": 1, "eos_token_id": 2, "model_type": "mistral", "neuron": {"task": "text-generation", "batch_size": 32, "num_cores": 8, "auto_cast_type": "bf16", "sequence_length": 4096, "compiler_type": "neuronx-cc", "compiler_version": "2.13.66.0+6dfecc895", "checkpoint_id": "mistralai/Mistral-7B-v0.1", "checkpoint_revision": "26bca36bde8333b5d7f72e9ed20ccda6a618af24"}}
|
neuronxcc-2.13.66.0+6dfecc895/0_REGISTRY/0.0.21.dev0/inference/mistral/mistralai/Mistral-7B-v0.1/81be8abab2484836ea97.json
ADDED
@@ -0,0 +1 @@
|
|
|
|
|
1 |
+
{"vocab_size": 32000, "max_position_embeddings": 32768, "hidden_size": 4096, "intermediate_size": 14336, "num_hidden_layers": 32, "num_attention_heads": 32, "sliding_window": 4096, "num_key_value_heads": 8, "hidden_act": "silu", "initializer_range": 0.02, "rms_norm_eps": 1e-05, "use_cache": true, "rope_theta": 10000.0, "attention_dropout": 0.0, "torch_dtype": "bfloat16", "tie_word_embeddings": false, "architectures": ["MistralForCausalLM"], "bos_token_id": 1, "eos_token_id": 2, "model_type": "mistral", "neuron": {"task": "text-generation", "batch_size": 4, "num_cores": 8, "auto_cast_type": "bf16", "sequence_length": 4096, "compiler_type": "neuronx-cc", "compiler_version": "2.13.66.0+6dfecc895", "checkpoint_id": "mistralai/Mistral-7B-v0.1", "checkpoint_revision": "26bca36bde8333b5d7f72e9ed20ccda6a618af24"}}
|
neuronxcc-2.13.66.0+6dfecc895/0_REGISTRY/0.0.21.dev0/inference/mistral/mistralai/Mistral-7B-v0.1/9e73f279a5220e46dcf1.json
ADDED
@@ -0,0 +1 @@
|
|
|
|
|
1 |
+
{"vocab_size": 32000, "max_position_embeddings": 32768, "hidden_size": 4096, "intermediate_size": 14336, "num_hidden_layers": 32, "num_attention_heads": 32, "sliding_window": 4096, "num_key_value_heads": 8, "hidden_act": "silu", "initializer_range": 0.02, "rms_norm_eps": 1e-05, "use_cache": true, "rope_theta": 10000.0, "attention_dropout": 0.0, "torch_dtype": "bfloat16", "tie_word_embeddings": false, "architectures": ["MistralForCausalLM"], "bos_token_id": 1, "eos_token_id": 2, "model_type": "mistral", "neuron": {"auto_cast_type": "bf16", "batch_size": 1, "checkpoint_id": "mistralai/Mistral-7B-v0.1", "checkpoint_revision": "26bca36bde8333b5d7f72e9ed20ccda6a618af24", "compiler_type": "neuronx-cc", "compiler_version": "2.13.66.0+6dfecc895", "num_cores": 8, "sequence_length": 4096, "task": "text-generation"}}
|
neuronxcc-2.13.66.0+6dfecc895/0_REGISTRY/0.0.21.dev0/inference/mistral/mistralai/Mistral-7B-v0.1/9eb34c5c51dae5739571.json
ADDED
@@ -0,0 +1 @@
|
|
|
|
|
1 |
+
{"vocab_size": 32000, "max_position_embeddings": 32768, "hidden_size": 4096, "intermediate_size": 14336, "num_hidden_layers": 32, "num_attention_heads": 32, "sliding_window": 4096, "num_key_value_heads": 8, "hidden_act": "silu", "initializer_range": 0.02, "rms_norm_eps": 1e-05, "use_cache": true, "rope_theta": 10000.0, "attention_dropout": 0.0, "torch_dtype": "bfloat16", "tie_word_embeddings": false, "architectures": ["MistralForCausalLM"], "bos_token_id": 1, "eos_token_id": 2, "model_type": "mistral", "neuron": {"task": "text-generation", "batch_size": 16, "num_cores": 8, "auto_cast_type": "bf16", "sequence_length": 4096, "compiler_type": "neuronx-cc", "compiler_version": "2.13.66.0+6dfecc895", "checkpoint_id": "mistralai/Mistral-7B-v0.1", "checkpoint_revision": "26bca36bde8333b5d7f72e9ed20ccda6a618af24"}}
|
neuronxcc-2.13.66.0+6dfecc895/0_REGISTRY/0.0.21.dev0/inference/opt/hf-internal-testing/tiny-random-OPTForCausalLM/0eb6d808c895dd98f1d9.json
ADDED
@@ -0,0 +1 @@
|
|
|
|
|
1 |
+
{"torch_dtype": "float32", "is_decoder": true, "architectures": ["OPTForCausalLM"], "bos_token_id": 2, "pad_token_id": 1, "eos_token_id": 2, "embed_dim": 16, "model_type": "opt", "neuron": {"auto_cast_type": "fp32", "batch_size": 2, "checkpoint_id": "hf-internal-testing/tiny-random-OPTForCausalLM", "checkpoint_revision": "190d1f4fc0011d2eaeaa05282e0fbd2445e4b11f", "compiler_type": "neuronx-cc", "compiler_version": "2.13.66.0+6dfecc895", "num_cores": 2, "sequence_length": 100, "task": "text-generation"}, "vocab_size": 50265, "max_position_embeddings": 100, "num_attention_heads": 4, "word_embed_proj_dim": 16, "ffn_dim": 4, "hidden_size": 16, "num_hidden_layers": 5, "dropout": 0.1, "attention_dropout": 0.1, "activation_function": "relu", "init_std": 0.02, "layerdrop": 0.0, "use_cache": true, "do_layer_norm_before": true, "enable_bias": true, "layer_norm_elementwise_affine": true, "_remove_final_layer_norm": false}
|
neuronxcc-2.13.66.0+6dfecc895/0_REGISTRY/0.0.21.dev0/inference/opt/hf-internal-testing/tiny-random-OPTForCausalLM/244d78257ade535a7f03.json
ADDED
@@ -0,0 +1 @@
|
|
|
|
|
1 |
+
{"torch_dtype": "float32", "is_decoder": true, "architectures": ["OPTForCausalLM"], "bos_token_id": 2, "pad_token_id": 1, "eos_token_id": 2, "embed_dim": 16, "model_type": "opt", "vocab_size": 50265, "max_position_embeddings": 100, "num_attention_heads": 4, "word_embed_proj_dim": 16, "ffn_dim": 4, "hidden_size": 16, "num_hidden_layers": 5, "dropout": 0.1, "attention_dropout": 0.1, "activation_function": "relu", "init_std": 0.02, "layerdrop": 0.0, "use_cache": true, "do_layer_norm_before": true, "enable_bias": true, "layer_norm_elementwise_affine": true, "_remove_final_layer_norm": false, "neuron": {"task": "text-generation", "batch_size": 2, "num_cores": 2, "auto_cast_type": "fp32", "sequence_length": 100, "compiler_type": "neuronx-cc", "compiler_version": "2.13.66.0+6dfecc895", "checkpoint_id": "hf-internal-testing/tiny-random-OPTForCausalLM", "checkpoint_revision": "190d1f4fc0011d2eaeaa05282e0fbd2445e4b11f"}}
|
neuronxcc-2.13.66.0+6dfecc895/MODULE_03e6f14805ce902580b5+2c2d707e/compile_flags.txt
ADDED
@@ -0,0 +1 @@
|
|
|
|
|
1 |
+
--model-type=transformer --auto-cast=none
|
neuronxcc-2.13.66.0+6dfecc895/MODULE_03e6f14805ce902580b5+2c2d707e/model.hlo.pb
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:35163c9fe615e4417ea3264c7affccccb93c9873e79b89af900f55fdb350f6a8
|
3 |
+
size 8715129
|
neuronxcc-2.13.66.0+6dfecc895/MODULE_03e6f14805ce902580b5+2c2d707e/model.neff
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:1f8e696d3ad615b5494da26e6a3258b63a5a5afd095d962a2e150f2c33fd3009
|
3 |
+
size 3615744
|
neuronxcc-2.13.66.0+6dfecc895/MODULE_05115dff8003eb0253a4+2c2d707e/compile_flags.txt
ADDED
@@ -0,0 +1 @@
|
|
|
|
|
1 |
+
--model-type=transformer --auto-cast=none
|