dacorvo HF staff commited on
Commit
619547a
1 Parent(s): 04d89bb

Synchronizing local compiler cache.

Browse files
This view is limited to 50 files because it contains too many changes.   See raw diff
Files changed (50) hide show
  1. .gitattributes +34 -0
  2. neuronxcc-2.12.68.0+4480452af/0_REGISTRY/0.0.19.dev0/bloom/hf-internal-testing/tiny-random-BloomForCausalLM/099759272d993c40a970.json +1 -0
  3. neuronxcc-2.12.68.0+4480452af/0_REGISTRY/0.0.19.dev0/bloom/hf-internal-testing/tiny-random-BloomForCausalLM/239f41145d1b767d334a.json +1 -0
  4. neuronxcc-2.12.68.0+4480452af/0_REGISTRY/0.0.19.dev0/bloom/hf-internal-testing/tiny-random-BloomForCausalLM/4d8722cb197e46e38b27.json +1 -0
  5. neuronxcc-2.12.68.0+4480452af/0_REGISTRY/0.0.19.dev0/bloom/hf-internal-testing/tiny-random-BloomForCausalLM/748b2fdfaddfd8af9e97.json +1 -0
  6. neuronxcc-2.12.68.0+4480452af/0_REGISTRY/0.0.19.dev0/bloom/hf-internal-testing/tiny-random-BloomForCausalLM/796b0a69b6cfb147681c.json +1 -0
  7. neuronxcc-2.12.68.0+4480452af/0_REGISTRY/0.0.19.dev0/bloom/hf-internal-testing/tiny-random-BloomForCausalLM/cf86851df25066dfc4a8.json +1 -0
  8. neuronxcc-2.12.68.0+4480452af/0_REGISTRY/0.0.19.dev0/bloom/hf-internal-testing/tiny-random-BloomForCausalLM/fc67e4e8571c0e5d61c0.json +1 -0
  9. neuronxcc-2.12.68.0+4480452af/0_REGISTRY/0.0.19.dev0/bloom/hf-internal-testing/tiny-random-BloomForCausalLM/ff669ec4237cf4cf05af.json +1 -0
  10. neuronxcc-2.12.68.0+4480452af/0_REGISTRY/0.0.19.dev0/gpt2/gpt2/6c685e79c2c5aee24a81.json +1 -0
  11. neuronxcc-2.12.68.0+4480452af/0_REGISTRY/0.0.19.dev0/gpt2/hf-internal-testing/tiny-random-gpt2/0300ef28a839709a0896.json +1 -0
  12. neuronxcc-2.12.68.0+4480452af/0_REGISTRY/0.0.19.dev0/gpt2/hf-internal-testing/tiny-random-gpt2/0d1c153d6b4d1312bc77.json +1 -0
  13. neuronxcc-2.12.68.0+4480452af/0_REGISTRY/0.0.19.dev0/gpt2/hf-internal-testing/tiny-random-gpt2/13764fb2969c5a940316.json +1 -0
  14. neuronxcc-2.12.68.0+4480452af/0_REGISTRY/0.0.19.dev0/gpt2/hf-internal-testing/tiny-random-gpt2/24c47890e4ce14857524.json +1 -0
  15. neuronxcc-2.12.68.0+4480452af/0_REGISTRY/0.0.19.dev0/gpt2/hf-internal-testing/tiny-random-gpt2/6c1fb523f74e4521f986.json +1 -0
  16. neuronxcc-2.12.68.0+4480452af/0_REGISTRY/0.0.19.dev0/gpt2/hf-internal-testing/tiny-random-gpt2/80e0b6325bbb2858eb14.json +1 -0
  17. neuronxcc-2.12.68.0+4480452af/0_REGISTRY/0.0.19.dev0/gpt2/hf-internal-testing/tiny-random-gpt2/ca8d99bc6374cb0ea37c.json +1 -0
  18. neuronxcc-2.12.68.0+4480452af/0_REGISTRY/0.0.19.dev0/gpt2/hf-internal-testing/tiny-random-gpt2/d70210c8fe7cbaf377ab.json +1 -0
  19. neuronxcc-2.12.68.0+4480452af/0_REGISTRY/0.0.19.dev0/gpt2/hf-internal-testing/tiny-random-gpt2/eb42bab3c5050bb02069.json +1 -0
  20. neuronxcc-2.12.68.0+4480452af/0_REGISTRY/0.0.19.dev0/gpt2/hf-internal-testing/tiny-random-gpt2/ecdc60511e87d1dda9fb.json +1 -0
  21. neuronxcc-2.12.68.0+4480452af/0_REGISTRY/0.0.19.dev0/llama/NousResearch/Llama-2-7b-chat-hf/024f7a52b94f5d2bb43f.json +1 -0
  22. neuronxcc-2.12.68.0+4480452af/0_REGISTRY/0.0.19.dev0/llama/dacorvo/tiny-random-llama/0c5dcee71eb09cd78ca3.json +1 -0
  23. neuronxcc-2.12.68.0+4480452af/0_REGISTRY/0.0.19.dev0/llama/dacorvo/tiny-random-llama/2ea9c51c3f55022f558f.json +1 -0
  24. neuronxcc-2.12.68.0+4480452af/0_REGISTRY/0.0.19.dev0/llama/dacorvo/tiny-random-llama/36b96843d581cc86fdd1.json +1 -0
  25. neuronxcc-2.12.68.0+4480452af/0_REGISTRY/0.0.19.dev0/llama/dacorvo/tiny-random-llama/4f598bddf7d4afda210f.json +1 -0
  26. neuronxcc-2.12.68.0+4480452af/0_REGISTRY/0.0.19.dev0/llama/dacorvo/tiny-random-llama/823f635bc0b9b2e27147.json +1 -0
  27. neuronxcc-2.12.68.0+4480452af/0_REGISTRY/0.0.19.dev0/llama/dacorvo/tiny-random-llama/d9d1f24f085b3e1ee495.json +1 -0
  28. neuronxcc-2.12.68.0+4480452af/0_REGISTRY/0.0.19.dev0/llama/dacorvo/tiny-random-llama/e74544f6ce86f925e132.json +1 -0
  29. neuronxcc-2.12.68.0+4480452af/0_REGISTRY/0.0.19.dev0/llama/dacorvo/tiny-random-llama/f409bda0862113a1e014.json +1 -0
  30. neuronxcc-2.12.68.0+4480452af/0_REGISTRY/0.0.19.dev0/llama/meta-llama/Llama-2-7b-chat-hf/18f6f0c38b0c07980e4e.json +1 -0
  31. neuronxcc-2.12.68.0+4480452af/0_REGISTRY/0.0.19.dev0/llama/meta-llama/Llama-2-7b-chat-hf/58d1e81d4b6ef21222d8.json +1 -0
  32. neuronxcc-2.12.68.0+4480452af/0_REGISTRY/0.0.19.dev0/llama/meta-llama/Llama-2-7b-chat-hf/a23bae888fd7fe3142b6.json +1 -0
  33. neuronxcc-2.12.68.0+4480452af/0_REGISTRY/0.0.19.dev0/llama/meta-llama/Llama-2-7b-chat-hf/a8458afb96b3a458065e.json +1 -0
  34. neuronxcc-2.12.68.0+4480452af/0_REGISTRY/0.0.19.dev0/llama/meta-llama/Llama-2-7b-chat-hf/cba933997ecb39d1bd8c.json +1 -0
  35. neuronxcc-2.12.68.0+4480452af/0_REGISTRY/0.0.19.dev0/llama/meta-llama/Llama-2-7b-chat-hf/d1c9d1c755775ce29399.json +1 -0
  36. neuronxcc-2.12.68.0+4480452af/0_REGISTRY/0.0.19.dev0/llama/princeton-nlp/Sheared-LLaMA-1.3B/8d6bbf381f3c1b57ac17.json +1 -0
  37. neuronxcc-2.12.68.0+4480452af/0_REGISTRY/0.0.19.dev0/mistral/dacorvo/tiny-random-MistralForCausalLM/06fc7c62762607ab744e.json +1 -0
  38. neuronxcc-2.12.68.0+4480452af/0_REGISTRY/0.0.19.dev0/mistral/dacorvo/tiny-random-MistralForCausalLM/5975d2858f9bce8ef290.json +1 -0
  39. neuronxcc-2.12.68.0+4480452af/0_REGISTRY/0.0.19.dev0/mistral/dacorvo/tiny-random-MistralForCausalLM/87a7a99e2c2090700c5f.json +1 -0
  40. neuronxcc-2.12.68.0+4480452af/0_REGISTRY/0.0.19.dev0/mistral/dacorvo/tiny-random-MistralForCausalLM/9a8b7079ddb8a7611d5a.json +1 -0
  41. neuronxcc-2.12.68.0+4480452af/0_REGISTRY/0.0.19.dev0/mistral/dacorvo/tiny-random-MistralForCausalLM/fbf45e5cda8a458d37d3.json +1 -0
  42. neuronxcc-2.12.68.0+4480452af/0_REGISTRY/0.0.19.dev0/mistral/dacorvo/tiny-random-MistralForCausalLM/fdf8481549a30230656d.json +1 -0
  43. neuronxcc-2.12.68.0+4480452af/0_REGISTRY/0.0.19.dev0/opt/hf-internal-testing/tiny-random-OPTForCausalLM/0128c1baa44ad491d0f4.json +1 -0
  44. neuronxcc-2.12.68.0+4480452af/0_REGISTRY/0.0.19.dev0/opt/hf-internal-testing/tiny-random-OPTForCausalLM/187350afb13bf6aee97c.json +1 -0
  45. neuronxcc-2.12.68.0+4480452af/0_REGISTRY/0.0.19.dev0/opt/hf-internal-testing/tiny-random-OPTForCausalLM/45cb29bd56d7cd7ddc1a.json +1 -0
  46. neuronxcc-2.12.68.0+4480452af/0_REGISTRY/0.0.19.dev0/opt/hf-internal-testing/tiny-random-OPTForCausalLM/6f2444cfa075d1f210d0.json +1 -0
  47. neuronxcc-2.12.68.0+4480452af/0_REGISTRY/0.0.19.dev0/opt/hf-internal-testing/tiny-random-OPTForCausalLM/ea89a1b8198a05d1e71d.json +1 -0
  48. neuronxcc-2.12.68.0+4480452af/0_REGISTRY/0.0.19.dev0/opt/hf-internal-testing/tiny-random-OPTForCausalLM/f3ff2a63bf38db3d9595.json +1 -0
  49. neuronxcc-2.12.68.0+4480452af/MODULE_05c54d3634daf421dee1+2c2d707e/model.neff +0 -0
  50. neuronxcc-2.12.68.0+4480452af/MODULE_066004de210eb2d9c913+eefbc487/compile_flags.txt +1 -0
.gitattributes CHANGED
@@ -956,3 +956,37 @@ neuronxcc-2.12.68.0+4480452af/MODULE_acab3c397b3109c486af+2c2d707e/model.neff fi
956
  neuronxcc-2.12.68.0+4480452af/MODULE_b26a1377ff5739dbcc2d+2c2d707e/model.neff filter=lfs diff=lfs merge=lfs -text
957
  neuronxcc-2.12.68.0+4480452af/MODULE_dbf20078a89040997a97+2c2d707e/model.neff filter=lfs diff=lfs merge=lfs -text
958
  neuronxcc-2.12.68.0+4480452af/MODULE_eefa74de40f1b0605cd4+2c2d707e/model.neff filter=lfs diff=lfs merge=lfs -text
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
956
  neuronxcc-2.12.68.0+4480452af/MODULE_b26a1377ff5739dbcc2d+2c2d707e/model.neff filter=lfs diff=lfs merge=lfs -text
957
  neuronxcc-2.12.68.0+4480452af/MODULE_dbf20078a89040997a97+2c2d707e/model.neff filter=lfs diff=lfs merge=lfs -text
958
  neuronxcc-2.12.68.0+4480452af/MODULE_eefa74de40f1b0605cd4+2c2d707e/model.neff filter=lfs diff=lfs merge=lfs -text
959
+ neuronxcc-2.12.68.0+4480452af/MODULE_066004de210eb2d9c913+eefbc487/model.neff filter=lfs diff=lfs merge=lfs -text
960
+ neuronxcc-2.12.68.0+4480452af/MODULE_0942b3859682581dd1fe+2c2d707e/model.neff filter=lfs diff=lfs merge=lfs -text
961
+ neuronxcc-2.12.68.0+4480452af/MODULE_0f2c9d580b08a38164e1+eefbc487/model.neff filter=lfs diff=lfs merge=lfs -text
962
+ neuronxcc-2.12.68.0+4480452af/MODULE_15575de81949a4b92aa2+eefbc487/model.neff filter=lfs diff=lfs merge=lfs -text
963
+ neuronxcc-2.12.68.0+4480452af/MODULE_1d81285968f437f261e3+2c2d707e/model.neff filter=lfs diff=lfs merge=lfs -text
964
+ neuronxcc-2.12.68.0+4480452af/MODULE_223754db4990973e39fb+2c2d707e/model.neff filter=lfs diff=lfs merge=lfs -text
965
+ neuronxcc-2.12.68.0+4480452af/MODULE_226b245cfaf7762e6d9b+2c2d707e/model.neff filter=lfs diff=lfs merge=lfs -text
966
+ neuronxcc-2.12.68.0+4480452af/MODULE_2ca1e585d26d3652aa58+2c2d707e/model.neff filter=lfs diff=lfs merge=lfs -text
967
+ neuronxcc-2.12.68.0+4480452af/MODULE_513d7135f4d5bf996f2f+eefbc487/model.neff filter=lfs diff=lfs merge=lfs -text
968
+ neuronxcc-2.12.68.0+4480452af/MODULE_546eacf1e08561576cce+2c2d707e/model.neff filter=lfs diff=lfs merge=lfs -text
969
+ neuronxcc-2.12.68.0+4480452af/MODULE_6550f48aa0ef2604b21a+2c2d707e/model.neff filter=lfs diff=lfs merge=lfs -text
970
+ neuronxcc-2.12.68.0+4480452af/MODULE_667471b346635ae07c92+2c2d707e/model.neff filter=lfs diff=lfs merge=lfs -text
971
+ neuronxcc-2.12.68.0+4480452af/MODULE_686167d9ddedfddd84d6+2c2d707e/model.neff filter=lfs diff=lfs merge=lfs -text
972
+ neuronxcc-2.12.68.0+4480452af/MODULE_6af8f12e80b3c60ffd77+2c2d707e/model.neff filter=lfs diff=lfs merge=lfs -text
973
+ neuronxcc-2.12.68.0+4480452af/MODULE_76b368b3f15aff947a36+2c2d707e/model.neff filter=lfs diff=lfs merge=lfs -text
974
+ neuronxcc-2.12.68.0+4480452af/MODULE_77e951b968906564e8eb+2c2d707e/model.neff filter=lfs diff=lfs merge=lfs -text
975
+ neuronxcc-2.12.68.0+4480452af/MODULE_7ced8b335e90748dd532+2c2d707e/model.neff filter=lfs diff=lfs merge=lfs -text
976
+ neuronxcc-2.12.68.0+4480452af/MODULE_7d55a4b09b8600f2cfbf+2c2d707e/model.neff filter=lfs diff=lfs merge=lfs -text
977
+ neuronxcc-2.12.68.0+4480452af/MODULE_826a840ae5283ddff3fe+2c2d707e/model.neff filter=lfs diff=lfs merge=lfs -text
978
+ neuronxcc-2.12.68.0+4480452af/MODULE_845c788160f0d4d42a8e+eefbc487/model.neff filter=lfs diff=lfs merge=lfs -text
979
+ neuronxcc-2.12.68.0+4480452af/MODULE_9f559e16dfa4eb2c344f+eefbc487/model.neff filter=lfs diff=lfs merge=lfs -text
980
+ neuronxcc-2.12.68.0+4480452af/MODULE_a8123f8eb1ebcfd1b503+2c2d707e/model.neff filter=lfs diff=lfs merge=lfs -text
981
+ neuronxcc-2.12.68.0+4480452af/MODULE_b334dfeba0380c9d8e5c+2c2d707e/model.neff filter=lfs diff=lfs merge=lfs -text
982
+ neuronxcc-2.12.68.0+4480452af/MODULE_b45eacefccd8d1396063+2c2d707e/model.neff filter=lfs diff=lfs merge=lfs -text
983
+ neuronxcc-2.12.68.0+4480452af/MODULE_bc2b48e02226f9cc29a6+eefbc487/model.neff filter=lfs diff=lfs merge=lfs -text
984
+ neuronxcc-2.12.68.0+4480452af/MODULE_c093544c78cd27f1aef5+eefbc487/model.neff filter=lfs diff=lfs merge=lfs -text
985
+ neuronxcc-2.12.68.0+4480452af/MODULE_c2518cfecfe497358c10+2c2d707e/model.neff filter=lfs diff=lfs merge=lfs -text
986
+ neuronxcc-2.12.68.0+4480452af/MODULE_cd0fa8a24ba22743afe0+eefbc487/model.neff filter=lfs diff=lfs merge=lfs -text
987
+ neuronxcc-2.12.68.0+4480452af/MODULE_d49868da19e9b603a35d+2c2d707e/model.neff filter=lfs diff=lfs merge=lfs -text
988
+ neuronxcc-2.12.68.0+4480452af/MODULE_e7de26df3f2e7bddb371+2c2d707e/model.neff filter=lfs diff=lfs merge=lfs -text
989
+ neuronxcc-2.12.68.0+4480452af/MODULE_e8b4528eb5b3d2b6f000+2c2d707e/model.neff filter=lfs diff=lfs merge=lfs -text
990
+ neuronxcc-2.12.68.0+4480452af/MODULE_edd7f008ebd6e5b11f2a+2c2d707e/model.neff filter=lfs diff=lfs merge=lfs -text
991
+ neuronxcc-2.12.68.0+4480452af/MODULE_f734cee9a605225d8642+eefbc487/model.neff filter=lfs diff=lfs merge=lfs -text
992
+ neuronxcc-2.12.68.0+4480452af/MODULE_ffb4bd5d0af615ca83a1+2c2d707e/model.neff filter=lfs diff=lfs merge=lfs -text
neuronxcc-2.12.68.0+4480452af/0_REGISTRY/0.0.19.dev0/bloom/hf-internal-testing/tiny-random-BloomForCausalLM/099759272d993c40a970.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"vocab_size": 1024, "hidden_size": 32, "n_layer": 5, "n_head": 4, "layer_norm_epsilon": 1e-05, "initializer_range": 0.02, "use_cache": true, "pretraining_tp": 1, "apply_residual_connection_post_layernorm": false, "hidden_dropout": 0.1, "attention_dropout": 0.1, "bos_token_id": 1, "eos_token_id": 2, "slow_but_exact": true, "torch_dtype": "float32", "is_decoder": true, "architectures": ["BloomForCausalLM"], "id2label": {"0": "LABEL_0", "1": "LABEL_1", "2": "LABEL_2"}, "label2id": {"LABEL_0": 0, "LABEL_1": 1, "LABEL_2": 2}, "pad_token_id": 3, "dtype": "float32", "gradient_checkpointing": false, "model_type": "bloom", "n_positions": 512, "seq_length": 7, "type_vocab_size": 16, "neuron": {"task": "text-generation", "batch_size": 2, "num_cores": 2, "auto_cast_type": "fp16", "sequence_length": 100, "compiler_type": "neuronx-cc", "compiler_version": "2.12.68.0+4480452af", "checkpoint_id": "hf-internal-testing/tiny-random-BloomForCausalLM", "checkpoint_revision": "0f4f06f162cd67d34d03ee156484e4001d468500"}}
neuronxcc-2.12.68.0+4480452af/0_REGISTRY/0.0.19.dev0/bloom/hf-internal-testing/tiny-random-BloomForCausalLM/239f41145d1b767d334a.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"vocab_size": 1024, "hidden_size": 32, "n_layer": 5, "n_head": 4, "layer_norm_epsilon": 1e-05, "initializer_range": 0.02, "use_cache": true, "pretraining_tp": 1, "apply_residual_connection_post_layernorm": false, "hidden_dropout": 0.1, "attention_dropout": 0.1, "bos_token_id": 1, "eos_token_id": 2, "slow_but_exact": true, "torch_dtype": "float32", "is_decoder": true, "architectures": ["BloomForCausalLM"], "id2label": {"0": "LABEL_0", "1": "LABEL_1", "2": "LABEL_2"}, "label2id": {"LABEL_0": 0, "LABEL_1": 1, "LABEL_2": 2}, "pad_token_id": 3, "dtype": "float32", "gradient_checkpointing": false, "model_type": "bloom", "n_positions": 512, "neuron": {"auto_cast_type": "fp32", "batch_size": 1, "checkpoint_id": "hf-internal-testing/tiny-random-BloomForCausalLM", "checkpoint_revision": "0f4f06f162cd67d34d03ee156484e4001d468500", "compiler_type": "neuronx-cc", "compiler_version": "2.12.68.0+4480452af", "num_cores": 2, "sequence_length": 100, "task": "text-generation"}, "seq_length": 7, "type_vocab_size": 16}
neuronxcc-2.12.68.0+4480452af/0_REGISTRY/0.0.19.dev0/bloom/hf-internal-testing/tiny-random-BloomForCausalLM/4d8722cb197e46e38b27.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"vocab_size": 1024, "hidden_size": 32, "n_layer": 5, "n_head": 4, "layer_norm_epsilon": 1e-05, "initializer_range": 0.02, "use_cache": true, "pretraining_tp": 1, "apply_residual_connection_post_layernorm": false, "hidden_dropout": 0.1, "attention_dropout": 0.1, "bos_token_id": 1, "eos_token_id": 2, "slow_but_exact": true, "torch_dtype": "float32", "is_decoder": true, "architectures": ["BloomForCausalLM"], "id2label": {"0": "LABEL_0", "1": "LABEL_1", "2": "LABEL_2"}, "label2id": {"LABEL_0": 0, "LABEL_1": 1, "LABEL_2": 2}, "pad_token_id": 3, "dtype": "float32", "gradient_checkpointing": false, "model_type": "bloom", "n_positions": 512, "neuron": {"auto_cast_type": "fp32", "batch_size": 2, "checkpoint_id": "hf-internal-testing/tiny-random-BloomForCausalLM", "checkpoint_revision": "0f4f06f162cd67d34d03ee156484e4001d468500", "compiler_type": "neuronx-cc", "compiler_version": "2.12.68.0+4480452af", "num_cores": 2, "sequence_length": 256, "task": "text-generation"}, "seq_length": 7, "type_vocab_size": 16}
neuronxcc-2.12.68.0+4480452af/0_REGISTRY/0.0.19.dev0/bloom/hf-internal-testing/tiny-random-BloomForCausalLM/748b2fdfaddfd8af9e97.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"vocab_size": 1024, "hidden_size": 32, "n_layer": 5, "n_head": 4, "layer_norm_epsilon": 1e-05, "initializer_range": 0.02, "use_cache": true, "pretraining_tp": 1, "apply_residual_connection_post_layernorm": false, "hidden_dropout": 0.1, "attention_dropout": 0.1, "bos_token_id": 1, "eos_token_id": 2, "slow_but_exact": true, "torch_dtype": "float32", "is_decoder": true, "architectures": ["BloomForCausalLM"], "id2label": {"0": "LABEL_0", "1": "LABEL_1", "2": "LABEL_2"}, "label2id": {"LABEL_0": 0, "LABEL_1": 1, "LABEL_2": 2}, "pad_token_id": 3, "dtype": "float32", "gradient_checkpointing": false, "model_type": "bloom", "n_positions": 512, "seq_length": 7, "type_vocab_size": 16, "neuron": {"task": "text-generation", "batch_size": 2, "num_cores": 2, "auto_cast_type": "fp32", "sequence_length": 100, "compiler_type": "neuronx-cc", "compiler_version": "2.12.68.0+4480452af", "checkpoint_id": "hf-internal-testing/tiny-random-BloomForCausalLM", "checkpoint_revision": "0f4f06f162cd67d34d03ee156484e4001d468500"}}
neuronxcc-2.12.68.0+4480452af/0_REGISTRY/0.0.19.dev0/bloom/hf-internal-testing/tiny-random-BloomForCausalLM/796b0a69b6cfb147681c.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"vocab_size": 1024, "hidden_size": 32, "n_layer": 5, "n_head": 4, "layer_norm_epsilon": 1e-05, "initializer_range": 0.02, "use_cache": true, "pretraining_tp": 1, "apply_residual_connection_post_layernorm": false, "hidden_dropout": 0.1, "attention_dropout": 0.1, "bos_token_id": 1, "eos_token_id": 2, "slow_but_exact": true, "torch_dtype": "float32", "is_decoder": true, "architectures": ["BloomForCausalLM"], "id2label": {"0": "LABEL_0", "1": "LABEL_1", "2": "LABEL_2"}, "label2id": {"LABEL_0": 0, "LABEL_1": 1, "LABEL_2": 2}, "pad_token_id": 3, "dtype": "float32", "gradient_checkpointing": false, "model_type": "bloom", "n_positions": 512, "seq_length": 7, "type_vocab_size": 16, "neuron": {"task": "text-generation", "batch_size": 1, "num_cores": 2, "auto_cast_type": "fp16", "sequence_length": 100, "compiler_type": "neuronx-cc", "compiler_version": "2.12.68.0+4480452af", "checkpoint_id": "hf-internal-testing/tiny-random-BloomForCausalLM", "checkpoint_revision": "0f4f06f162cd67d34d03ee156484e4001d468500"}}
neuronxcc-2.12.68.0+4480452af/0_REGISTRY/0.0.19.dev0/bloom/hf-internal-testing/tiny-random-BloomForCausalLM/cf86851df25066dfc4a8.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"vocab_size": 1024, "hidden_size": 32, "n_layer": 5, "n_head": 4, "layer_norm_epsilon": 1e-05, "initializer_range": 0.02, "use_cache": true, "pretraining_tp": 1, "apply_residual_connection_post_layernorm": false, "hidden_dropout": 0.1, "attention_dropout": 0.1, "bos_token_id": 1, "eos_token_id": 2, "slow_but_exact": true, "torch_dtype": "float32", "is_decoder": true, "architectures": ["BloomForCausalLM"], "id2label": {"0": "LABEL_0", "1": "LABEL_1", "2": "LABEL_2"}, "label2id": {"LABEL_0": 0, "LABEL_1": 1, "LABEL_2": 2}, "pad_token_id": 3, "dtype": "float32", "gradient_checkpointing": false, "model_type": "bloom", "n_positions": 512, "seq_length": 7, "type_vocab_size": 16, "neuron": {"task": "text-generation", "batch_size": 1, "num_cores": 2, "auto_cast_type": "fp32", "sequence_length": 100, "compiler_type": "neuronx-cc", "compiler_version": "2.12.68.0+4480452af", "checkpoint_id": "hf-internal-testing/tiny-random-BloomForCausalLM", "checkpoint_revision": "0f4f06f162cd67d34d03ee156484e4001d468500"}}
neuronxcc-2.12.68.0+4480452af/0_REGISTRY/0.0.19.dev0/bloom/hf-internal-testing/tiny-random-BloomForCausalLM/fc67e4e8571c0e5d61c0.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"vocab_size": 1024, "hidden_size": 32, "n_layer": 5, "n_head": 4, "layer_norm_epsilon": 1e-05, "initializer_range": 0.02, "use_cache": true, "pretraining_tp": 1, "apply_residual_connection_post_layernorm": false, "hidden_dropout": 0.1, "attention_dropout": 0.1, "bos_token_id": 1, "eos_token_id": 2, "slow_but_exact": true, "torch_dtype": "float32", "is_decoder": true, "architectures": ["BloomForCausalLM"], "id2label": {"0": "LABEL_0", "1": "LABEL_1", "2": "LABEL_2"}, "label2id": {"LABEL_0": 0, "LABEL_1": 1, "LABEL_2": 2}, "pad_token_id": 3, "dtype": "float32", "gradient_checkpointing": false, "model_type": "bloom", "n_positions": 512, "neuron": {"auto_cast_type": "fp32", "batch_size": 2, "checkpoint_id": "hf-internal-testing/tiny-random-BloomForCausalLM", "checkpoint_revision": "0f4f06f162cd67d34d03ee156484e4001d468500", "compiler_type": "neuronx-cc", "compiler_version": "2.12.68.0+4480452af", "num_cores": 2, "sequence_length": 100, "task": "text-generation"}, "seq_length": 7, "type_vocab_size": 16}
neuronxcc-2.12.68.0+4480452af/0_REGISTRY/0.0.19.dev0/bloom/hf-internal-testing/tiny-random-BloomForCausalLM/ff669ec4237cf4cf05af.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"vocab_size": 1024, "hidden_size": 32, "n_layer": 5, "n_head": 4, "layer_norm_epsilon": 1e-05, "initializer_range": 0.02, "use_cache": true, "pretraining_tp": 1, "apply_residual_connection_post_layernorm": false, "hidden_dropout": 0.1, "attention_dropout": 0.1, "bos_token_id": 1, "eos_token_id": 2, "slow_but_exact": true, "torch_dtype": "float32", "is_decoder": true, "architectures": ["BloomForCausalLM"], "id2label": {"0": "LABEL_0", "1": "LABEL_1", "2": "LABEL_2"}, "label2id": {"LABEL_0": 0, "LABEL_1": 1, "LABEL_2": 2}, "pad_token_id": 3, "dtype": "float32", "gradient_checkpointing": false, "model_type": "bloom", "n_positions": 512, "seq_length": 7, "type_vocab_size": 16, "neuron": {"task": "text-generation", "batch_size": 2, "num_cores": 2, "auto_cast_type": "fp32", "sequence_length": 256, "compiler_type": "neuronx-cc", "compiler_version": "2.12.68.0+4480452af", "checkpoint_id": "hf-internal-testing/tiny-random-BloomForCausalLM", "checkpoint_revision": "0f4f06f162cd67d34d03ee156484e4001d468500"}}
neuronxcc-2.12.68.0+4480452af/0_REGISTRY/0.0.19.dev0/gpt2/gpt2/6c685e79c2c5aee24a81.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"vocab_size": 50257, "n_positions": 1024, "n_embd": 768, "n_layer": 12, "n_head": 12, "n_inner": null, "activation_function": "gelu_new", "resid_pdrop": 0.1, "embd_pdrop": 0.1, "attn_pdrop": 0.1, "layer_norm_epsilon": 1e-05, "initializer_range": 0.02, "summary_type": "cls_index", "summary_use_proj": true, "summary_activation": null, "summary_first_dropout": 0.1, "summary_proj_to_labels": true, "scale_attn_weights": true, "use_cache": true, "scale_attn_by_inverse_layer_idx": false, "reorder_and_upcast_attn": false, "bos_token_id": 50256, "eos_token_id": 50256, "architectures": ["GPT2LMHeadModel"], "task_specific_params": {"text-generation": {"do_sample": true, "max_length": 50}}, "model_type": "gpt2", "n_ctx": 1024, "neuron": {"auto_cast_type": "fp32", "batch_size": 4, "checkpoint_id": "gpt2", "checkpoint_revision": "11c5a3d5811f50298f278a704980280950aedb10", "compiler_type": "neuronx-cc", "compiler_version": "2.12.68.0+4480452af", "num_cores": 2, "sequence_length": 1024, "task": "text-generation"}}
neuronxcc-2.12.68.0+4480452af/0_REGISTRY/0.0.19.dev0/gpt2/hf-internal-testing/tiny-random-gpt2/0300ef28a839709a0896.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"vocab_size": 1000, "n_positions": 512, "n_embd": 32, "n_layer": 5, "n_head": 4, "n_inner": null, "activation_function": "gelu_new", "resid_pdrop": 0.1, "embd_pdrop": 0.1, "attn_pdrop": 0.1, "layer_norm_epsilon": 1e-05, "initializer_range": 0.02, "summary_type": "cls_index", "summary_use_proj": true, "summary_activation": null, "summary_first_dropout": 0.1, "summary_proj_to_labels": true, "scale_attn_weights": true, "use_cache": true, "scale_attn_by_inverse_layer_idx": false, "reorder_and_upcast_attn": false, "bos_token_id": 98, "eos_token_id": 98, "pad_token_id": 98, "attention_probs_dropout_prob": 0.1, "gradient_checkpointing": false, "hidden_act": "gelu", "hidden_dropout_prob": 0.1, "intermediate_size": 37, "model_type": "gpt2", "n_ctx": 512, "type_vocab_size": 16, "neuron": {"task": "text-generation", "batch_size": 1, "num_cores": 2, "auto_cast_type": "fp16", "sequence_length": 100, "compiler_type": "neuronx-cc", "compiler_version": "2.12.68.0+4480452af", "checkpoint_id": "hf-internal-testing/tiny-random-gpt2", "checkpoint_revision": "91c0fe31d692dd8448d9bc06e8d1877345009e3b"}}
neuronxcc-2.12.68.0+4480452af/0_REGISTRY/0.0.19.dev0/gpt2/hf-internal-testing/tiny-random-gpt2/0d1c153d6b4d1312bc77.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"vocab_size": 1000, "n_positions": 512, "n_embd": 32, "n_layer": 5, "n_head": 4, "n_inner": null, "activation_function": "gelu_new", "resid_pdrop": 0.1, "embd_pdrop": 0.1, "attn_pdrop": 0.1, "layer_norm_epsilon": 1e-05, "initializer_range": 0.02, "summary_type": "cls_index", "summary_use_proj": true, "summary_activation": null, "summary_first_dropout": 0.1, "summary_proj_to_labels": true, "scale_attn_weights": true, "use_cache": true, "scale_attn_by_inverse_layer_idx": false, "reorder_and_upcast_attn": false, "bos_token_id": 98, "eos_token_id": 98, "pad_token_id": 98, "attention_probs_dropout_prob": 0.1, "gradient_checkpointing": false, "hidden_act": "gelu", "hidden_dropout_prob": 0.1, "intermediate_size": 37, "model_type": "gpt2", "n_ctx": 512, "neuron": {"auto_cast_type": "fp32", "batch_size": 2, "checkpoint_id": "hf-internal-testing/tiny-random-gpt2", "checkpoint_revision": "91c0fe31d692dd8448d9bc06e8d1877345009e3b", "compiler_type": "neuronx-cc", "compiler_version": "2.12.68.0+4480452af", "num_cores": 2, "sequence_length": 100, "task": "text-generation"}, "type_vocab_size": 16}
neuronxcc-2.12.68.0+4480452af/0_REGISTRY/0.0.19.dev0/gpt2/hf-internal-testing/tiny-random-gpt2/13764fb2969c5a940316.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"vocab_size": 1000, "n_positions": 512, "n_embd": 32, "n_layer": 5, "n_head": 4, "n_inner": null, "activation_function": "gelu_new", "resid_pdrop": 0.1, "embd_pdrop": 0.1, "attn_pdrop": 0.1, "layer_norm_epsilon": 1e-05, "initializer_range": 0.02, "summary_type": "cls_index", "summary_use_proj": true, "summary_activation": null, "summary_first_dropout": 0.1, "summary_proj_to_labels": true, "scale_attn_weights": true, "use_cache": true, "scale_attn_by_inverse_layer_idx": false, "reorder_and_upcast_attn": false, "bos_token_id": 98, "eos_token_id": 98, "pad_token_id": 98, "attention_probs_dropout_prob": 0.1, "gradient_checkpointing": false, "hidden_act": "gelu", "hidden_dropout_prob": 0.1, "intermediate_size": 37, "model_type": "gpt2", "n_ctx": 512, "neuron": {"auto_cast_type": "fp32", "batch_size": 16, "checkpoint_id": "hf-internal-testing/tiny-random-gpt2", "checkpoint_revision": "91c0fe31d692dd8448d9bc06e8d1877345009e3b", "compiler_type": "neuronx-cc", "compiler_version": "2.12.54.0+f631c2365", "num_cores": 2, "sequence_length": 512, "task": "text-generation"}, "type_vocab_size": 16}
neuronxcc-2.12.68.0+4480452af/0_REGISTRY/0.0.19.dev0/gpt2/hf-internal-testing/tiny-random-gpt2/24c47890e4ce14857524.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"vocab_size": 1000, "n_positions": 512, "n_embd": 32, "n_layer": 5, "n_head": 4, "n_inner": null, "activation_function": "gelu_new", "resid_pdrop": 0.1, "embd_pdrop": 0.1, "attn_pdrop": 0.1, "layer_norm_epsilon": 1e-05, "initializer_range": 0.02, "summary_type": "cls_index", "summary_use_proj": true, "summary_activation": null, "summary_first_dropout": 0.1, "summary_proj_to_labels": true, "scale_attn_weights": true, "use_cache": true, "scale_attn_by_inverse_layer_idx": false, "reorder_and_upcast_attn": false, "bos_token_id": 98, "eos_token_id": 98, "pad_token_id": 98, "attention_probs_dropout_prob": 0.1, "gradient_checkpointing": false, "hidden_act": "gelu", "hidden_dropout_prob": 0.1, "intermediate_size": 37, "model_type": "gpt2", "n_ctx": 512, "type_vocab_size": 16, "neuron": {"task": "text-generation", "batch_size": 1, "num_cores": 24, "auto_cast_type": "fp32", "sequence_length": 512, "compiler_type": "neuronx-cc", "compiler_version": "2.12.68.0+4480452af", "checkpoint_id": "hf-internal-testing/tiny-random-gpt2", "checkpoint_revision": "91c0fe31d692dd8448d9bc06e8d1877345009e3b"}}
neuronxcc-2.12.68.0+4480452af/0_REGISTRY/0.0.19.dev0/gpt2/hf-internal-testing/tiny-random-gpt2/6c1fb523f74e4521f986.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"vocab_size": 1000, "n_positions": 512, "n_embd": 32, "n_layer": 5, "n_head": 4, "n_inner": null, "activation_function": "gelu_new", "resid_pdrop": 0.1, "embd_pdrop": 0.1, "attn_pdrop": 0.1, "layer_norm_epsilon": 1e-05, "initializer_range": 0.02, "summary_type": "cls_index", "summary_use_proj": true, "summary_activation": null, "summary_first_dropout": 0.1, "summary_proj_to_labels": true, "scale_attn_weights": true, "use_cache": true, "scale_attn_by_inverse_layer_idx": false, "reorder_and_upcast_attn": false, "bos_token_id": 98, "eos_token_id": 98, "pad_token_id": 98, "attention_probs_dropout_prob": 0.1, "gradient_checkpointing": false, "hidden_act": "gelu", "hidden_dropout_prob": 0.1, "intermediate_size": 37, "model_type": "gpt2", "n_ctx": 512, "neuron": {"auto_cast_type": "fp32", "batch_size": 2, "checkpoint_id": "hf-internal-testing/tiny-random-gpt2", "checkpoint_revision": "91c0fe31d692dd8448d9bc06e8d1877345009e3b", "compiler_type": "neuronx-cc", "compiler_version": "2.12.68.0+4480452af", "num_cores": 2, "sequence_length": 256, "task": "text-generation"}, "type_vocab_size": 16}
neuronxcc-2.12.68.0+4480452af/0_REGISTRY/0.0.19.dev0/gpt2/hf-internal-testing/tiny-random-gpt2/80e0b6325bbb2858eb14.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"vocab_size": 1000, "n_positions": 512, "n_embd": 32, "n_layer": 5, "n_head": 4, "n_inner": null, "activation_function": "gelu_new", "resid_pdrop": 0.1, "embd_pdrop": 0.1, "attn_pdrop": 0.1, "layer_norm_epsilon": 1e-05, "initializer_range": 0.02, "summary_type": "cls_index", "summary_use_proj": true, "summary_activation": null, "summary_first_dropout": 0.1, "summary_proj_to_labels": true, "scale_attn_weights": true, "use_cache": true, "scale_attn_by_inverse_layer_idx": false, "reorder_and_upcast_attn": false, "bos_token_id": 98, "eos_token_id": 98, "pad_token_id": 98, "attention_probs_dropout_prob": 0.1, "gradient_checkpointing": false, "hidden_act": "gelu", "hidden_dropout_prob": 0.1, "intermediate_size": 37, "model_type": "gpt2", "n_ctx": 512, "type_vocab_size": 16, "neuron": {"task": "text-generation", "batch_size": 2, "num_cores": 2, "auto_cast_type": "fp32", "sequence_length": 256, "compiler_type": "neuronx-cc", "compiler_version": "2.12.68.0+4480452af", "checkpoint_id": "hf-internal-testing/tiny-random-gpt2", "checkpoint_revision": "91c0fe31d692dd8448d9bc06e8d1877345009e3b"}}
neuronxcc-2.12.68.0+4480452af/0_REGISTRY/0.0.19.dev0/gpt2/hf-internal-testing/tiny-random-gpt2/ca8d99bc6374cb0ea37c.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"vocab_size": 1000, "n_positions": 512, "n_embd": 32, "n_layer": 5, "n_head": 4, "n_inner": null, "activation_function": "gelu_new", "resid_pdrop": 0.1, "embd_pdrop": 0.1, "attn_pdrop": 0.1, "layer_norm_epsilon": 1e-05, "initializer_range": 0.02, "summary_type": "cls_index", "summary_use_proj": true, "summary_activation": null, "summary_first_dropout": 0.1, "summary_proj_to_labels": true, "scale_attn_weights": true, "use_cache": true, "scale_attn_by_inverse_layer_idx": false, "reorder_and_upcast_attn": false, "bos_token_id": 98, "eos_token_id": 98, "pad_token_id": 98, "attention_probs_dropout_prob": 0.1, "gradient_checkpointing": false, "hidden_act": "gelu", "hidden_dropout_prob": 0.1, "intermediate_size": 37, "model_type": "gpt2", "n_ctx": 512, "neuron": {"auto_cast_type": "fp32", "batch_size": 1, "checkpoint_id": "hf-internal-testing/tiny-random-gpt2", "checkpoint_revision": "91c0fe31d692dd8448d9bc06e8d1877345009e3b", "compiler_type": "neuronx-cc", "compiler_version": "2.12.68.0+4480452af", "num_cores": 2, "sequence_length": 100, "task": "text-generation"}, "type_vocab_size": 16}
neuronxcc-2.12.68.0+4480452af/0_REGISTRY/0.0.19.dev0/gpt2/hf-internal-testing/tiny-random-gpt2/d70210c8fe7cbaf377ab.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"vocab_size": 1000, "n_positions": 512, "n_embd": 32, "n_layer": 5, "n_head": 4, "n_inner": null, "activation_function": "gelu_new", "resid_pdrop": 0.1, "embd_pdrop": 0.1, "attn_pdrop": 0.1, "layer_norm_epsilon": 1e-05, "initializer_range": 0.02, "summary_type": "cls_index", "summary_use_proj": true, "summary_activation": null, "summary_first_dropout": 0.1, "summary_proj_to_labels": true, "scale_attn_weights": true, "use_cache": true, "scale_attn_by_inverse_layer_idx": false, "reorder_and_upcast_attn": false, "bos_token_id": 98, "eos_token_id": 98, "pad_token_id": 98, "attention_probs_dropout_prob": 0.1, "gradient_checkpointing": false, "hidden_act": "gelu", "hidden_dropout_prob": 0.1, "intermediate_size": 37, "model_type": "gpt2", "n_ctx": 512, "type_vocab_size": 16, "neuron": {"task": "text-generation", "batch_size": 2, "num_cores": 2, "auto_cast_type": "fp16", "sequence_length": 100, "compiler_type": "neuronx-cc", "compiler_version": "2.12.68.0+4480452af", "checkpoint_id": "hf-internal-testing/tiny-random-gpt2", "checkpoint_revision": "91c0fe31d692dd8448d9bc06e8d1877345009e3b"}}
neuronxcc-2.12.68.0+4480452af/0_REGISTRY/0.0.19.dev0/gpt2/hf-internal-testing/tiny-random-gpt2/eb42bab3c5050bb02069.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"vocab_size": 1000, "n_positions": 512, "n_embd": 32, "n_layer": 5, "n_head": 4, "n_inner": null, "activation_function": "gelu_new", "resid_pdrop": 0.1, "embd_pdrop": 0.1, "attn_pdrop": 0.1, "layer_norm_epsilon": 1e-05, "initializer_range": 0.02, "summary_type": "cls_index", "summary_use_proj": true, "summary_activation": null, "summary_first_dropout": 0.1, "summary_proj_to_labels": true, "scale_attn_weights": true, "use_cache": true, "scale_attn_by_inverse_layer_idx": false, "reorder_and_upcast_attn": false, "bos_token_id": 98, "eos_token_id": 98, "pad_token_id": 98, "attention_probs_dropout_prob": 0.1, "gradient_checkpointing": false, "hidden_act": "gelu", "hidden_dropout_prob": 0.1, "intermediate_size": 37, "model_type": "gpt2", "n_ctx": 512, "type_vocab_size": 16, "neuron": {"task": "text-generation", "batch_size": 1, "num_cores": 2, "auto_cast_type": "fp32", "sequence_length": 100, "compiler_type": "neuronx-cc", "compiler_version": "2.12.68.0+4480452af", "checkpoint_id": "hf-internal-testing/tiny-random-gpt2", "checkpoint_revision": "91c0fe31d692dd8448d9bc06e8d1877345009e3b"}}
neuronxcc-2.12.68.0+4480452af/0_REGISTRY/0.0.19.dev0/gpt2/hf-internal-testing/tiny-random-gpt2/ecdc60511e87d1dda9fb.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"vocab_size": 1000, "n_positions": 512, "n_embd": 32, "n_layer": 5, "n_head": 4, "n_inner": null, "activation_function": "gelu_new", "resid_pdrop": 0.1, "embd_pdrop": 0.1, "attn_pdrop": 0.1, "layer_norm_epsilon": 1e-05, "initializer_range": 0.02, "summary_type": "cls_index", "summary_use_proj": true, "summary_activation": null, "summary_first_dropout": 0.1, "summary_proj_to_labels": true, "scale_attn_weights": true, "use_cache": true, "scale_attn_by_inverse_layer_idx": false, "reorder_and_upcast_attn": false, "bos_token_id": 98, "eos_token_id": 98, "pad_token_id": 98, "attention_probs_dropout_prob": 0.1, "gradient_checkpointing": false, "hidden_act": "gelu", "hidden_dropout_prob": 0.1, "intermediate_size": 37, "model_type": "gpt2", "n_ctx": 512, "type_vocab_size": 16, "neuron": {"task": "text-generation", "batch_size": 2, "num_cores": 2, "auto_cast_type": "fp32", "sequence_length": 100, "compiler_type": "neuronx-cc", "compiler_version": "2.12.68.0+4480452af", "checkpoint_id": "hf-internal-testing/tiny-random-gpt2", "checkpoint_revision": "91c0fe31d692dd8448d9bc06e8d1877345009e3b"}}
neuronxcc-2.12.68.0+4480452af/0_REGISTRY/0.0.19.dev0/llama/NousResearch/Llama-2-7b-chat-hf/024f7a52b94f5d2bb43f.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"vocab_size": 32000, "max_position_embeddings": 4096, "hidden_size": 4096, "intermediate_size": 11008, "num_hidden_layers": 32, "num_attention_heads": 32, "num_key_value_heads": 32, "hidden_act": "silu", "initializer_range": 0.02, "rms_norm_eps": 1e-05, "pretraining_tp": 1, "use_cache": true, "rope_theta": 10000.0, "rope_scaling": null, "attention_bias": false, "attention_dropout": 0.0, "torch_dtype": "float16", "tie_word_embeddings": false, "architectures": ["LlamaForCausalLM"], "bos_token_id": 1, "pad_token_id": 0, "eos_token_id": 2, "model_type": "llama", "neuron": {"task": "text-generation", "batch_size": 2, "num_cores": 2, "auto_cast_type": "f16", "sequence_length": 2048, "compiler_type": "neuronx-cc", "compiler_version": "2.12.68.0+4480452af", "checkpoint_id": "NousResearch/Llama-2-7b-chat-hf", "checkpoint_revision": "37892f30c23786c0d5367d80481fa0d9fba93cf8"}}
neuronxcc-2.12.68.0+4480452af/0_REGISTRY/0.0.19.dev0/llama/dacorvo/tiny-random-llama/0c5dcee71eb09cd78ca3.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"vocab_size": 32000, "max_position_embeddings": 512, "hidden_size": 128, "intermediate_size": 256, "num_hidden_layers": 1, "num_attention_heads": 1, "num_key_value_heads": 1, "hidden_act": "silu", "initializer_range": 0.02, "rms_norm_eps": 1e-06, "pretraining_tp": 1, "use_cache": true, "rope_theta": 10000.0, "rope_scaling": null, "attention_bias": false, "attention_dropout": 0.0, "torch_dtype": "float32", "tie_word_embeddings": false, "architectures": ["LlamaForCausalLM"], "bos_token_id": 1, "eos_token_id": 2, "model_type": "llama", "neuron": {"auto_cast_type": "fp32", "batch_size": 1, "checkpoint_id": "dacorvo/tiny-random-llama", "checkpoint_revision": "7fdafd2fe6a2d31c6abb72ae60db606d8bb23196", "compiler_type": "neuronx-cc", "compiler_version": "2.12.68.0+4480452af", "num_cores": 2, "sequence_length": 100, "task": "text-generation"}}
neuronxcc-2.12.68.0+4480452af/0_REGISTRY/0.0.19.dev0/llama/dacorvo/tiny-random-llama/2ea9c51c3f55022f558f.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"vocab_size": 32000, "max_position_embeddings": 512, "hidden_size": 128, "intermediate_size": 256, "num_hidden_layers": 1, "num_attention_heads": 1, "num_key_value_heads": 1, "hidden_act": "silu", "initializer_range": 0.02, "rms_norm_eps": 1e-06, "pretraining_tp": 1, "use_cache": true, "rope_theta": 10000.0, "rope_scaling": null, "attention_bias": false, "attention_dropout": 0.0, "torch_dtype": "float32", "tie_word_embeddings": false, "architectures": ["LlamaForCausalLM"], "bos_token_id": 1, "eos_token_id": 2, "model_type": "llama", "neuron": {"auto_cast_type": "fp32", "batch_size": 2, "checkpoint_id": "dacorvo/tiny-random-llama", "checkpoint_revision": "7fdafd2fe6a2d31c6abb72ae60db606d8bb23196", "compiler_type": "neuronx-cc", "compiler_version": "2.12.68.0+4480452af", "num_cores": 2, "sequence_length": 100, "task": "text-generation"}}
neuronxcc-2.12.68.0+4480452af/0_REGISTRY/0.0.19.dev0/llama/dacorvo/tiny-random-llama/36b96843d581cc86fdd1.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"vocab_size": 32000, "max_position_embeddings": 512, "hidden_size": 128, "intermediate_size": 256, "num_hidden_layers": 1, "num_attention_heads": 1, "num_key_value_heads": 1, "hidden_act": "silu", "initializer_range": 0.02, "rms_norm_eps": 1e-06, "pretraining_tp": 1, "use_cache": true, "rope_theta": 10000.0, "rope_scaling": null, "attention_bias": false, "attention_dropout": 0.0, "torch_dtype": "float32", "tie_word_embeddings": false, "architectures": ["LlamaForCausalLM"], "bos_token_id": 1, "eos_token_id": 2, "model_type": "llama", "neuron": {"task": "text-generation", "batch_size": 1, "num_cores": 2, "auto_cast_type": "fp32", "sequence_length": 100, "compiler_type": "neuronx-cc", "compiler_version": "2.12.68.0+4480452af", "checkpoint_id": "dacorvo/tiny-random-llama", "checkpoint_revision": "7fdafd2fe6a2d31c6abb72ae60db606d8bb23196"}}
neuronxcc-2.12.68.0+4480452af/0_REGISTRY/0.0.19.dev0/llama/dacorvo/tiny-random-llama/4f598bddf7d4afda210f.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"vocab_size": 32000, "max_position_embeddings": 512, "hidden_size": 128, "intermediate_size": 256, "num_hidden_layers": 1, "num_attention_heads": 1, "num_key_value_heads": 1, "hidden_act": "silu", "initializer_range": 0.02, "rms_norm_eps": 1e-06, "pretraining_tp": 1, "use_cache": true, "rope_theta": 10000.0, "rope_scaling": null, "attention_bias": false, "attention_dropout": 0.0, "torch_dtype": "float32", "tie_word_embeddings": false, "architectures": ["LlamaForCausalLM"], "bos_token_id": 1, "eos_token_id": 2, "model_type": "llama", "neuron": {"task": "text-generation", "batch_size": 2, "num_cores": 2, "auto_cast_type": "fp16", "sequence_length": 100, "compiler_type": "neuronx-cc", "compiler_version": "2.12.68.0+4480452af", "checkpoint_id": "dacorvo/tiny-random-llama", "checkpoint_revision": "7fdafd2fe6a2d31c6abb72ae60db606d8bb23196"}}
neuronxcc-2.12.68.0+4480452af/0_REGISTRY/0.0.19.dev0/llama/dacorvo/tiny-random-llama/823f635bc0b9b2e27147.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"vocab_size": 32000, "max_position_embeddings": 512, "hidden_size": 128, "intermediate_size": 256, "num_hidden_layers": 1, "num_attention_heads": 1, "num_key_value_heads": 1, "hidden_act": "silu", "initializer_range": 0.02, "rms_norm_eps": 1e-06, "pretraining_tp": 1, "use_cache": true, "rope_theta": 10000.0, "rope_scaling": null, "attention_bias": false, "attention_dropout": 0.0, "torch_dtype": "float32", "tie_word_embeddings": false, "architectures": ["LlamaForCausalLM"], "bos_token_id": 1, "eos_token_id": 2, "model_type": "llama", "neuron": {"task": "text-generation", "batch_size": 2, "num_cores": 2, "auto_cast_type": "fp32", "sequence_length": 100, "compiler_type": "neuronx-cc", "compiler_version": "2.12.68.0+4480452af", "checkpoint_id": "dacorvo/tiny-random-llama", "checkpoint_revision": "7fdafd2fe6a2d31c6abb72ae60db606d8bb23196"}}
neuronxcc-2.12.68.0+4480452af/0_REGISTRY/0.0.19.dev0/llama/dacorvo/tiny-random-llama/d9d1f24f085b3e1ee495.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"vocab_size": 32000, "max_position_embeddings": 512, "hidden_size": 128, "intermediate_size": 256, "num_hidden_layers": 1, "num_attention_heads": 1, "num_key_value_heads": 1, "hidden_act": "silu", "initializer_range": 0.02, "rms_norm_eps": 1e-06, "pretraining_tp": 1, "use_cache": true, "rope_theta": 10000.0, "rope_scaling": null, "attention_bias": false, "attention_dropout": 0.0, "torch_dtype": "float32", "tie_word_embeddings": false, "architectures": ["LlamaForCausalLM"], "bos_token_id": 1, "eos_token_id": 2, "model_type": "llama", "neuron": {"task": "text-generation", "batch_size": 1, "num_cores": 2, "auto_cast_type": "fp16", "sequence_length": 100, "compiler_type": "neuronx-cc", "compiler_version": "2.12.68.0+4480452af", "checkpoint_id": "dacorvo/tiny-random-llama", "checkpoint_revision": "7fdafd2fe6a2d31c6abb72ae60db606d8bb23196"}}
neuronxcc-2.12.68.0+4480452af/0_REGISTRY/0.0.19.dev0/llama/dacorvo/tiny-random-llama/e74544f6ce86f925e132.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"vocab_size": 32000, "max_position_embeddings": 512, "hidden_size": 128, "intermediate_size": 256, "num_hidden_layers": 1, "num_attention_heads": 1, "num_key_value_heads": 1, "hidden_act": "silu", "initializer_range": 0.02, "rms_norm_eps": 1e-06, "pretraining_tp": 1, "use_cache": true, "rope_theta": 10000.0, "rope_scaling": null, "attention_bias": false, "attention_dropout": 0.0, "torch_dtype": "float32", "tie_word_embeddings": false, "architectures": ["LlamaForCausalLM"], "bos_token_id": 1, "eos_token_id": 2, "model_type": "llama", "neuron": {"auto_cast_type": "fp32", "batch_size": 2, "checkpoint_id": "dacorvo/tiny-random-llama", "checkpoint_revision": "7fdafd2fe6a2d31c6abb72ae60db606d8bb23196", "compiler_type": "neuronx-cc", "compiler_version": "2.12.68.0+4480452af", "num_cores": 2, "sequence_length": 256, "task": "text-generation"}}
neuronxcc-2.12.68.0+4480452af/0_REGISTRY/0.0.19.dev0/llama/dacorvo/tiny-random-llama/f409bda0862113a1e014.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"vocab_size": 32000, "max_position_embeddings": 512, "hidden_size": 128, "intermediate_size": 256, "num_hidden_layers": 1, "num_attention_heads": 1, "num_key_value_heads": 1, "hidden_act": "silu", "initializer_range": 0.02, "rms_norm_eps": 1e-06, "pretraining_tp": 1, "use_cache": true, "rope_theta": 10000.0, "rope_scaling": null, "attention_bias": false, "attention_dropout": 0.0, "torch_dtype": "float32", "tie_word_embeddings": false, "architectures": ["LlamaForCausalLM"], "bos_token_id": 1, "eos_token_id": 2, "model_type": "llama", "neuron": {"task": "text-generation", "batch_size": 2, "num_cores": 2, "auto_cast_type": "fp32", "sequence_length": 256, "compiler_type": "neuronx-cc", "compiler_version": "2.12.68.0+4480452af", "checkpoint_id": "dacorvo/tiny-random-llama", "checkpoint_revision": "7fdafd2fe6a2d31c6abb72ae60db606d8bb23196"}}
neuronxcc-2.12.68.0+4480452af/0_REGISTRY/0.0.19.dev0/llama/meta-llama/Llama-2-7b-chat-hf/18f6f0c38b0c07980e4e.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"vocab_size": 32000, "max_position_embeddings": 4096, "hidden_size": 4096, "intermediate_size": 11008, "num_hidden_layers": 32, "num_attention_heads": 32, "num_key_value_heads": 32, "hidden_act": "silu", "initializer_range": 0.02, "rms_norm_eps": 1e-05, "pretraining_tp": 1, "use_cache": true, "rope_theta": 10000.0, "rope_scaling": null, "attention_bias": false, "attention_dropout": 0.0, "torch_dtype": "float16", "tie_word_embeddings": false, "architectures": ["LlamaForCausalLM"], "bos_token_id": 1, "eos_token_id": 2, "model_type": "llama", "neuron": {"task": "text-generation", "batch_size": 2, "num_cores": 2, "auto_cast_type": "fp16", "sequence_length": 2048, "compiler_type": "neuronx-cc", "compiler_version": "2.12.68.0+4480452af", "checkpoint_id": "meta-llama/Llama-2-7b-chat-hf", "checkpoint_revision": "c1b0db933684edbfe29a06fa47eb19cc48025e93"}}
neuronxcc-2.12.68.0+4480452af/0_REGISTRY/0.0.19.dev0/llama/meta-llama/Llama-2-7b-chat-hf/58d1e81d4b6ef21222d8.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"vocab_size": 32000, "max_position_embeddings": 4096, "hidden_size": 4096, "intermediate_size": 11008, "num_hidden_layers": 32, "num_attention_heads": 32, "num_key_value_heads": 32, "hidden_act": "silu", "initializer_range": 0.02, "rms_norm_eps": 1e-05, "pretraining_tp": 1, "use_cache": true, "rope_theta": 10000.0, "rope_scaling": null, "attention_bias": false, "attention_dropout": 0.0, "torch_dtype": "float16", "tie_word_embeddings": false, "architectures": ["LlamaForCausalLM"], "bos_token_id": 1, "eos_token_id": 2, "model_type": "llama", "neuron": {"auto_cast_type": "fp16", "batch_size": 2, "checkpoint_id": "meta-llama/Llama-2-7b-chat-hf", "checkpoint_revision": "c1b0db933684edbfe29a06fa47eb19cc48025e93", "compiler_type": "neuronx-cc", "compiler_version": "2.12.68.0+4480452af", "num_cores": 2, "sequence_length": 2048, "task": "text-generation"}}
neuronxcc-2.12.68.0+4480452af/0_REGISTRY/0.0.19.dev0/llama/meta-llama/Llama-2-7b-chat-hf/a23bae888fd7fe3142b6.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"vocab_size": 32000, "max_position_embeddings": 4096, "hidden_size": 4096, "intermediate_size": 11008, "num_hidden_layers": 32, "num_attention_heads": 32, "num_key_value_heads": 32, "hidden_act": "silu", "initializer_range": 0.02, "rms_norm_eps": 1e-05, "pretraining_tp": 1, "use_cache": true, "rope_theta": 10000.0, "rope_scaling": null, "attention_bias": false, "attention_dropout": 0.0, "torch_dtype": "float16", "tie_word_embeddings": false, "architectures": ["LlamaForCausalLM"], "bos_token_id": 1, "eos_token_id": 2, "model_type": "llama", "neuron": {"task": "text-generation", "batch_size": 1, "num_cores": 2, "auto_cast_type": "fp16", "sequence_length": 2048, "compiler_type": "neuronx-cc", "compiler_version": "2.12.68.0+4480452af", "checkpoint_id": "meta-llama/Llama-2-7b-chat-hf", "checkpoint_revision": "c1b0db933684edbfe29a06fa47eb19cc48025e93"}}
neuronxcc-2.12.68.0+4480452af/0_REGISTRY/0.0.19.dev0/llama/meta-llama/Llama-2-7b-chat-hf/a8458afb96b3a458065e.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"vocab_size": 32000, "max_position_embeddings": 4096, "hidden_size": 4096, "intermediate_size": 11008, "num_hidden_layers": 32, "num_attention_heads": 32, "num_key_value_heads": 32, "hidden_act": "silu", "initializer_range": 0.02, "rms_norm_eps": 1e-05, "pretraining_tp": 1, "use_cache": true, "rope_theta": 10000.0, "rope_scaling": null, "attention_bias": false, "attention_dropout": 0.0, "torch_dtype": "float16", "tie_word_embeddings": false, "architectures": ["LlamaForCausalLM"], "bos_token_id": 1, "eos_token_id": 2, "model_type": "llama", "neuron": {"auto_cast_type": "fp16", "batch_size": 1, "checkpoint_id": "meta-llama/Llama-2-7b-chat-hf", "checkpoint_revision": "c1b0db933684edbfe29a06fa47eb19cc48025e93", "compiler_type": "neuronx-cc", "compiler_version": "2.12.68.0+4480452af", "num_cores": 2, "sequence_length": 2048, "task": "text-generation"}}
neuronxcc-2.12.68.0+4480452af/0_REGISTRY/0.0.19.dev0/llama/meta-llama/Llama-2-7b-chat-hf/cba933997ecb39d1bd8c.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"vocab_size": 32000, "max_position_embeddings": 4096, "hidden_size": 4096, "intermediate_size": 11008, "num_hidden_layers": 32, "num_attention_heads": 32, "num_key_value_heads": 32, "hidden_act": "silu", "initializer_range": 0.02, "rms_norm_eps": 1e-05, "pretraining_tp": 1, "use_cache": true, "rope_theta": 10000.0, "rope_scaling": null, "attention_bias": false, "attention_dropout": 0.0, "torch_dtype": "float16", "tie_word_embeddings": false, "architectures": ["LlamaForCausalLM"], "bos_token_id": 1, "eos_token_id": 2, "model_type": "llama", "neuron": {"task": "text-generation", "batch_size": 4, "num_cores": 8, "auto_cast_type": "fp16", "sequence_length": 2048, "compiler_type": "neuronx-cc", "compiler_version": "2.12.68.0+4480452af", "checkpoint_id": "meta-llama/Llama-2-7b-chat-hf", "checkpoint_revision": "c1b0db933684edbfe29a06fa47eb19cc48025e93"}}
neuronxcc-2.12.68.0+4480452af/0_REGISTRY/0.0.19.dev0/llama/meta-llama/Llama-2-7b-chat-hf/d1c9d1c755775ce29399.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"vocab_size": 32000, "max_position_embeddings": 4096, "hidden_size": 4096, "intermediate_size": 11008, "num_hidden_layers": 32, "num_attention_heads": 32, "num_key_value_heads": 32, "hidden_act": "silu", "initializer_range": 0.02, "rms_norm_eps": 1e-05, "pretraining_tp": 1, "use_cache": true, "rope_theta": 10000.0, "rope_scaling": null, "attention_bias": false, "attention_dropout": 0.0, "torch_dtype": "float16", "tie_word_embeddings": false, "architectures": ["LlamaForCausalLM"], "bos_token_id": 1, "eos_token_id": 2, "model_type": "llama", "neuron": {"auto_cast_type": "fp16", "batch_size": 4, "checkpoint_id": "meta-llama/Llama-2-7b-chat-hf", "checkpoint_revision": "c1b0db933684edbfe29a06fa47eb19cc48025e93", "compiler_type": "neuronx-cc", "compiler_version": "2.12.68.0+4480452af", "num_cores": 8, "sequence_length": 2048, "task": "text-generation"}}
neuronxcc-2.12.68.0+4480452af/0_REGISTRY/0.0.19.dev0/llama/princeton-nlp/Sheared-LLaMA-1.3B/8d6bbf381f3c1b57ac17.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"vocab_size": 32000, "max_position_embeddings": 4096, "hidden_size": 2048, "intermediate_size": 5504, "num_hidden_layers": 24, "num_attention_heads": 16, "num_key_value_heads": 16, "hidden_act": "silu", "initializer_range": 0.02, "rms_norm_eps": 1e-05, "pretraining_tp": 1, "use_cache": true, "rope_theta": 10000.0, "rope_scaling": null, "attention_bias": false, "attention_dropout": 0.0, "torch_dtype": "float32", "tie_word_embeddings": false, "architectures": ["LlamaForCausalLM"], "bos_token_id": 1, "pad_token_id": 0, "eos_token_id": 2, "model_type": "llama", "neuron": {"task": "text-generation", "batch_size": 4, "num_cores": 2, "auto_cast_type": "fp32", "sequence_length": 2048, "compiler_type": "neuronx-cc", "compiler_version": "2.12.68.0+4480452af", "checkpoint_id": "princeton-nlp/Sheared-LLaMA-1.3B", "checkpoint_revision": "a4b76938edbf571ea7d7d9904861cbdca08809b4"}}
neuronxcc-2.12.68.0+4480452af/0_REGISTRY/0.0.19.dev0/mistral/dacorvo/tiny-random-MistralForCausalLM/06fc7c62762607ab744e.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"vocab_size": 32000, "max_position_embeddings": 512, "hidden_size": 32, "intermediate_size": 37, "num_hidden_layers": 2, "num_attention_heads": 4, "sliding_window": 4096, "num_key_value_heads": 2, "hidden_act": "gelu", "initializer_range": 0.02, "rms_norm_eps": 1e-06, "use_cache": true, "rope_theta": 10000.0, "attention_dropout": 0.0, "torch_dtype": "float32", "tie_word_embeddings": false, "is_decoder": true, "architectures": ["MistralForCausalLM"], "bos_token_id": 1, "pad_token_id": 0, "eos_token_id": 2, "attention_probs_dropout_prob": 0.1, "hidden_dropout_prob": 0.1, "model_type": "mistral", "type_vocab_size": 16, "neuron": {"task": "text-generation", "batch_size": 2, "num_cores": 2, "auto_cast_type": "fp16", "sequence_length": 100, "compiler_type": "neuronx-cc", "compiler_version": "2.12.68.0+4480452af", "checkpoint_id": "dacorvo/tiny-random-MistralForCausalLM", "checkpoint_revision": "81d453e3c8985649e9ee3d4c9378461029d1c73a"}}
neuronxcc-2.12.68.0+4480452af/0_REGISTRY/0.0.19.dev0/mistral/dacorvo/tiny-random-MistralForCausalLM/5975d2858f9bce8ef290.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"vocab_size": 32000, "max_position_embeddings": 512, "hidden_size": 32, "intermediate_size": 37, "num_hidden_layers": 2, "num_attention_heads": 4, "sliding_window": 4096, "num_key_value_heads": 2, "hidden_act": "gelu", "initializer_range": 0.02, "rms_norm_eps": 1e-06, "use_cache": true, "rope_theta": 10000.0, "attention_dropout": 0.0, "torch_dtype": "float32", "tie_word_embeddings": false, "is_decoder": true, "architectures": ["MistralForCausalLM"], "bos_token_id": 1, "pad_token_id": 0, "eos_token_id": 2, "attention_probs_dropout_prob": 0.1, "hidden_dropout_prob": 0.1, "model_type": "mistral", "type_vocab_size": 16, "neuron": {"task": "text-generation", "batch_size": 1, "num_cores": 2, "auto_cast_type": "fp16", "sequence_length": 100, "compiler_type": "neuronx-cc", "compiler_version": "2.12.68.0+4480452af", "checkpoint_id": "dacorvo/tiny-random-MistralForCausalLM", "checkpoint_revision": "81d453e3c8985649e9ee3d4c9378461029d1c73a"}}
neuronxcc-2.12.68.0+4480452af/0_REGISTRY/0.0.19.dev0/mistral/dacorvo/tiny-random-MistralForCausalLM/87a7a99e2c2090700c5f.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"vocab_size": 32000, "max_position_embeddings": 512, "hidden_size": 32, "intermediate_size": 37, "num_hidden_layers": 2, "num_attention_heads": 4, "sliding_window": 4096, "num_key_value_heads": 2, "hidden_act": "gelu", "initializer_range": 0.02, "rms_norm_eps": 1e-06, "use_cache": true, "rope_theta": 10000.0, "attention_dropout": 0.0, "torch_dtype": "float32", "tie_word_embeddings": false, "is_decoder": true, "architectures": ["MistralForCausalLM"], "bos_token_id": 1, "pad_token_id": 0, "eos_token_id": 2, "attention_probs_dropout_prob": 0.1, "hidden_dropout_prob": 0.1, "model_type": "mistral", "type_vocab_size": 16, "neuron": {"task": "text-generation", "batch_size": 2, "num_cores": 2, "auto_cast_type": "fp32", "sequence_length": 100, "compiler_type": "neuronx-cc", "compiler_version": "2.12.68.0+4480452af", "checkpoint_id": "dacorvo/tiny-random-MistralForCausalLM", "checkpoint_revision": "81d453e3c8985649e9ee3d4c9378461029d1c73a"}}
neuronxcc-2.12.68.0+4480452af/0_REGISTRY/0.0.19.dev0/mistral/dacorvo/tiny-random-MistralForCausalLM/9a8b7079ddb8a7611d5a.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"vocab_size": 32000, "max_position_embeddings": 512, "hidden_size": 32, "intermediate_size": 37, "num_hidden_layers": 2, "num_attention_heads": 4, "sliding_window": 4096, "num_key_value_heads": 2, "hidden_act": "gelu", "initializer_range": 0.02, "rms_norm_eps": 1e-06, "use_cache": true, "rope_theta": 10000.0, "attention_dropout": 0.0, "torch_dtype": "float32", "tie_word_embeddings": false, "is_decoder": true, "architectures": ["MistralForCausalLM"], "bos_token_id": 1, "pad_token_id": 0, "eos_token_id": 2, "attention_probs_dropout_prob": 0.1, "hidden_dropout_prob": 0.1, "model_type": "mistral", "type_vocab_size": 16, "neuron": {"task": "text-generation", "batch_size": 1, "num_cores": 2, "auto_cast_type": "fp32", "sequence_length": 100, "compiler_type": "neuronx-cc", "compiler_version": "2.12.68.0+4480452af", "checkpoint_id": "dacorvo/tiny-random-MistralForCausalLM", "checkpoint_revision": "81d453e3c8985649e9ee3d4c9378461029d1c73a"}}
neuronxcc-2.12.68.0+4480452af/0_REGISTRY/0.0.19.dev0/mistral/dacorvo/tiny-random-MistralForCausalLM/fbf45e5cda8a458d37d3.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"vocab_size": 32000, "max_position_embeddings": 512, "hidden_size": 32, "intermediate_size": 37, "num_hidden_layers": 2, "num_attention_heads": 4, "sliding_window": 4096, "num_key_value_heads": 2, "hidden_act": "gelu", "initializer_range": 0.02, "rms_norm_eps": 1e-06, "use_cache": true, "rope_theta": 10000.0, "attention_dropout": 0.0, "torch_dtype": "float32", "tie_word_embeddings": false, "is_decoder": true, "architectures": ["MistralForCausalLM"], "bos_token_id": 1, "pad_token_id": 0, "eos_token_id": 2, "attention_probs_dropout_prob": 0.1, "hidden_dropout_prob": 0.1, "model_type": "mistral", "neuron": {"auto_cast_type": "fp32", "batch_size": 1, "checkpoint_id": "dacorvo/tiny-random-MistralForCausalLM", "checkpoint_revision": "81d453e3c8985649e9ee3d4c9378461029d1c73a", "compiler_type": "neuronx-cc", "compiler_version": "2.12.68.0+4480452af", "num_cores": 2, "sequence_length": 100, "task": "text-generation"}, "type_vocab_size": 16}
neuronxcc-2.12.68.0+4480452af/0_REGISTRY/0.0.19.dev0/mistral/dacorvo/tiny-random-MistralForCausalLM/fdf8481549a30230656d.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"vocab_size": 32000, "max_position_embeddings": 512, "hidden_size": 32, "intermediate_size": 37, "num_hidden_layers": 2, "num_attention_heads": 4, "sliding_window": 4096, "num_key_value_heads": 2, "hidden_act": "gelu", "initializer_range": 0.02, "rms_norm_eps": 1e-06, "use_cache": true, "rope_theta": 10000.0, "attention_dropout": 0.0, "torch_dtype": "float32", "tie_word_embeddings": false, "is_decoder": true, "architectures": ["MistralForCausalLM"], "bos_token_id": 1, "pad_token_id": 0, "eos_token_id": 2, "attention_probs_dropout_prob": 0.1, "hidden_dropout_prob": 0.1, "model_type": "mistral", "neuron": {"auto_cast_type": "fp32", "batch_size": 2, "checkpoint_id": "dacorvo/tiny-random-MistralForCausalLM", "checkpoint_revision": "81d453e3c8985649e9ee3d4c9378461029d1c73a", "compiler_type": "neuronx-cc", "compiler_version": "2.12.68.0+4480452af", "num_cores": 2, "sequence_length": 100, "task": "text-generation"}, "type_vocab_size": 16}
neuronxcc-2.12.68.0+4480452af/0_REGISTRY/0.0.19.dev0/opt/hf-internal-testing/tiny-random-OPTForCausalLM/0128c1baa44ad491d0f4.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"torch_dtype": "float32", "is_decoder": true, "architectures": ["OPTForCausalLM"], "bos_token_id": 2, "pad_token_id": 1, "eos_token_id": 2, "embed_dim": 16, "model_type": "opt", "neuron": {"auto_cast_type": "fp32", "batch_size": 1, "checkpoint_id": "hf-internal-testing/tiny-random-OPTForCausalLM", "checkpoint_revision": "190d1f4fc0011d2eaeaa05282e0fbd2445e4b11f", "compiler_type": "neuronx-cc", "compiler_version": "2.12.68.0+4480452af", "num_cores": 2, "sequence_length": 100, "task": "text-generation"}, "vocab_size": 50265, "max_position_embeddings": 100, "num_attention_heads": 4, "word_embed_proj_dim": 16, "ffn_dim": 4, "hidden_size": 16, "num_hidden_layers": 5, "dropout": 0.1, "attention_dropout": 0.1, "activation_function": "relu", "init_std": 0.02, "layerdrop": 0.0, "use_cache": true, "do_layer_norm_before": true, "enable_bias": true, "layer_norm_elementwise_affine": true, "_remove_final_layer_norm": false}
neuronxcc-2.12.68.0+4480452af/0_REGISTRY/0.0.19.dev0/opt/hf-internal-testing/tiny-random-OPTForCausalLM/187350afb13bf6aee97c.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"torch_dtype": "float32", "is_decoder": true, "architectures": ["OPTForCausalLM"], "bos_token_id": 2, "pad_token_id": 1, "eos_token_id": 2, "embed_dim": 16, "model_type": "opt", "vocab_size": 50265, "max_position_embeddings": 100, "num_attention_heads": 4, "word_embed_proj_dim": 16, "ffn_dim": 4, "hidden_size": 16, "num_hidden_layers": 5, "dropout": 0.1, "attention_dropout": 0.1, "activation_function": "relu", "init_std": 0.02, "layerdrop": 0.0, "use_cache": true, "do_layer_norm_before": true, "enable_bias": true, "layer_norm_elementwise_affine": true, "_remove_final_layer_norm": false, "neuron": {"task": "text-generation", "batch_size": 2, "num_cores": 2, "auto_cast_type": "fp32", "sequence_length": 100, "compiler_type": "neuronx-cc", "compiler_version": "2.12.68.0+4480452af", "checkpoint_id": "hf-internal-testing/tiny-random-OPTForCausalLM", "checkpoint_revision": "190d1f4fc0011d2eaeaa05282e0fbd2445e4b11f"}}
neuronxcc-2.12.68.0+4480452af/0_REGISTRY/0.0.19.dev0/opt/hf-internal-testing/tiny-random-OPTForCausalLM/45cb29bd56d7cd7ddc1a.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"torch_dtype": "float32", "is_decoder": true, "architectures": ["OPTForCausalLM"], "bos_token_id": 2, "pad_token_id": 1, "eos_token_id": 2, "embed_dim": 16, "model_type": "opt", "neuron": {"auto_cast_type": "fp32", "batch_size": 2, "checkpoint_id": "hf-internal-testing/tiny-random-OPTForCausalLM", "checkpoint_revision": "190d1f4fc0011d2eaeaa05282e0fbd2445e4b11f", "compiler_type": "neuronx-cc", "compiler_version": "2.12.68.0+4480452af", "num_cores": 2, "sequence_length": 100, "task": "text-generation"}, "vocab_size": 50265, "max_position_embeddings": 100, "num_attention_heads": 4, "word_embed_proj_dim": 16, "ffn_dim": 4, "hidden_size": 16, "num_hidden_layers": 5, "dropout": 0.1, "attention_dropout": 0.1, "activation_function": "relu", "init_std": 0.02, "layerdrop": 0.0, "use_cache": true, "do_layer_norm_before": true, "enable_bias": true, "layer_norm_elementwise_affine": true, "_remove_final_layer_norm": false}
neuronxcc-2.12.68.0+4480452af/0_REGISTRY/0.0.19.dev0/opt/hf-internal-testing/tiny-random-OPTForCausalLM/6f2444cfa075d1f210d0.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"torch_dtype": "float32", "is_decoder": true, "architectures": ["OPTForCausalLM"], "bos_token_id": 2, "pad_token_id": 1, "eos_token_id": 2, "embed_dim": 16, "model_type": "opt", "vocab_size": 50265, "max_position_embeddings": 100, "num_attention_heads": 4, "word_embed_proj_dim": 16, "ffn_dim": 4, "hidden_size": 16, "num_hidden_layers": 5, "dropout": 0.1, "attention_dropout": 0.1, "activation_function": "relu", "init_std": 0.02, "layerdrop": 0.0, "use_cache": true, "do_layer_norm_before": true, "enable_bias": true, "layer_norm_elementwise_affine": true, "_remove_final_layer_norm": false, "neuron": {"task": "text-generation", "batch_size": 1, "num_cores": 2, "auto_cast_type": "fp16", "sequence_length": 100, "compiler_type": "neuronx-cc", "compiler_version": "2.12.68.0+4480452af", "checkpoint_id": "hf-internal-testing/tiny-random-OPTForCausalLM", "checkpoint_revision": "190d1f4fc0011d2eaeaa05282e0fbd2445e4b11f"}}
neuronxcc-2.12.68.0+4480452af/0_REGISTRY/0.0.19.dev0/opt/hf-internal-testing/tiny-random-OPTForCausalLM/ea89a1b8198a05d1e71d.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"torch_dtype": "float32", "is_decoder": true, "architectures": ["OPTForCausalLM"], "bos_token_id": 2, "pad_token_id": 1, "eos_token_id": 2, "embed_dim": 16, "model_type": "opt", "vocab_size": 50265, "max_position_embeddings": 100, "num_attention_heads": 4, "word_embed_proj_dim": 16, "ffn_dim": 4, "hidden_size": 16, "num_hidden_layers": 5, "dropout": 0.1, "attention_dropout": 0.1, "activation_function": "relu", "init_std": 0.02, "layerdrop": 0.0, "use_cache": true, "do_layer_norm_before": true, "enable_bias": true, "layer_norm_elementwise_affine": true, "_remove_final_layer_norm": false, "neuron": {"task": "text-generation", "batch_size": 2, "num_cores": 2, "auto_cast_type": "fp16", "sequence_length": 100, "compiler_type": "neuronx-cc", "compiler_version": "2.12.68.0+4480452af", "checkpoint_id": "hf-internal-testing/tiny-random-OPTForCausalLM", "checkpoint_revision": "190d1f4fc0011d2eaeaa05282e0fbd2445e4b11f"}}
neuronxcc-2.12.68.0+4480452af/0_REGISTRY/0.0.19.dev0/opt/hf-internal-testing/tiny-random-OPTForCausalLM/f3ff2a63bf38db3d9595.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"torch_dtype": "float32", "is_decoder": true, "architectures": ["OPTForCausalLM"], "bos_token_id": 2, "pad_token_id": 1, "eos_token_id": 2, "embed_dim": 16, "model_type": "opt", "vocab_size": 50265, "max_position_embeddings": 100, "num_attention_heads": 4, "word_embed_proj_dim": 16, "ffn_dim": 4, "hidden_size": 16, "num_hidden_layers": 5, "dropout": 0.1, "attention_dropout": 0.1, "activation_function": "relu", "init_std": 0.02, "layerdrop": 0.0, "use_cache": true, "do_layer_norm_before": true, "enable_bias": true, "layer_norm_elementwise_affine": true, "_remove_final_layer_norm": false, "neuron": {"task": "text-generation", "batch_size": 1, "num_cores": 2, "auto_cast_type": "fp32", "sequence_length": 100, "compiler_type": "neuronx-cc", "compiler_version": "2.12.68.0+4480452af", "checkpoint_id": "hf-internal-testing/tiny-random-OPTForCausalLM", "checkpoint_revision": "190d1f4fc0011d2eaeaa05282e0fbd2445e4b11f"}}
neuronxcc-2.12.68.0+4480452af/MODULE_05c54d3634daf421dee1+2c2d707e/model.neff CHANGED
Binary files a/neuronxcc-2.12.68.0+4480452af/MODULE_05c54d3634daf421dee1+2c2d707e/model.neff and b/neuronxcc-2.12.68.0+4480452af/MODULE_05c54d3634daf421dee1+2c2d707e/model.neff differ
 
neuronxcc-2.12.68.0+4480452af/MODULE_066004de210eb2d9c913+eefbc487/compile_flags.txt ADDED
@@ -0,0 +1 @@
 
 
1
+ --model-type=transformer