dacorvo HF staff commited on
Commit
a825eb7
1 Parent(s): c86d4b0

Synchronizing local compiler cache.

Browse files
This view is limited to 50 files because it contains too many changes.   See raw diff
Files changed (50) hide show
  1. .gitattributes +6 -0
  2. neuronxcc-2.13.66.0+6dfecc895/0_REGISTRY/0.0.23.dev0/inference/bloom/hf-internal-testing/tiny-random-BloomForCausalLM/4efe4ec09e4490308cc0.json +1 -0
  3. neuronxcc-2.13.66.0+6dfecc895/0_REGISTRY/0.0.23.dev0/inference/bloom/hf-internal-testing/tiny-random-BloomForCausalLM/9d5f94b71f6ac3455612.json +1 -0
  4. neuronxcc-2.13.66.0+6dfecc895/0_REGISTRY/0.0.23.dev0/inference/bloom/hf-internal-testing/tiny-random-BloomForCausalLM/b9db5e6da88e9f07ee3e.json +1 -0
  5. neuronxcc-2.13.66.0+6dfecc895/0_REGISTRY/0.0.23.dev0/inference/bloom/hf-internal-testing/tiny-random-BloomForCausalLM/f9a5439ee67f962284df.json +1 -0
  6. neuronxcc-2.13.66.0+6dfecc895/0_REGISTRY/0.0.23.dev0/inference/gpt2/hf-internal-testing/tiny-random-gpt2/2713954a7f357d88b849.json +1 -0
  7. neuronxcc-2.13.66.0+6dfecc895/0_REGISTRY/0.0.23.dev0/inference/gpt2/hf-internal-testing/tiny-random-gpt2/7e574905d10c185da6d3.json +1 -0
  8. neuronxcc-2.13.66.0+6dfecc895/0_REGISTRY/0.0.23.dev0/inference/gpt2/hf-internal-testing/tiny-random-gpt2/8e370463308e6d06a809.json +1 -0
  9. neuronxcc-2.13.66.0+6dfecc895/0_REGISTRY/0.0.23.dev0/inference/gpt2/hf-internal-testing/tiny-random-gpt2/b99d0747aa4e9af275fb.json +1 -0
  10. neuronxcc-2.13.66.0+6dfecc895/0_REGISTRY/0.0.23.dev0/inference/gpt2/hf-internal-testing/tiny-random-gpt2/e62dd30300cbb6a57da5.json +1 -0
  11. neuronxcc-2.13.66.0+6dfecc895/0_REGISTRY/0.0.23.dev0/inference/gpt2/openai-community/gpt2/e31ae96d70b5049c0036.json +1 -0
  12. neuronxcc-2.13.66.0+6dfecc895/0_REGISTRY/0.0.23.dev0/inference/llama/dacorvo/tiny-random-llama/1ff321163e68478a1ce2.json +1 -0
  13. neuronxcc-2.13.66.0+6dfecc895/0_REGISTRY/0.0.23.dev0/inference/llama/dacorvo/tiny-random-llama/311a29bf60fb4f6082c6.json +1 -0
  14. neuronxcc-2.13.66.0+6dfecc895/0_REGISTRY/0.0.23.dev0/inference/llama/dacorvo/tiny-random-llama/60954a11dd58b586c979.json +1 -0
  15. neuronxcc-2.13.66.0+6dfecc895/0_REGISTRY/0.0.23.dev0/inference/llama/dacorvo/tiny-random-llama/fb4b47ea0eea68e736b2.json +1 -0
  16. neuronxcc-2.13.66.0+6dfecc895/0_REGISTRY/0.0.23.dev0/inference/llama/meta-llama/Meta-Llama-3-8B/c2f28b3cdba011f81417.json +1 -0
  17. neuronxcc-2.13.66.0+6dfecc895/0_REGISTRY/0.0.23.dev0/inference/mistral/dacorvo/tiny-random-MistralForCausalLM/52fbcfd7fedd024371f9.json +1 -0
  18. neuronxcc-2.13.66.0+6dfecc895/0_REGISTRY/0.0.23.dev0/inference/mistral/dacorvo/tiny-random-MistralForCausalLM/61da37a594d2621c84f3.json +1 -0
  19. neuronxcc-2.13.66.0+6dfecc895/0_REGISTRY/0.0.23.dev0/inference/mistral/dacorvo/tiny-random-MistralForCausalLM/8ab25062c2316baf852b.json +1 -0
  20. neuronxcc-2.13.66.0+6dfecc895/0_REGISTRY/0.0.23.dev0/inference/mistral/dacorvo/tiny-random-MistralForCausalLM/8adb053370d3399fa36d.json +1 -0
  21. neuronxcc-2.13.66.0+6dfecc895/0_REGISTRY/0.0.23.dev0/inference/mixtral/dacorvo/Mixtral-tiny/43944cbea36886b7c858.json +1 -0
  22. neuronxcc-2.13.66.0+6dfecc895/0_REGISTRY/0.0.23.dev0/inference/mixtral/dacorvo/Mixtral-tiny/52680b2cdc233c828caf.json +1 -0
  23. neuronxcc-2.13.66.0+6dfecc895/0_REGISTRY/0.0.23.dev0/inference/mixtral/dacorvo/Mixtral-tiny/57b16e8fab66604f85eb.json +1 -0
  24. neuronxcc-2.13.66.0+6dfecc895/0_REGISTRY/0.0.23.dev0/inference/mixtral/dacorvo/Mixtral-tiny/f6fbd647258f2c561a81.json +1 -0
  25. neuronxcc-2.13.66.0+6dfecc895/0_REGISTRY/0.0.23.dev0/inference/opt/hf-internal-testing/tiny-random-OPTForCausalLM/765507817dbb6125de6c.json +1 -0
  26. neuronxcc-2.13.66.0+6dfecc895/0_REGISTRY/0.0.23.dev0/inference/opt/hf-internal-testing/tiny-random-OPTForCausalLM/7f2d5e84740d788567e2.json +1 -0
  27. neuronxcc-2.13.66.0+6dfecc895/0_REGISTRY/0.0.23.dev0/inference/opt/hf-internal-testing/tiny-random-OPTForCausalLM/a8f94bb3f4eb93f8bd2e.json +1 -0
  28. neuronxcc-2.13.66.0+6dfecc895/0_REGISTRY/0.0.23.dev0/inference/opt/hf-internal-testing/tiny-random-OPTForCausalLM/f3adfe5a9c79b5a36fd7.json +1 -0
  29. neuronxcc-2.13.66.0+6dfecc895/MODULE_007ddbaadfdea74c4db5+2c2d707e/model.neff +1 -1
  30. neuronxcc-2.13.66.0+6dfecc895/MODULE_051daf37f1af2e155915+2c2d707e/compile_flags.txt +1 -0
  31. neuronxcc-2.13.66.0+6dfecc895/MODULE_051daf37f1af2e155915+2c2d707e/model.hlo_module.pb +3 -0
  32. neuronxcc-2.13.66.0+6dfecc895/MODULE_051daf37f1af2e155915+2c2d707e/model.neff +0 -0
  33. neuronxcc-2.13.66.0+6dfecc895/MODULE_06ec99d5ddf55b152eee+2c2d707e/compile_flags.txt +1 -0
  34. neuronxcc-2.13.66.0+6dfecc895/MODULE_06ec99d5ddf55b152eee+2c2d707e/model.hlo_module.pb +3 -0
  35. neuronxcc-2.13.66.0+6dfecc895/MODULE_06ec99d5ddf55b152eee+2c2d707e/model.neff +0 -0
  36. neuronxcc-2.13.66.0+6dfecc895/MODULE_0d2fdaf02cce617fdd0c+2c2d707e/compile_flags.txt +1 -0
  37. neuronxcc-2.13.66.0+6dfecc895/MODULE_0d2fdaf02cce617fdd0c+2c2d707e/model.hlo_module.pb +3 -0
  38. neuronxcc-2.13.66.0+6dfecc895/MODULE_0d2fdaf02cce617fdd0c+2c2d707e/model.neff +0 -0
  39. neuronxcc-2.13.66.0+6dfecc895/MODULE_250ee69b911da3bdc8d2+2c2d707e/compile_flags.txt +1 -0
  40. neuronxcc-2.13.66.0+6dfecc895/MODULE_250ee69b911da3bdc8d2+2c2d707e/model.hlo_module.pb +3 -0
  41. neuronxcc-2.13.66.0+6dfecc895/MODULE_250ee69b911da3bdc8d2+2c2d707e/model.neff +3 -0
  42. neuronxcc-2.13.66.0+6dfecc895/MODULE_2648f9d12f524b67bd88+2c2d707e/compile_flags.txt +1 -0
  43. neuronxcc-2.13.66.0+6dfecc895/MODULE_2648f9d12f524b67bd88+2c2d707e/model.hlo_module.pb +3 -0
  44. neuronxcc-2.13.66.0+6dfecc895/MODULE_2648f9d12f524b67bd88+2c2d707e/model.neff +0 -0
  45. neuronxcc-2.13.66.0+6dfecc895/MODULE_29cca2b907e9f3fdb2f8+2c2d707e/model.neff +1 -1
  46. neuronxcc-2.13.66.0+6dfecc895/MODULE_301a41e1f343ba3bec71+2c2d707e/model.neff +1 -1
  47. neuronxcc-2.13.66.0+6dfecc895/MODULE_34d558c33d28a2f62eeb+2c2d707e/compile_flags.txt +1 -0
  48. neuronxcc-2.13.66.0+6dfecc895/MODULE_34d558c33d28a2f62eeb+2c2d707e/model.hlo_module.pb +3 -0
  49. neuronxcc-2.13.66.0+6dfecc895/MODULE_34d558c33d28a2f62eeb+2c2d707e/model.neff +0 -0
  50. neuronxcc-2.13.66.0+6dfecc895/MODULE_35a5e2c0699bcf01dfe7+2c2d707e/compile_flags.txt +1 -0
.gitattributes CHANGED
@@ -2999,3 +2999,9 @@ neuronxcc-2.13.66.0+6dfecc895/MODULE_51908aacfbf851d85af4+2c2d707e/model.neff fi
2999
  neuronxcc-2.13.66.0+6dfecc895/MODULE_cef8491fe79ba64897ec+2c2d707e/model.neff filter=lfs diff=lfs merge=lfs -text
3000
  neuronxcc-2.13.66.0+6dfecc895/MODULE_00fe98d615fb072392ce+2c2d707e/model.neff filter=lfs diff=lfs merge=lfs -text
3001
  neuronxcc-2.13.66.0+6dfecc895/MODULE_05ffcb959c9e232f6071+2c2d707e/model.neff filter=lfs diff=lfs merge=lfs -text
 
 
 
 
 
 
 
2999
  neuronxcc-2.13.66.0+6dfecc895/MODULE_cef8491fe79ba64897ec+2c2d707e/model.neff filter=lfs diff=lfs merge=lfs -text
3000
  neuronxcc-2.13.66.0+6dfecc895/MODULE_00fe98d615fb072392ce+2c2d707e/model.neff filter=lfs diff=lfs merge=lfs -text
3001
  neuronxcc-2.13.66.0+6dfecc895/MODULE_05ffcb959c9e232f6071+2c2d707e/model.neff filter=lfs diff=lfs merge=lfs -text
3002
+ neuronxcc-2.13.66.0+6dfecc895/MODULE_250ee69b911da3bdc8d2+2c2d707e/model.neff filter=lfs diff=lfs merge=lfs -text
3003
+ neuronxcc-2.13.66.0+6dfecc895/MODULE_35a5e2c0699bcf01dfe7+2c2d707e/model.neff filter=lfs diff=lfs merge=lfs -text
3004
+ neuronxcc-2.13.66.0+6dfecc895/MODULE_42df9e1a521f5732af09+2c2d707e/model.neff filter=lfs diff=lfs merge=lfs -text
3005
+ neuronxcc-2.13.66.0+6dfecc895/MODULE_4cc7cdbf96054f333830+2c2d707e/model.neff filter=lfs diff=lfs merge=lfs -text
3006
+ neuronxcc-2.13.66.0+6dfecc895/MODULE_80b37e060185910ca232+2c2d707e/model.neff filter=lfs diff=lfs merge=lfs -text
3007
+ neuronxcc-2.13.66.0+6dfecc895/MODULE_e805a82c8d8a8e7ebfe1+2c2d707e/model.neff filter=lfs diff=lfs merge=lfs -text
neuronxcc-2.13.66.0+6dfecc895/0_REGISTRY/0.0.23.dev0/inference/bloom/hf-internal-testing/tiny-random-BloomForCausalLM/4efe4ec09e4490308cc0.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"apply_residual_connection_post_layernorm": false, "architectures": ["BloomForCausalLM"], "attention_dropout": 0.1, "bos_token_id": 1, "dtype": "float32", "eos_token_id": 2, "gradient_checkpointing": false, "hidden_dropout": 0.1, "hidden_size": 32, "id2label": {"0": "LABEL_0", "1": "LABEL_1", "2": "LABEL_2"}, "initializer_range": 0.02, "is_decoder": true, "label2id": {"LABEL_0": 0, "LABEL_1": 1, "LABEL_2": 2}, "layer_norm_epsilon": 1e-05, "model_type": "bloom", "n_head": 4, "n_layer": 5, "n_positions": 512, "neuron": {"auto_cast_type": "fp32", "batch_size": 1, "checkpoint_id": "hf-internal-testing/tiny-random-BloomForCausalLM", "checkpoint_revision": "92b07e9b7b4f986fa7c54e2ac3b9201b4ba5212e", "compiler_type": "neuronx-cc", "compiler_version": "2.13.66.0+6dfecc895", "num_cores": 2, "sequence_length": 100, "task": "text-generation"}, "pad_token_id": 3, "pretraining_tp": 1, "seq_length": 7, "slow_but_exact": true, "torch_dtype": "float32", "type_vocab_size": 16, "use_cache": true, "vocab_size": 1024}
neuronxcc-2.13.66.0+6dfecc895/0_REGISTRY/0.0.23.dev0/inference/bloom/hf-internal-testing/tiny-random-BloomForCausalLM/9d5f94b71f6ac3455612.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"apply_residual_connection_post_layernorm": false, "architectures": ["BloomForCausalLM"], "attention_dropout": 0.1, "bos_token_id": 1, "dtype": "float32", "eos_token_id": 2, "gradient_checkpointing": false, "hidden_dropout": 0.1, "hidden_size": 32, "id2label": {"0": "LABEL_0", "1": "LABEL_1", "2": "LABEL_2"}, "initializer_range": 0.02, "is_decoder": true, "label2id": {"LABEL_0": 0, "LABEL_1": 1, "LABEL_2": 2}, "layer_norm_epsilon": 1e-05, "model_type": "bloom", "n_head": 4, "n_layer": 5, "n_positions": 512, "neuron": {"auto_cast_type": "fp16", "batch_size": 2, "checkpoint_id": "hf-internal-testing/tiny-random-BloomForCausalLM", "checkpoint_revision": "92b07e9b7b4f986fa7c54e2ac3b9201b4ba5212e", "compiler_type": "neuronx-cc", "compiler_version": "2.13.66.0+6dfecc895", "num_cores": 2, "sequence_length": 100, "task": "text-generation"}, "pad_token_id": 3, "pretraining_tp": 1, "seq_length": 7, "slow_but_exact": true, "torch_dtype": "float32", "type_vocab_size": 16, "use_cache": true, "vocab_size": 1024}
neuronxcc-2.13.66.0+6dfecc895/0_REGISTRY/0.0.23.dev0/inference/bloom/hf-internal-testing/tiny-random-BloomForCausalLM/b9db5e6da88e9f07ee3e.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"apply_residual_connection_post_layernorm": false, "architectures": ["BloomForCausalLM"], "attention_dropout": 0.1, "bos_token_id": 1, "dtype": "float32", "eos_token_id": 2, "gradient_checkpointing": false, "hidden_dropout": 0.1, "hidden_size": 32, "id2label": {"0": "LABEL_0", "1": "LABEL_1", "2": "LABEL_2"}, "initializer_range": 0.02, "is_decoder": true, "label2id": {"LABEL_0": 0, "LABEL_1": 1, "LABEL_2": 2}, "layer_norm_epsilon": 1e-05, "model_type": "bloom", "n_head": 4, "n_layer": 5, "n_positions": 512, "neuron": {"auto_cast_type": "fp16", "batch_size": 1, "checkpoint_id": "hf-internal-testing/tiny-random-BloomForCausalLM", "checkpoint_revision": "92b07e9b7b4f986fa7c54e2ac3b9201b4ba5212e", "compiler_type": "neuronx-cc", "compiler_version": "2.13.66.0+6dfecc895", "num_cores": 2, "sequence_length": 100, "task": "text-generation"}, "pad_token_id": 3, "pretraining_tp": 1, "seq_length": 7, "slow_but_exact": true, "torch_dtype": "float32", "type_vocab_size": 16, "use_cache": true, "vocab_size": 1024}
neuronxcc-2.13.66.0+6dfecc895/0_REGISTRY/0.0.23.dev0/inference/bloom/hf-internal-testing/tiny-random-BloomForCausalLM/f9a5439ee67f962284df.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"apply_residual_connection_post_layernorm": false, "architectures": ["BloomForCausalLM"], "attention_dropout": 0.1, "bos_token_id": 1, "dtype": "float32", "eos_token_id": 2, "gradient_checkpointing": false, "hidden_dropout": 0.1, "hidden_size": 32, "id2label": {"0": "LABEL_0", "1": "LABEL_1", "2": "LABEL_2"}, "initializer_range": 0.02, "is_decoder": true, "label2id": {"LABEL_0": 0, "LABEL_1": 1, "LABEL_2": 2}, "layer_norm_epsilon": 1e-05, "model_type": "bloom", "n_head": 4, "n_layer": 5, "n_positions": 512, "neuron": {"auto_cast_type": "fp32", "batch_size": 2, "checkpoint_id": "hf-internal-testing/tiny-random-BloomForCausalLM", "checkpoint_revision": "92b07e9b7b4f986fa7c54e2ac3b9201b4ba5212e", "compiler_type": "neuronx-cc", "compiler_version": "2.13.66.0+6dfecc895", "num_cores": 2, "sequence_length": 512, "task": "text-generation"}, "pad_token_id": 3, "pretraining_tp": 1, "seq_length": 7, "slow_but_exact": true, "torch_dtype": "float32", "type_vocab_size": 16, "use_cache": true, "vocab_size": 1024}
neuronxcc-2.13.66.0+6dfecc895/0_REGISTRY/0.0.23.dev0/inference/gpt2/hf-internal-testing/tiny-random-gpt2/2713954a7f357d88b849.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"activation_function": "gelu_new", "attention_probs_dropout_prob": 0.1, "attn_pdrop": 0.1, "bos_token_id": 98, "embd_pdrop": 0.1, "eos_token_id": 98, "gradient_checkpointing": false, "hidden_act": "gelu", "hidden_dropout_prob": 0.1, "initializer_range": 0.02, "intermediate_size": 37, "layer_norm_epsilon": 1e-05, "model_type": "gpt2", "n_ctx": 512, "n_embd": 32, "n_head": 4, "n_inner": null, "n_layer": 5, "n_positions": 512, "neuron": {"auto_cast_type": "fp32", "batch_size": 2, "checkpoint_id": "hf-internal-testing/tiny-random-gpt2", "checkpoint_revision": "71034c5d8bde858ff824298bdedc65515b97d2b9", "compiler_type": "neuronx-cc", "compiler_version": "2.13.66.0+6dfecc895", "num_cores": 2, "sequence_length": 512, "task": "text-generation"}, "pad_token_id": 98, "reorder_and_upcast_attn": false, "resid_pdrop": 0.1, "scale_attn_by_inverse_layer_idx": false, "scale_attn_weights": true, "summary_activation": null, "summary_first_dropout": 0.1, "summary_proj_to_labels": true, "summary_type": "cls_index", "summary_use_proj": true, "type_vocab_size": 16, "use_cache": true, "vocab_size": 1000}
neuronxcc-2.13.66.0+6dfecc895/0_REGISTRY/0.0.23.dev0/inference/gpt2/hf-internal-testing/tiny-random-gpt2/7e574905d10c185da6d3.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"activation_function": "gelu_new", "attention_probs_dropout_prob": 0.1, "attn_pdrop": 0.1, "bos_token_id": 98, "embd_pdrop": 0.1, "eos_token_id": 98, "gradient_checkpointing": false, "hidden_act": "gelu", "hidden_dropout_prob": 0.1, "initializer_range": 0.02, "intermediate_size": 37, "layer_norm_epsilon": 1e-05, "model_type": "gpt2", "n_ctx": 512, "n_embd": 32, "n_head": 4, "n_inner": null, "n_layer": 5, "n_positions": 512, "neuron": {"auto_cast_type": "fp16", "batch_size": 1, "checkpoint_id": "hf-internal-testing/tiny-random-gpt2", "checkpoint_revision": "71034c5d8bde858ff824298bdedc65515b97d2b9", "compiler_type": "neuronx-cc", "compiler_version": "2.13.66.0+6dfecc895", "num_cores": 2, "sequence_length": 100, "task": "text-generation"}, "pad_token_id": 98, "reorder_and_upcast_attn": false, "resid_pdrop": 0.1, "scale_attn_by_inverse_layer_idx": false, "scale_attn_weights": true, "summary_activation": null, "summary_first_dropout": 0.1, "summary_proj_to_labels": true, "summary_type": "cls_index", "summary_use_proj": true, "type_vocab_size": 16, "use_cache": true, "vocab_size": 1000}
neuronxcc-2.13.66.0+6dfecc895/0_REGISTRY/0.0.23.dev0/inference/gpt2/hf-internal-testing/tiny-random-gpt2/8e370463308e6d06a809.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"activation_function": "gelu_new", "attention_probs_dropout_prob": 0.1, "attn_pdrop": 0.1, "bos_token_id": 98, "embd_pdrop": 0.1, "eos_token_id": 98, "gradient_checkpointing": false, "hidden_act": "gelu", "hidden_dropout_prob": 0.1, "initializer_range": 0.02, "intermediate_size": 37, "layer_norm_epsilon": 1e-05, "model_type": "gpt2", "n_ctx": 512, "n_embd": 32, "n_head": 4, "n_inner": null, "n_layer": 5, "n_positions": 512, "neuron": {"auto_cast_type": "fp32", "batch_size": 16, "checkpoint_id": "hf-internal-testing/tiny-random-gpt2", "checkpoint_revision": "91c0fe31d692dd8448d9bc06e8d1877345009e3b", "compiler_type": "neuronx-cc", "compiler_version": "2.12.54.0+f631c2365", "num_cores": 2, "sequence_length": 512, "task": "text-generation"}, "pad_token_id": 98, "reorder_and_upcast_attn": false, "resid_pdrop": 0.1, "scale_attn_by_inverse_layer_idx": false, "scale_attn_weights": true, "summary_activation": null, "summary_first_dropout": 0.1, "summary_proj_to_labels": true, "summary_type": "cls_index", "summary_use_proj": true, "type_vocab_size": 16, "use_cache": true, "vocab_size": 1000}
neuronxcc-2.13.66.0+6dfecc895/0_REGISTRY/0.0.23.dev0/inference/gpt2/hf-internal-testing/tiny-random-gpt2/b99d0747aa4e9af275fb.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"activation_function": "gelu_new", "attention_probs_dropout_prob": 0.1, "attn_pdrop": 0.1, "bos_token_id": 98, "embd_pdrop": 0.1, "eos_token_id": 98, "gradient_checkpointing": false, "hidden_act": "gelu", "hidden_dropout_prob": 0.1, "initializer_range": 0.02, "intermediate_size": 37, "layer_norm_epsilon": 1e-05, "model_type": "gpt2", "n_ctx": 512, "n_embd": 32, "n_head": 4, "n_inner": null, "n_layer": 5, "n_positions": 512, "neuron": {"auto_cast_type": "fp32", "batch_size": 1, "checkpoint_id": "hf-internal-testing/tiny-random-gpt2", "checkpoint_revision": "71034c5d8bde858ff824298bdedc65515b97d2b9", "compiler_type": "neuronx-cc", "compiler_version": "2.13.66.0+6dfecc895", "num_cores": 2, "sequence_length": 100, "task": "text-generation"}, "pad_token_id": 98, "reorder_and_upcast_attn": false, "resid_pdrop": 0.1, "scale_attn_by_inverse_layer_idx": false, "scale_attn_weights": true, "summary_activation": null, "summary_first_dropout": 0.1, "summary_proj_to_labels": true, "summary_type": "cls_index", "summary_use_proj": true, "type_vocab_size": 16, "use_cache": true, "vocab_size": 1000}
neuronxcc-2.13.66.0+6dfecc895/0_REGISTRY/0.0.23.dev0/inference/gpt2/hf-internal-testing/tiny-random-gpt2/e62dd30300cbb6a57da5.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"activation_function": "gelu_new", "attention_probs_dropout_prob": 0.1, "attn_pdrop": 0.1, "bos_token_id": 98, "embd_pdrop": 0.1, "eos_token_id": 98, "gradient_checkpointing": false, "hidden_act": "gelu", "hidden_dropout_prob": 0.1, "initializer_range": 0.02, "intermediate_size": 37, "layer_norm_epsilon": 1e-05, "model_type": "gpt2", "n_ctx": 512, "n_embd": 32, "n_head": 4, "n_inner": null, "n_layer": 5, "n_positions": 512, "neuron": {"auto_cast_type": "fp16", "batch_size": 2, "checkpoint_id": "hf-internal-testing/tiny-random-gpt2", "checkpoint_revision": "71034c5d8bde858ff824298bdedc65515b97d2b9", "compiler_type": "neuronx-cc", "compiler_version": "2.13.66.0+6dfecc895", "num_cores": 2, "sequence_length": 100, "task": "text-generation"}, "pad_token_id": 98, "reorder_and_upcast_attn": false, "resid_pdrop": 0.1, "scale_attn_by_inverse_layer_idx": false, "scale_attn_weights": true, "summary_activation": null, "summary_first_dropout": 0.1, "summary_proj_to_labels": true, "summary_type": "cls_index", "summary_use_proj": true, "type_vocab_size": 16, "use_cache": true, "vocab_size": 1000}
neuronxcc-2.13.66.0+6dfecc895/0_REGISTRY/0.0.23.dev0/inference/gpt2/openai-community/gpt2/e31ae96d70b5049c0036.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"activation_function": "gelu_new", "architectures": ["GPT2LMHeadModel"], "attn_pdrop": 0.1, "bos_token_id": 50256, "embd_pdrop": 0.1, "eos_token_id": 50256, "initializer_range": 0.02, "layer_norm_epsilon": 1e-05, "model_type": "gpt2", "n_ctx": 1024, "n_embd": 768, "n_head": 12, "n_inner": null, "n_layer": 12, "n_positions": 1024, "neuron": {"auto_cast_type": "fp16", "batch_size": 1, "checkpoint_id": "openai-community/gpt2", "checkpoint_revision": "607a30d783dfa663caf39e06633721c8d4cfcd7e", "compiler_type": "neuronx-cc", "compiler_version": "2.13.66.0+6dfecc895", "num_cores": 2, "sequence_length": 1024, "task": "text-generation"}, "reorder_and_upcast_attn": false, "resid_pdrop": 0.1, "scale_attn_by_inverse_layer_idx": false, "scale_attn_weights": true, "summary_activation": null, "summary_first_dropout": 0.1, "summary_proj_to_labels": true, "summary_type": "cls_index", "summary_use_proj": true, "task_specific_params": {"text-generation": {"do_sample": true, "max_length": 50}}, "use_cache": true, "vocab_size": 50257}
neuronxcc-2.13.66.0+6dfecc895/0_REGISTRY/0.0.23.dev0/inference/llama/dacorvo/tiny-random-llama/1ff321163e68478a1ce2.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"architectures": ["LlamaForCausalLM"], "attention_bias": false, "attention_dropout": 0.0, "bos_token_id": 1, "eos_token_id": 2, "hidden_act": "silu", "hidden_size": 128, "initializer_range": 0.02, "intermediate_size": 256, "max_position_embeddings": 512, "model_type": "llama", "neuron": {"auto_cast_type": "fp16", "batch_size": 1, "checkpoint_id": "dacorvo/tiny-random-llama", "checkpoint_revision": "7fdafd2fe6a2d31c6abb72ae60db606d8bb23196", "compiler_type": "neuronx-cc", "compiler_version": "2.13.66.0+6dfecc895", "num_cores": 2, "sequence_length": 100, "task": "text-generation"}, "num_attention_heads": 1, "num_hidden_layers": 1, "num_key_value_heads": 1, "pretraining_tp": 1, "rms_norm_eps": 1e-06, "rope_scaling": null, "rope_theta": 10000.0, "tie_word_embeddings": false, "torch_dtype": "float32", "use_cache": true, "vocab_size": 32000}
neuronxcc-2.13.66.0+6dfecc895/0_REGISTRY/0.0.23.dev0/inference/llama/dacorvo/tiny-random-llama/311a29bf60fb4f6082c6.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"architectures": ["LlamaForCausalLM"], "attention_bias": false, "attention_dropout": 0.0, "bos_token_id": 1, "eos_token_id": 2, "hidden_act": "silu", "hidden_size": 128, "initializer_range": 0.02, "intermediate_size": 256, "max_position_embeddings": 512, "model_type": "llama", "neuron": {"auto_cast_type": "fp32", "batch_size": 1, "checkpoint_id": "dacorvo/tiny-random-llama", "checkpoint_revision": "7fdafd2fe6a2d31c6abb72ae60db606d8bb23196", "compiler_type": "neuronx-cc", "compiler_version": "2.13.66.0+6dfecc895", "num_cores": 2, "sequence_length": 100, "task": "text-generation"}, "num_attention_heads": 1, "num_hidden_layers": 1, "num_key_value_heads": 1, "pretraining_tp": 1, "rms_norm_eps": 1e-06, "rope_scaling": null, "rope_theta": 10000.0, "tie_word_embeddings": false, "torch_dtype": "float32", "use_cache": true, "vocab_size": 32000}
neuronxcc-2.13.66.0+6dfecc895/0_REGISTRY/0.0.23.dev0/inference/llama/dacorvo/tiny-random-llama/60954a11dd58b586c979.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"architectures": ["LlamaForCausalLM"], "attention_bias": false, "attention_dropout": 0.0, "bos_token_id": 1, "eos_token_id": 2, "hidden_act": "silu", "hidden_size": 128, "initializer_range": 0.02, "intermediate_size": 256, "max_position_embeddings": 512, "model_type": "llama", "neuron": {"auto_cast_type": "fp16", "batch_size": 2, "checkpoint_id": "dacorvo/tiny-random-llama", "checkpoint_revision": "7fdafd2fe6a2d31c6abb72ae60db606d8bb23196", "compiler_type": "neuronx-cc", "compiler_version": "2.13.66.0+6dfecc895", "num_cores": 2, "sequence_length": 100, "task": "text-generation"}, "num_attention_heads": 1, "num_hidden_layers": 1, "num_key_value_heads": 1, "pretraining_tp": 1, "rms_norm_eps": 1e-06, "rope_scaling": null, "rope_theta": 10000.0, "tie_word_embeddings": false, "torch_dtype": "float32", "use_cache": true, "vocab_size": 32000}
neuronxcc-2.13.66.0+6dfecc895/0_REGISTRY/0.0.23.dev0/inference/llama/dacorvo/tiny-random-llama/fb4b47ea0eea68e736b2.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"architectures": ["LlamaForCausalLM"], "attention_bias": false, "attention_dropout": 0.0, "bos_token_id": 1, "eos_token_id": 2, "hidden_act": "silu", "hidden_size": 128, "initializer_range": 0.02, "intermediate_size": 256, "max_position_embeddings": 512, "model_type": "llama", "neuron": {"auto_cast_type": "fp32", "batch_size": 2, "checkpoint_id": "dacorvo/tiny-random-llama", "checkpoint_revision": "7fdafd2fe6a2d31c6abb72ae60db606d8bb23196", "compiler_type": "neuronx-cc", "compiler_version": "2.13.66.0+6dfecc895", "num_cores": 2, "sequence_length": 512, "task": "text-generation"}, "num_attention_heads": 1, "num_hidden_layers": 1, "num_key_value_heads": 1, "pretraining_tp": 1, "rms_norm_eps": 1e-06, "rope_scaling": null, "rope_theta": 10000.0, "tie_word_embeddings": false, "torch_dtype": "float32", "use_cache": true, "vocab_size": 32000}
neuronxcc-2.13.66.0+6dfecc895/0_REGISTRY/0.0.23.dev0/inference/llama/meta-llama/Meta-Llama-3-8B/c2f28b3cdba011f81417.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"architectures": ["LlamaForCausalLM"], "attention_bias": false, "attention_dropout": 0.0, "bos_token_id": 128000, "eos_token_id": 128001, "hidden_act": "silu", "hidden_size": 4096, "initializer_range": 0.02, "intermediate_size": 14336, "max_position_embeddings": 8192, "model_type": "llama", "neuron": {"auto_cast_type": "fp16", "batch_size": 32, "checkpoint_id": "meta-llama/Meta-Llama-3-8B", "checkpoint_revision": "62bd457b6fe961a42a631306577e622c83876cb6", "compiler_type": "neuronx-cc", "compiler_version": "2.13.66.0+6dfecc895", "num_cores": 8, "sequence_length": 4096, "task": "text-generation"}, "num_attention_heads": 32, "num_hidden_layers": 32, "num_key_value_heads": 8, "pretraining_tp": 1, "rms_norm_eps": 1e-05, "rope_scaling": null, "rope_theta": 500000.0, "tie_word_embeddings": false, "torch_dtype": "bfloat16", "use_cache": true, "vocab_size": 128256}
neuronxcc-2.13.66.0+6dfecc895/0_REGISTRY/0.0.23.dev0/inference/mistral/dacorvo/tiny-random-MistralForCausalLM/52fbcfd7fedd024371f9.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"architectures": ["MistralForCausalLM"], "attention_dropout": 0.0, "attention_probs_dropout_prob": 0.1, "bos_token_id": 1, "eos_token_id": 2, "hidden_act": "gelu", "hidden_dropout_prob": 0.1, "hidden_size": 32, "initializer_range": 0.02, "intermediate_size": 37, "is_decoder": true, "max_position_embeddings": 512, "model_type": "mistral", "neuron": {"auto_cast_type": "fp16", "batch_size": 2, "checkpoint_id": "dacorvo/tiny-random-MistralForCausalLM", "checkpoint_revision": "81d453e3c8985649e9ee3d4c9378461029d1c73a", "compiler_type": "neuronx-cc", "compiler_version": "2.13.66.0+6dfecc895", "num_cores": 2, "sequence_length": 100, "task": "text-generation"}, "num_attention_heads": 4, "num_hidden_layers": 2, "num_key_value_heads": 2, "pad_token_id": 0, "rms_norm_eps": 1e-06, "rope_theta": 10000.0, "sliding_window": 4096, "tie_word_embeddings": false, "torch_dtype": "float32", "type_vocab_size": 16, "use_cache": true, "vocab_size": 32000}
neuronxcc-2.13.66.0+6dfecc895/0_REGISTRY/0.0.23.dev0/inference/mistral/dacorvo/tiny-random-MistralForCausalLM/61da37a594d2621c84f3.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"architectures": ["MistralForCausalLM"], "attention_dropout": 0.0, "attention_probs_dropout_prob": 0.1, "bos_token_id": 1, "eos_token_id": 2, "hidden_act": "gelu", "hidden_dropout_prob": 0.1, "hidden_size": 32, "initializer_range": 0.02, "intermediate_size": 37, "is_decoder": true, "max_position_embeddings": 512, "model_type": "mistral", "neuron": {"auto_cast_type": "fp16", "batch_size": 1, "checkpoint_id": "dacorvo/tiny-random-MistralForCausalLM", "checkpoint_revision": "81d453e3c8985649e9ee3d4c9378461029d1c73a", "compiler_type": "neuronx-cc", "compiler_version": "2.13.66.0+6dfecc895", "num_cores": 2, "sequence_length": 100, "task": "text-generation"}, "num_attention_heads": 4, "num_hidden_layers": 2, "num_key_value_heads": 2, "pad_token_id": 0, "rms_norm_eps": 1e-06, "rope_theta": 10000.0, "sliding_window": 4096, "tie_word_embeddings": false, "torch_dtype": "float32", "type_vocab_size": 16, "use_cache": true, "vocab_size": 32000}
neuronxcc-2.13.66.0+6dfecc895/0_REGISTRY/0.0.23.dev0/inference/mistral/dacorvo/tiny-random-MistralForCausalLM/8ab25062c2316baf852b.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"architectures": ["MistralForCausalLM"], "attention_dropout": 0.0, "attention_probs_dropout_prob": 0.1, "bos_token_id": 1, "eos_token_id": 2, "hidden_act": "gelu", "hidden_dropout_prob": 0.1, "hidden_size": 32, "initializer_range": 0.02, "intermediate_size": 37, "is_decoder": true, "max_position_embeddings": 512, "model_type": "mistral", "neuron": {"auto_cast_type": "fp32", "batch_size": 2, "checkpoint_id": "dacorvo/tiny-random-MistralForCausalLM", "checkpoint_revision": "81d453e3c8985649e9ee3d4c9378461029d1c73a", "compiler_type": "neuronx-cc", "compiler_version": "2.13.66.0+6dfecc895", "num_cores": 2, "sequence_length": 512, "task": "text-generation"}, "num_attention_heads": 4, "num_hidden_layers": 2, "num_key_value_heads": 2, "pad_token_id": 0, "rms_norm_eps": 1e-06, "rope_theta": 10000.0, "sliding_window": 4096, "tie_word_embeddings": false, "torch_dtype": "float32", "type_vocab_size": 16, "use_cache": true, "vocab_size": 32000}
neuronxcc-2.13.66.0+6dfecc895/0_REGISTRY/0.0.23.dev0/inference/mistral/dacorvo/tiny-random-MistralForCausalLM/8adb053370d3399fa36d.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"architectures": ["MistralForCausalLM"], "attention_dropout": 0.0, "attention_probs_dropout_prob": 0.1, "bos_token_id": 1, "eos_token_id": 2, "hidden_act": "gelu", "hidden_dropout_prob": 0.1, "hidden_size": 32, "initializer_range": 0.02, "intermediate_size": 37, "is_decoder": true, "max_position_embeddings": 512, "model_type": "mistral", "neuron": {"auto_cast_type": "fp32", "batch_size": 1, "checkpoint_id": "dacorvo/tiny-random-MistralForCausalLM", "checkpoint_revision": "81d453e3c8985649e9ee3d4c9378461029d1c73a", "compiler_type": "neuronx-cc", "compiler_version": "2.13.66.0+6dfecc895", "num_cores": 2, "sequence_length": 100, "task": "text-generation"}, "num_attention_heads": 4, "num_hidden_layers": 2, "num_key_value_heads": 2, "pad_token_id": 0, "rms_norm_eps": 1e-06, "rope_theta": 10000.0, "sliding_window": 4096, "tie_word_embeddings": false, "torch_dtype": "float32", "type_vocab_size": 16, "use_cache": true, "vocab_size": 32000}
neuronxcc-2.13.66.0+6dfecc895/0_REGISTRY/0.0.23.dev0/inference/mixtral/dacorvo/Mixtral-tiny/43944cbea36886b7c858.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"architectures": ["MixtralForCausalLM"], "attention_dropout": 0.0, "bos_token_id": 1, "eos_token_id": 2, "hidden_act": "silu", "hidden_size": 1024, "initializer_range": 0.02, "intermediate_size": 3584, "max_position_embeddings": 1024, "model_type": "mixtral", "neuron": {"auto_cast_type": "fp32", "batch_size": 1, "checkpoint_id": "dacorvo/Mixtral-tiny", "checkpoint_revision": "c557ba205ddff6ea911f4719e0d543d6c08356b6", "compiler_type": "neuronx-cc", "compiler_version": "2.13.66.0+6dfecc895", "num_cores": 2, "sequence_length": 100, "task": "text-generation"}, "num_attention_heads": 32, "num_experts_per_tok": 2, "num_hidden_layers": 2, "num_key_value_heads": 8, "num_local_experts": 8, "output_router_logits": false, "rms_norm_eps": 1e-05, "rope_theta": 10000.0, "router_aux_loss_coef": 0.001, "router_jitter_noise": 0.0, "sliding_window": 4096, "tie_word_embeddings": false, "torch_dtype": "float16", "use_cache": true, "vocab_size": 32000}
neuronxcc-2.13.66.0+6dfecc895/0_REGISTRY/0.0.23.dev0/inference/mixtral/dacorvo/Mixtral-tiny/52680b2cdc233c828caf.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"architectures": ["MixtralForCausalLM"], "attention_dropout": 0.0, "bos_token_id": 1, "eos_token_id": 2, "hidden_act": "silu", "hidden_size": 1024, "initializer_range": 0.02, "intermediate_size": 3584, "max_position_embeddings": 1024, "model_type": "mixtral", "neuron": {"auto_cast_type": "fp32", "batch_size": 2, "checkpoint_id": "dacorvo/Mixtral-tiny", "checkpoint_revision": "c557ba205ddff6ea911f4719e0d543d6c08356b6", "compiler_type": "neuronx-cc", "compiler_version": "2.13.66.0+6dfecc895", "num_cores": 2, "sequence_length": 1024, "task": "text-generation"}, "num_attention_heads": 32, "num_experts_per_tok": 2, "num_hidden_layers": 2, "num_key_value_heads": 8, "num_local_experts": 8, "output_router_logits": false, "rms_norm_eps": 1e-05, "rope_theta": 10000.0, "router_aux_loss_coef": 0.001, "router_jitter_noise": 0.0, "sliding_window": 4096, "tie_word_embeddings": false, "torch_dtype": "float16", "use_cache": true, "vocab_size": 32000}
neuronxcc-2.13.66.0+6dfecc895/0_REGISTRY/0.0.23.dev0/inference/mixtral/dacorvo/Mixtral-tiny/57b16e8fab66604f85eb.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"architectures": ["MixtralForCausalLM"], "attention_dropout": 0.0, "bos_token_id": 1, "eos_token_id": 2, "hidden_act": "silu", "hidden_size": 1024, "initializer_range": 0.02, "intermediate_size": 3584, "max_position_embeddings": 1024, "model_type": "mixtral", "neuron": {"auto_cast_type": "fp16", "batch_size": 2, "checkpoint_id": "dacorvo/Mixtral-tiny", "checkpoint_revision": "c557ba205ddff6ea911f4719e0d543d6c08356b6", "compiler_type": "neuronx-cc", "compiler_version": "2.13.66.0+6dfecc895", "num_cores": 2, "sequence_length": 100, "task": "text-generation"}, "num_attention_heads": 32, "num_experts_per_tok": 2, "num_hidden_layers": 2, "num_key_value_heads": 8, "num_local_experts": 8, "output_router_logits": false, "rms_norm_eps": 1e-05, "rope_theta": 10000.0, "router_aux_loss_coef": 0.001, "router_jitter_noise": 0.0, "sliding_window": 4096, "tie_word_embeddings": false, "torch_dtype": "float16", "use_cache": true, "vocab_size": 32000}
neuronxcc-2.13.66.0+6dfecc895/0_REGISTRY/0.0.23.dev0/inference/mixtral/dacorvo/Mixtral-tiny/f6fbd647258f2c561a81.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"architectures": ["MixtralForCausalLM"], "attention_dropout": 0.0, "bos_token_id": 1, "eos_token_id": 2, "hidden_act": "silu", "hidden_size": 1024, "initializer_range": 0.02, "intermediate_size": 3584, "max_position_embeddings": 1024, "model_type": "mixtral", "neuron": {"auto_cast_type": "fp16", "batch_size": 1, "checkpoint_id": "dacorvo/Mixtral-tiny", "checkpoint_revision": "c557ba205ddff6ea911f4719e0d543d6c08356b6", "compiler_type": "neuronx-cc", "compiler_version": "2.13.66.0+6dfecc895", "num_cores": 2, "sequence_length": 100, "task": "text-generation"}, "num_attention_heads": 32, "num_experts_per_tok": 2, "num_hidden_layers": 2, "num_key_value_heads": 8, "num_local_experts": 8, "output_router_logits": false, "rms_norm_eps": 1e-05, "rope_theta": 10000.0, "router_aux_loss_coef": 0.001, "router_jitter_noise": 0.0, "sliding_window": 4096, "tie_word_embeddings": false, "torch_dtype": "float16", "use_cache": true, "vocab_size": 32000}
neuronxcc-2.13.66.0+6dfecc895/0_REGISTRY/0.0.23.dev0/inference/opt/hf-internal-testing/tiny-random-OPTForCausalLM/765507817dbb6125de6c.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"_remove_final_layer_norm": false, "activation_function": "relu", "architectures": ["OPTForCausalLM"], "attention_dropout": 0.1, "bos_token_id": 2, "do_layer_norm_before": true, "dropout": 0.1, "embed_dim": 16, "enable_bias": true, "eos_token_id": 2, "ffn_dim": 4, "hidden_size": 16, "init_std": 0.02, "is_decoder": true, "layer_norm_elementwise_affine": true, "layerdrop": 0.0, "max_position_embeddings": 100, "model_type": "opt", "neuron": {"auto_cast_type": "fp16", "batch_size": 2, "checkpoint_id": "hf-internal-testing/tiny-random-OPTForCausalLM", "checkpoint_revision": "190d1f4fc0011d2eaeaa05282e0fbd2445e4b11f", "compiler_type": "neuronx-cc", "compiler_version": "2.13.66.0+6dfecc895", "num_cores": 2, "sequence_length": 100, "task": "text-generation"}, "num_attention_heads": 4, "num_hidden_layers": 5, "pad_token_id": 1, "torch_dtype": "float32", "use_cache": true, "vocab_size": 50265, "word_embed_proj_dim": 16}
neuronxcc-2.13.66.0+6dfecc895/0_REGISTRY/0.0.23.dev0/inference/opt/hf-internal-testing/tiny-random-OPTForCausalLM/7f2d5e84740d788567e2.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"_remove_final_layer_norm": false, "activation_function": "relu", "architectures": ["OPTForCausalLM"], "attention_dropout": 0.1, "bos_token_id": 2, "do_layer_norm_before": true, "dropout": 0.1, "embed_dim": 16, "enable_bias": true, "eos_token_id": 2, "ffn_dim": 4, "hidden_size": 16, "init_std": 0.02, "is_decoder": true, "layer_norm_elementwise_affine": true, "layerdrop": 0.0, "max_position_embeddings": 100, "model_type": "opt", "neuron": {"auto_cast_type": "fp32", "batch_size": 1, "checkpoint_id": "hf-internal-testing/tiny-random-OPTForCausalLM", "checkpoint_revision": "190d1f4fc0011d2eaeaa05282e0fbd2445e4b11f", "compiler_type": "neuronx-cc", "compiler_version": "2.13.66.0+6dfecc895", "num_cores": 2, "sequence_length": 100, "task": "text-generation"}, "num_attention_heads": 4, "num_hidden_layers": 5, "pad_token_id": 1, "torch_dtype": "float32", "use_cache": true, "vocab_size": 50265, "word_embed_proj_dim": 16}
neuronxcc-2.13.66.0+6dfecc895/0_REGISTRY/0.0.23.dev0/inference/opt/hf-internal-testing/tiny-random-OPTForCausalLM/a8f94bb3f4eb93f8bd2e.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"_remove_final_layer_norm": false, "activation_function": "relu", "architectures": ["OPTForCausalLM"], "attention_dropout": 0.1, "bos_token_id": 2, "do_layer_norm_before": true, "dropout": 0.1, "embed_dim": 16, "enable_bias": true, "eos_token_id": 2, "ffn_dim": 4, "hidden_size": 16, "init_std": 0.02, "is_decoder": true, "layer_norm_elementwise_affine": true, "layerdrop": 0.0, "max_position_embeddings": 100, "model_type": "opt", "neuron": {"auto_cast_type": "fp16", "batch_size": 1, "checkpoint_id": "hf-internal-testing/tiny-random-OPTForCausalLM", "checkpoint_revision": "190d1f4fc0011d2eaeaa05282e0fbd2445e4b11f", "compiler_type": "neuronx-cc", "compiler_version": "2.13.66.0+6dfecc895", "num_cores": 2, "sequence_length": 100, "task": "text-generation"}, "num_attention_heads": 4, "num_hidden_layers": 5, "pad_token_id": 1, "torch_dtype": "float32", "use_cache": true, "vocab_size": 50265, "word_embed_proj_dim": 16}
neuronxcc-2.13.66.0+6dfecc895/0_REGISTRY/0.0.23.dev0/inference/opt/hf-internal-testing/tiny-random-OPTForCausalLM/f3adfe5a9c79b5a36fd7.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"_remove_final_layer_norm": false, "activation_function": "relu", "architectures": ["OPTForCausalLM"], "attention_dropout": 0.1, "bos_token_id": 2, "do_layer_norm_before": true, "dropout": 0.1, "embed_dim": 16, "enable_bias": true, "eos_token_id": 2, "ffn_dim": 4, "hidden_size": 16, "init_std": 0.02, "is_decoder": true, "layer_norm_elementwise_affine": true, "layerdrop": 0.0, "max_position_embeddings": 100, "model_type": "opt", "neuron": {"auto_cast_type": "fp32", "batch_size": 2, "checkpoint_id": "hf-internal-testing/tiny-random-OPTForCausalLM", "checkpoint_revision": "190d1f4fc0011d2eaeaa05282e0fbd2445e4b11f", "compiler_type": "neuronx-cc", "compiler_version": "2.13.66.0+6dfecc895", "num_cores": 2, "sequence_length": 100, "task": "text-generation"}, "num_attention_heads": 4, "num_hidden_layers": 5, "pad_token_id": 1, "torch_dtype": "float32", "use_cache": true, "vocab_size": 50265, "word_embed_proj_dim": 16}
neuronxcc-2.13.66.0+6dfecc895/MODULE_007ddbaadfdea74c4db5+2c2d707e/model.neff CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:8559beec2df842857f647f9921c6ff30f0f3f20a4e17b0e3c7142e24902f7d90
3
  size 13282304
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:3ad504235ae58759a762486abdfcb61460190608c0d5d68eb7748bef1e742863
3
  size 13282304
neuronxcc-2.13.66.0+6dfecc895/MODULE_051daf37f1af2e155915+2c2d707e/compile_flags.txt ADDED
@@ -0,0 +1 @@
 
 
1
+ --model-type=transformer --auto-cast=none
neuronxcc-2.13.66.0+6dfecc895/MODULE_051daf37f1af2e155915+2c2d707e/model.hlo_module.pb ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:7df0b4f289f40a056e30678ff8894da8b4f011484c65dcdd4b349501489fde16
3
+ size 55020
neuronxcc-2.13.66.0+6dfecc895/MODULE_051daf37f1af2e155915+2c2d707e/model.neff ADDED
Binary file (134 kB). View file
 
neuronxcc-2.13.66.0+6dfecc895/MODULE_06ec99d5ddf55b152eee+2c2d707e/compile_flags.txt ADDED
@@ -0,0 +1 @@
 
 
1
+ --model-type=transformer --auto-cast=none
neuronxcc-2.13.66.0+6dfecc895/MODULE_06ec99d5ddf55b152eee+2c2d707e/model.hlo_module.pb ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:5cd7d9000790363bc2bc8daf985e53256813dbb2c2d49cd49cc43f7ad66d0f68
3
+ size 63596
neuronxcc-2.13.66.0+6dfecc895/MODULE_06ec99d5ddf55b152eee+2c2d707e/model.neff ADDED
Binary file (390 kB). View file
 
neuronxcc-2.13.66.0+6dfecc895/MODULE_0d2fdaf02cce617fdd0c+2c2d707e/compile_flags.txt ADDED
@@ -0,0 +1 @@
 
 
1
+ --model-type=transformer --auto-cast=none
neuronxcc-2.13.66.0+6dfecc895/MODULE_0d2fdaf02cce617fdd0c+2c2d707e/model.hlo_module.pb ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:a76015cab1cefd2ce7ad45282d63cccfd3ba2e569fbebe0b3e1c4306c8926e28
3
+ size 26575
neuronxcc-2.13.66.0+6dfecc895/MODULE_0d2fdaf02cce617fdd0c+2c2d707e/model.neff ADDED
Binary file (155 kB). View file
 
neuronxcc-2.13.66.0+6dfecc895/MODULE_250ee69b911da3bdc8d2+2c2d707e/compile_flags.txt ADDED
@@ -0,0 +1 @@
 
 
1
+ --model-type=transformer --auto-cast=none
neuronxcc-2.13.66.0+6dfecc895/MODULE_250ee69b911da3bdc8d2+2c2d707e/model.hlo_module.pb ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:8162ba0c165cbe0ab7728ec6be29f58dc0626ae38d2005a53a9d2fd46756add8
3
+ size 414385
neuronxcc-2.13.66.0+6dfecc895/MODULE_250ee69b911da3bdc8d2+2c2d707e/model.neff ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:ccdca52360649cf8b54dcd3b331b2c27e0263f518e05194bd528d5c79fa29e68
3
+ size 4353024
neuronxcc-2.13.66.0+6dfecc895/MODULE_2648f9d12f524b67bd88+2c2d707e/compile_flags.txt ADDED
@@ -0,0 +1 @@
 
 
1
+ --model-type=transformer --auto-cast=none
neuronxcc-2.13.66.0+6dfecc895/MODULE_2648f9d12f524b67bd88+2c2d707e/model.hlo_module.pb ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:db7c07a04fc41c5c7a8b8eaba703c030ffb05f384cbd71dda98c9b6891bd8f5a
3
+ size 43576
neuronxcc-2.13.66.0+6dfecc895/MODULE_2648f9d12f524b67bd88+2c2d707e/model.neff ADDED
Binary file (144 kB). View file
 
neuronxcc-2.13.66.0+6dfecc895/MODULE_29cca2b907e9f3fdb2f8+2c2d707e/model.neff CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:753001cf46d17c61e6a201fce7187d69dc171906233fd1ca901699c87a25e0da
3
  size 2673664
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:0a9cc3e71b2241a2217ea0e978cdb750fa8b0a3a67add3fe6157fc5cd4eb0252
3
  size 2673664
neuronxcc-2.13.66.0+6dfecc895/MODULE_301a41e1f343ba3bec71+2c2d707e/model.neff CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:ccefdbc0fb27cfc2b03553b673f94f8a878fbbda3ac43ab3a6a51dd65ca6005c
3
  size 6401024
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:f050ef0f05acef33da5e88b825e08f595aaf60895d77f37e2658872224d99e62
3
  size 6401024
neuronxcc-2.13.66.0+6dfecc895/MODULE_34d558c33d28a2f62eeb+2c2d707e/compile_flags.txt ADDED
@@ -0,0 +1 @@
 
 
1
+ --model-type=transformer --auto-cast=none
neuronxcc-2.13.66.0+6dfecc895/MODULE_34d558c33d28a2f62eeb+2c2d707e/model.hlo_module.pb ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:bdfe95bf633997292599522cd22afcacace6638893329d36d3587b2a3c73b7ff
3
+ size 47581
neuronxcc-2.13.66.0+6dfecc895/MODULE_34d558c33d28a2f62eeb+2c2d707e/model.neff ADDED
Binary file (144 kB). View file
 
neuronxcc-2.13.66.0+6dfecc895/MODULE_35a5e2c0699bcf01dfe7+2c2d707e/compile_flags.txt ADDED
@@ -0,0 +1 @@
 
 
1
+ --model-type=transformer --auto-cast=none