Synchronizing local compiler cache.
Browse filesThis view is limited to 50 files because it contains too many changes.
See raw diff
- .gitattributes +4 -0
- neuronxcc-2.15.128.0+56dc5a86/0_REGISTRY/0.0.25/inference/llama/meta-llama/Meta-Llama-3-8b-Instruct/3170ca34242cf5513bbc.json +1 -0
- neuronxcc-2.15.128.0+56dc5a86/MODULE_27ff51eb90a211a76ddb+39f12043/compile_flags.json +1 -0
- neuronxcc-2.15.128.0+56dc5a86/MODULE_27ff51eb90a211a76ddb+39f12043/model.done +0 -0
- neuronxcc-2.15.128.0+56dc5a86/MODULE_27ff51eb90a211a76ddb+39f12043/model.hlo_module.pb +3 -0
- neuronxcc-2.15.128.0+56dc5a86/MODULE_27ff51eb90a211a76ddb+39f12043/model.neff +3 -0
- neuronxcc-2.15.128.0+56dc5a86/MODULE_76c6ee0ec53e9e3d108a+39f12043/compile_flags.json +1 -0
- neuronxcc-2.15.128.0+56dc5a86/MODULE_76c6ee0ec53e9e3d108a+39f12043/model.done +0 -0
- neuronxcc-2.15.128.0+56dc5a86/MODULE_76c6ee0ec53e9e3d108a+39f12043/model.hlo_module.pb +3 -0
- neuronxcc-2.15.128.0+56dc5a86/MODULE_76c6ee0ec53e9e3d108a+39f12043/model.neff +3 -0
- neuronxcc-2.15.141.0+d3cfc8ca/0_REGISTRY/0.0.25.dev0/inference/llama/meta-llama/Meta-Llama-3-8b-Instruct/1c477614ffee9812c607.json +1 -0
- neuronxcc-2.15.141.0+d3cfc8ca/0_REGISTRY/0.0.26.dev0/inference/bloom/hf-internal-testing/tiny-random-BloomForCausalLM/303eb4f62b8444fc342f.json +1 -0
- neuronxcc-2.15.141.0+d3cfc8ca/0_REGISTRY/0.0.26.dev0/inference/bloom/hf-internal-testing/tiny-random-BloomForCausalLM/782eb90724ffa6fd54df.json +1 -0
- neuronxcc-2.15.141.0+d3cfc8ca/0_REGISTRY/0.0.26.dev0/inference/bloom/hf-internal-testing/tiny-random-BloomForCausalLM/d414beb329300baec570.json +1 -0
- neuronxcc-2.15.141.0+d3cfc8ca/0_REGISTRY/0.0.26.dev0/inference/gpt2/gpt2/f75be174b21f2bce687f.json +1 -0
- neuronxcc-2.15.141.0+d3cfc8ca/0_REGISTRY/0.0.26.dev0/inference/gpt2/hf-internal-testing/tiny-random-gpt2/937b6b94976a23465cef.json +1 -0
- neuronxcc-2.15.141.0+d3cfc8ca/0_REGISTRY/0.0.26.dev0/inference/gpt2/hf-internal-testing/tiny-random-gpt2/947ea57bcc4424b6e318.json +1 -0
- neuronxcc-2.15.141.0+d3cfc8ca/0_REGISTRY/0.0.26.dev0/inference/gpt2/hf-internal-testing/tiny-random-gpt2/9df09e46d71d66c2c146.json +1 -0
- neuronxcc-2.15.141.0+d3cfc8ca/0_REGISTRY/0.0.26.dev0/inference/llama/llamafactory/tiny-random-Llama-3/7a89fd3a9340237271d6.json +1 -0
- neuronxcc-2.15.141.0+d3cfc8ca/0_REGISTRY/0.0.26.dev0/inference/llama/llamafactory/tiny-random-Llama-3/e3f51e783923d10d0584.json +1 -0
- neuronxcc-2.15.141.0+d3cfc8ca/0_REGISTRY/0.0.26.dev0/inference/llama/llamafactory/tiny-random-Llama-3/ecc12811687e476f6e4e.json +1 -0
- neuronxcc-2.15.141.0+d3cfc8ca/0_REGISTRY/0.0.26.dev0/inference/mistral/dacorvo/tiny-random-MistralForCausalLM/2d8ff9c1b68b56c4928c.json +1 -0
- neuronxcc-2.15.141.0+d3cfc8ca/0_REGISTRY/0.0.26.dev0/inference/mistral/dacorvo/tiny-random-MistralForCausalLM/9026f95ffe6ac8b787b3.json +1 -0
- neuronxcc-2.15.141.0+d3cfc8ca/0_REGISTRY/0.0.26.dev0/inference/mistral/dacorvo/tiny-random-MistralForCausalLM/b00855af4661e6b2790f.json +1 -0
- neuronxcc-2.15.141.0+d3cfc8ca/0_REGISTRY/0.0.26.dev0/inference/mixtral/dacorvo/Mixtral-tiny/25521949fb6b5d3d45d3.json +1 -0
- neuronxcc-2.15.141.0+d3cfc8ca/0_REGISTRY/0.0.26.dev0/inference/mixtral/dacorvo/Mixtral-tiny/4d2eff2be165c67a9f78.json +1 -0
- neuronxcc-2.15.141.0+d3cfc8ca/0_REGISTRY/0.0.26.dev0/inference/mixtral/dacorvo/Mixtral-tiny/79b28715925d5117dc06.json +1 -0
- neuronxcc-2.15.141.0+d3cfc8ca/0_REGISTRY/0.0.26.dev0/inference/opt/hf-internal-testing/tiny-random-OPTForCausalLM/804ab3bb740f659d9fbe.json +1 -0
- neuronxcc-2.15.141.0+d3cfc8ca/0_REGISTRY/0.0.26.dev0/inference/opt/hf-internal-testing/tiny-random-OPTForCausalLM/b39bdecd3bd4dfd67e74.json +1 -0
- neuronxcc-2.15.141.0+d3cfc8ca/0_REGISTRY/0.0.26.dev0/inference/opt/hf-internal-testing/tiny-random-OPTForCausalLM/e3aab8719127da843d98.json +1 -0
- neuronxcc-2.15.141.0+d3cfc8ca/MODULE_054a5de5e3be6e5388cf+39f12043/compile_flags.json +1 -0
- neuronxcc-2.15.141.0+d3cfc8ca/MODULE_054a5de5e3be6e5388cf+39f12043/model.done +0 -0
- neuronxcc-2.15.141.0+d3cfc8ca/MODULE_054a5de5e3be6e5388cf+39f12043/model.hlo_module.pb +3 -0
- neuronxcc-2.15.141.0+d3cfc8ca/MODULE_054a5de5e3be6e5388cf+39f12043/model.neff +0 -0
- neuronxcc-2.15.141.0+d3cfc8ca/MODULE_12981c07402f0005952b+39f12043/compile_flags.json +1 -0
- neuronxcc-2.15.141.0+d3cfc8ca/MODULE_12981c07402f0005952b+39f12043/model.done +0 -0
- neuronxcc-2.15.141.0+d3cfc8ca/MODULE_12981c07402f0005952b+39f12043/model.hlo_module.pb +3 -0
- neuronxcc-2.15.141.0+d3cfc8ca/MODULE_12981c07402f0005952b+39f12043/model.neff +0 -0
- neuronxcc-2.15.141.0+d3cfc8ca/MODULE_18903ade5ea61da2ce44+39f12043/compile_flags.json +1 -0
- neuronxcc-2.15.141.0+d3cfc8ca/MODULE_18903ade5ea61da2ce44+39f12043/model.done +0 -0
- neuronxcc-2.15.141.0+d3cfc8ca/MODULE_18903ade5ea61da2ce44+39f12043/model.hlo_module.pb +3 -0
- neuronxcc-2.15.141.0+d3cfc8ca/MODULE_18903ade5ea61da2ce44+39f12043/model.neff +0 -0
- neuronxcc-2.15.141.0+d3cfc8ca/MODULE_22991d85a6fe7d79cfb5+39f12043/compile_flags.json +1 -0
- neuronxcc-2.15.141.0+d3cfc8ca/MODULE_22991d85a6fe7d79cfb5+39f12043/model.done +0 -0
- neuronxcc-2.15.141.0+d3cfc8ca/MODULE_22991d85a6fe7d79cfb5+39f12043/model.hlo_module.pb +3 -0
- neuronxcc-2.15.141.0+d3cfc8ca/MODULE_22991d85a6fe7d79cfb5+39f12043/model.neff +0 -0
- neuronxcc-2.15.141.0+d3cfc8ca/MODULE_231f2efdd8e57cd0c9ef+39f12043/compile_flags.json +1 -0
- neuronxcc-2.15.141.0+d3cfc8ca/MODULE_231f2efdd8e57cd0c9ef+39f12043/model.done +0 -0
- neuronxcc-2.15.141.0+d3cfc8ca/MODULE_231f2efdd8e57cd0c9ef+39f12043/model.hlo_module.pb +3 -0
- neuronxcc-2.15.141.0+d3cfc8ca/MODULE_231f2efdd8e57cd0c9ef+39f12043/model.neff +0 -0
.gitattributes
CHANGED
@@ -1269,3 +1269,7 @@ neuronxcc-2.15.128.0+56dc5a86/MODULE_b0cf61abda16eb6dab0b+39f12043/model.neff fi
|
|
1269 |
neuronxcc-2.15.128.0+56dc5a86/MODULE_8c93cc9f80cd5bfc3ec7+39f12043/model.neff filter=lfs diff=lfs merge=lfs -text
|
1270 |
neuronxcc-2.15.128.0+56dc5a86/MODULE_204ebebf25942a01d1c2+39f12043/model.neff filter=lfs diff=lfs merge=lfs -text
|
1271 |
neuronxcc-2.15.128.0+56dc5a86/MODULE_2dfe883f70d43532f8a4+39f12043/model.neff filter=lfs diff=lfs merge=lfs -text
|
|
|
|
|
|
|
|
|
|
1269 |
neuronxcc-2.15.128.0+56dc5a86/MODULE_8c93cc9f80cd5bfc3ec7+39f12043/model.neff filter=lfs diff=lfs merge=lfs -text
|
1270 |
neuronxcc-2.15.128.0+56dc5a86/MODULE_204ebebf25942a01d1c2+39f12043/model.neff filter=lfs diff=lfs merge=lfs -text
|
1271 |
neuronxcc-2.15.128.0+56dc5a86/MODULE_2dfe883f70d43532f8a4+39f12043/model.neff filter=lfs diff=lfs merge=lfs -text
|
1272 |
+
neuronxcc-2.15.128.0+56dc5a86/MODULE_27ff51eb90a211a76ddb+39f12043/model.neff filter=lfs diff=lfs merge=lfs -text
|
1273 |
+
neuronxcc-2.15.128.0+56dc5a86/MODULE_76c6ee0ec53e9e3d108a+39f12043/model.neff filter=lfs diff=lfs merge=lfs -text
|
1274 |
+
neuronxcc-2.15.141.0+d3cfc8ca/MODULE_76c6ee0ec53e9e3d108a+39f12043/model.neff filter=lfs diff=lfs merge=lfs -text
|
1275 |
+
neuronxcc-2.15.141.0+d3cfc8ca/MODULE_9b632259b806bfddcba6+39f12043/model.neff filter=lfs diff=lfs merge=lfs -text
|
neuronxcc-2.15.128.0+56dc5a86/0_REGISTRY/0.0.25/inference/llama/meta-llama/Meta-Llama-3-8b-Instruct/3170ca34242cf5513bbc.json
ADDED
@@ -0,0 +1 @@
|
|
|
|
|
1 |
+
{"architectures": ["LlamaForCausalLM"], "attention_bias": false, "attention_dropout": 0.0, "bos_token_id": 128000, "eos_token_id": 128009, "hidden_act": "silu", "hidden_size": 4096, "initializer_range": 0.02, "intermediate_size": 14336, "max_position_embeddings": 8192, "mlp_bias": false, "model_type": "llama", "neuron": {"auto_cast_type": "bf16", "batch_size": 32, "checkpoint_id": "meta-llama/Meta-Llama-3-8b-Instruct", "checkpoint_revision": "5f0b02c75b57c5855da9ae460ce51323ea669d8a", "compiler_type": "neuronx-cc", "compiler_version": "2.15.128.0+56dc5a86", "num_cores": 24, "sequence_length": 8192, "task": "text-generation"}, "num_attention_heads": 32, "num_hidden_layers": 32, "num_key_value_heads": 8, "pretraining_tp": 1, "rms_norm_eps": 1e-05, "rope_scaling": null, "rope_theta": 500000.0, "tie_word_embeddings": false, "torch_dtype": "bfloat16", "use_cache": true, "vocab_size": 128256}
|
neuronxcc-2.15.128.0+56dc5a86/MODULE_27ff51eb90a211a76ddb+39f12043/compile_flags.json
ADDED
@@ -0,0 +1 @@
|
|
|
|
|
1 |
+
["--model-type=transformer", "--auto-cast=none", "--execute-repetition=1"]
|
neuronxcc-2.15.128.0+56dc5a86/MODULE_27ff51eb90a211a76ddb+39f12043/model.done
ADDED
File without changes
|
neuronxcc-2.15.128.0+56dc5a86/MODULE_27ff51eb90a211a76ddb+39f12043/model.hlo_module.pb
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:f0c0a0e6b33bd0d8072f82be430268b68d272d209845f221491f18cb8cb880ff
|
3 |
+
size 450045
|
neuronxcc-2.15.128.0+56dc5a86/MODULE_27ff51eb90a211a76ddb+39f12043/model.neff
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:198c91486b139fd0c5190a3b064aaa708e17d5f376a02fc0eead1da7db4cd9ce
|
3 |
+
size 40438784
|
neuronxcc-2.15.128.0+56dc5a86/MODULE_76c6ee0ec53e9e3d108a+39f12043/compile_flags.json
ADDED
@@ -0,0 +1 @@
|
|
|
|
|
1 |
+
["--model-type=transformer", "--auto-cast=none", "--execute-repetition=1"]
|
neuronxcc-2.15.128.0+56dc5a86/MODULE_76c6ee0ec53e9e3d108a+39f12043/model.done
ADDED
File without changes
|
neuronxcc-2.15.128.0+56dc5a86/MODULE_76c6ee0ec53e9e3d108a+39f12043/model.hlo_module.pb
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:587f52593d0480505013dd9691f5de5a87fe96b24c635dc46cf175194daa4a58
|
3 |
+
size 364579
|
neuronxcc-2.15.128.0+56dc5a86/MODULE_76c6ee0ec53e9e3d108a+39f12043/model.neff
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:02218369116272593d07465d52d1b0f2aa8265620eb4b0ba610e86c9d42c32fb
|
3 |
+
size 7465984
|
neuronxcc-2.15.141.0+d3cfc8ca/0_REGISTRY/0.0.25.dev0/inference/llama/meta-llama/Meta-Llama-3-8b-Instruct/1c477614ffee9812c607.json
ADDED
@@ -0,0 +1 @@
|
|
|
|
|
1 |
+
{"architectures": ["LlamaForCausalLM"], "attention_bias": false, "attention_dropout": 0.0, "bos_token_id": 128000, "eos_token_id": 128009, "hidden_act": "silu", "hidden_size": 4096, "initializer_range": 0.02, "intermediate_size": 14336, "max_position_embeddings": 8192, "mlp_bias": false, "model_type": "llama", "neuron": {"auto_cast_type": "bf16", "batch_size": 32, "checkpoint_id": "meta-llama/Meta-Llama-3-8b-Instruct", "checkpoint_revision": "5f0b02c75b57c5855da9ae460ce51323ea669d8a", "compiler_type": "neuronx-cc", "compiler_version": "2.15.141.0+d3cfc8ca", "num_cores": 24, "sequence_length": 8192, "task": "text-generation"}, "num_attention_heads": 32, "num_hidden_layers": 32, "num_key_value_heads": 8, "pretraining_tp": 1, "rms_norm_eps": 1e-05, "rope_scaling": null, "rope_theta": 500000.0, "tie_word_embeddings": false, "torch_dtype": "bfloat16", "use_cache": true, "vocab_size": 128256}
|
neuronxcc-2.15.141.0+d3cfc8ca/0_REGISTRY/0.0.26.dev0/inference/bloom/hf-internal-testing/tiny-random-BloomForCausalLM/303eb4f62b8444fc342f.json
ADDED
@@ -0,0 +1 @@
|
|
|
|
|
1 |
+
{"apply_residual_connection_post_layernorm": false, "architectures": ["BloomForCausalLM"], "attention_dropout": 0.1, "bos_token_id": 1, "dtype": "float32", "eos_token_id": 2, "gradient_checkpointing": false, "hidden_dropout": 0.1, "hidden_size": 32, "id2label": {"0": "LABEL_0", "1": "LABEL_1", "2": "LABEL_2"}, "initializer_range": 0.02, "is_decoder": true, "label2id": {"LABEL_0": 0, "LABEL_1": 1, "LABEL_2": 2}, "layer_norm_epsilon": 1e-05, "model_type": "bloom", "n_head": 4, "n_layer": 5, "n_positions": 512, "neuron": {"auto_cast_type": "fp16", "batch_size": 2, "checkpoint_id": "hf-internal-testing/tiny-random-BloomForCausalLM", "checkpoint_revision": "92b07e9b7b4f986fa7c54e2ac3b9201b4ba5212e", "compiler_type": "neuronx-cc", "compiler_version": "2.15.141.0+d3cfc8ca", "num_cores": 2, "sequence_length": 100, "task": "text-generation"}, "pad_token_id": 3, "pretraining_tp": 1, "seq_length": 7, "slow_but_exact": true, "torch_dtype": "float32", "type_vocab_size": 16, "use_cache": true, "vocab_size": 1024}
|
neuronxcc-2.15.141.0+d3cfc8ca/0_REGISTRY/0.0.26.dev0/inference/bloom/hf-internal-testing/tiny-random-BloomForCausalLM/782eb90724ffa6fd54df.json
ADDED
@@ -0,0 +1 @@
|
|
|
|
|
1 |
+
{"apply_residual_connection_post_layernorm": false, "architectures": ["BloomForCausalLM"], "attention_dropout": 0.1, "bos_token_id": 1, "dtype": "float32", "eos_token_id": 2, "gradient_checkpointing": false, "hidden_dropout": 0.1, "hidden_size": 32, "id2label": {"0": "LABEL_0", "1": "LABEL_1", "2": "LABEL_2"}, "initializer_range": 0.02, "is_decoder": true, "label2id": {"LABEL_0": 0, "LABEL_1": 1, "LABEL_2": 2}, "layer_norm_epsilon": 1e-05, "model_type": "bloom", "n_head": 4, "n_layer": 5, "n_positions": 512, "neuron": {"auto_cast_type": "fp32", "batch_size": 1, "checkpoint_id": "hf-internal-testing/tiny-random-BloomForCausalLM", "checkpoint_revision": "92b07e9b7b4f986fa7c54e2ac3b9201b4ba5212e", "compiler_type": "neuronx-cc", "compiler_version": "2.15.141.0+d3cfc8ca", "num_cores": 2, "sequence_length": 100, "task": "text-generation"}, "pad_token_id": 3, "pretraining_tp": 1, "seq_length": 7, "slow_but_exact": true, "torch_dtype": "float32", "type_vocab_size": 16, "use_cache": true, "vocab_size": 1024}
|
neuronxcc-2.15.141.0+d3cfc8ca/0_REGISTRY/0.0.26.dev0/inference/bloom/hf-internal-testing/tiny-random-BloomForCausalLM/d414beb329300baec570.json
ADDED
@@ -0,0 +1 @@
|
|
|
|
|
1 |
+
{"apply_residual_connection_post_layernorm": false, "architectures": ["BloomForCausalLM"], "attention_dropout": 0.1, "bos_token_id": 1, "dtype": "float32", "eos_token_id": 2, "gradient_checkpointing": false, "hidden_dropout": 0.1, "hidden_size": 32, "id2label": {"0": "LABEL_0", "1": "LABEL_1", "2": "LABEL_2"}, "initializer_range": 0.02, "is_decoder": true, "label2id": {"LABEL_0": 0, "LABEL_1": 1, "LABEL_2": 2}, "layer_norm_epsilon": 1e-05, "model_type": "bloom", "n_head": 4, "n_layer": 5, "n_positions": 512, "neuron": {"auto_cast_type": "fp16", "batch_size": 1, "checkpoint_id": "hf-internal-testing/tiny-random-BloomForCausalLM", "checkpoint_revision": "92b07e9b7b4f986fa7c54e2ac3b9201b4ba5212e", "compiler_type": "neuronx-cc", "compiler_version": "2.15.141.0+d3cfc8ca", "num_cores": 2, "sequence_length": 100, "task": "text-generation"}, "pad_token_id": 3, "pretraining_tp": 1, "seq_length": 7, "slow_but_exact": true, "torch_dtype": "float32", "type_vocab_size": 16, "use_cache": true, "vocab_size": 1024}
|
neuronxcc-2.15.141.0+d3cfc8ca/0_REGISTRY/0.0.26.dev0/inference/gpt2/gpt2/f75be174b21f2bce687f.json
ADDED
@@ -0,0 +1 @@
|
|
|
|
|
1 |
+
{"activation_function": "gelu_new", "architectures": ["GPT2LMHeadModel"], "attn_pdrop": 0.1, "bos_token_id": 50256, "embd_pdrop": 0.1, "eos_token_id": 50256, "initializer_range": 0.02, "layer_norm_epsilon": 1e-05, "model_type": "gpt2", "n_ctx": 1024, "n_embd": 768, "n_head": 12, "n_inner": null, "n_layer": 12, "n_positions": 1024, "neuron": {"auto_cast_type": "fp16", "batch_size": 4, "checkpoint_id": "gpt2", "checkpoint_revision": "607a30d783dfa663caf39e06633721c8d4cfcd7e", "compiler_type": "neuronx-cc", "compiler_version": "2.15.141.0+d3cfc8ca", "num_cores": 2, "sequence_length": 1024, "task": "text-generation"}, "reorder_and_upcast_attn": false, "resid_pdrop": 0.1, "scale_attn_by_inverse_layer_idx": false, "scale_attn_weights": true, "summary_activation": null, "summary_first_dropout": 0.1, "summary_proj_to_labels": true, "summary_type": "cls_index", "summary_use_proj": true, "task_specific_params": {"text-generation": {"do_sample": true, "max_length": 50}}, "use_cache": true, "vocab_size": 50257}
|
neuronxcc-2.15.141.0+d3cfc8ca/0_REGISTRY/0.0.26.dev0/inference/gpt2/hf-internal-testing/tiny-random-gpt2/937b6b94976a23465cef.json
ADDED
@@ -0,0 +1 @@
|
|
|
|
|
1 |
+
{"activation_function": "gelu_new", "attention_probs_dropout_prob": 0.1, "attn_pdrop": 0.1, "bos_token_id": 98, "embd_pdrop": 0.1, "eos_token_id": 98, "gradient_checkpointing": false, "hidden_act": "gelu", "hidden_dropout_prob": 0.1, "initializer_range": 0.02, "intermediate_size": 37, "layer_norm_epsilon": 1e-05, "model_type": "gpt2", "n_ctx": 512, "n_embd": 32, "n_head": 4, "n_inner": null, "n_layer": 5, "n_positions": 512, "neuron": {"auto_cast_type": "fp16", "batch_size": 2, "checkpoint_id": "hf-internal-testing/tiny-random-gpt2", "checkpoint_revision": "71034c5d8bde858ff824298bdedc65515b97d2b9", "compiler_type": "neuronx-cc", "compiler_version": "2.15.141.0+d3cfc8ca", "num_cores": 2, "sequence_length": 100, "task": "text-generation"}, "pad_token_id": 98, "reorder_and_upcast_attn": false, "resid_pdrop": 0.1, "scale_attn_by_inverse_layer_idx": false, "scale_attn_weights": true, "summary_activation": null, "summary_first_dropout": 0.1, "summary_proj_to_labels": true, "summary_type": "cls_index", "summary_use_proj": true, "type_vocab_size": 16, "use_cache": true, "vocab_size": 1000}
|
neuronxcc-2.15.141.0+d3cfc8ca/0_REGISTRY/0.0.26.dev0/inference/gpt2/hf-internal-testing/tiny-random-gpt2/947ea57bcc4424b6e318.json
ADDED
@@ -0,0 +1 @@
|
|
|
|
|
1 |
+
{"activation_function": "gelu_new", "attention_probs_dropout_prob": 0.1, "attn_pdrop": 0.1, "bos_token_id": 98, "embd_pdrop": 0.1, "eos_token_id": 98, "gradient_checkpointing": false, "hidden_act": "gelu", "hidden_dropout_prob": 0.1, "initializer_range": 0.02, "intermediate_size": 37, "layer_norm_epsilon": 1e-05, "model_type": "gpt2", "n_ctx": 512, "n_embd": 32, "n_head": 4, "n_inner": null, "n_layer": 5, "n_positions": 512, "neuron": {"auto_cast_type": "fp32", "batch_size": 1, "checkpoint_id": "hf-internal-testing/tiny-random-gpt2", "checkpoint_revision": "71034c5d8bde858ff824298bdedc65515b97d2b9", "compiler_type": "neuronx-cc", "compiler_version": "2.15.141.0+d3cfc8ca", "num_cores": 2, "sequence_length": 100, "task": "text-generation"}, "pad_token_id": 98, "reorder_and_upcast_attn": false, "resid_pdrop": 0.1, "scale_attn_by_inverse_layer_idx": false, "scale_attn_weights": true, "summary_activation": null, "summary_first_dropout": 0.1, "summary_proj_to_labels": true, "summary_type": "cls_index", "summary_use_proj": true, "type_vocab_size": 16, "use_cache": true, "vocab_size": 1000}
|
neuronxcc-2.15.141.0+d3cfc8ca/0_REGISTRY/0.0.26.dev0/inference/gpt2/hf-internal-testing/tiny-random-gpt2/9df09e46d71d66c2c146.json
ADDED
@@ -0,0 +1 @@
|
|
|
|
|
1 |
+
{"activation_function": "gelu_new", "attention_probs_dropout_prob": 0.1, "attn_pdrop": 0.1, "bos_token_id": 98, "embd_pdrop": 0.1, "eos_token_id": 98, "gradient_checkpointing": false, "hidden_act": "gelu", "hidden_dropout_prob": 0.1, "initializer_range": 0.02, "intermediate_size": 37, "layer_norm_epsilon": 1e-05, "model_type": "gpt2", "n_ctx": 512, "n_embd": 32, "n_head": 4, "n_inner": null, "n_layer": 5, "n_positions": 512, "neuron": {"auto_cast_type": "fp16", "batch_size": 1, "checkpoint_id": "hf-internal-testing/tiny-random-gpt2", "checkpoint_revision": "71034c5d8bde858ff824298bdedc65515b97d2b9", "compiler_type": "neuronx-cc", "compiler_version": "2.15.141.0+d3cfc8ca", "num_cores": 2, "sequence_length": 100, "task": "text-generation"}, "pad_token_id": 98, "reorder_and_upcast_attn": false, "resid_pdrop": 0.1, "scale_attn_by_inverse_layer_idx": false, "scale_attn_weights": true, "summary_activation": null, "summary_first_dropout": 0.1, "summary_proj_to_labels": true, "summary_type": "cls_index", "summary_use_proj": true, "type_vocab_size": 16, "use_cache": true, "vocab_size": 1000}
|
neuronxcc-2.15.141.0+d3cfc8ca/0_REGISTRY/0.0.26.dev0/inference/llama/llamafactory/tiny-random-Llama-3/7a89fd3a9340237271d6.json
ADDED
@@ -0,0 +1 @@
|
|
|
|
|
1 |
+
{"architectures": ["LlamaForCausalLM"], "attention_bias": false, "attention_dropout": 0.0, "bos_token_id": 128000, "eos_token_id": 128009, "hidden_act": "silu", "hidden_size": 16, "initializer_range": 0.02, "intermediate_size": 64, "max_position_embeddings": 4096, "mlp_bias": false, "model_type": "llama", "neuron": {"auto_cast_type": "fp32", "batch_size": 1, "checkpoint_id": "llamafactory/tiny-random-Llama-3", "checkpoint_revision": "2d9d8c1112e9cd4b0d66bb612e09be7da7997b18", "compiler_type": "neuronx-cc", "compiler_version": "2.15.141.0+d3cfc8ca", "num_cores": 2, "sequence_length": 100, "task": "text-generation"}, "num_attention_heads": 4, "num_hidden_layers": 2, "num_key_value_heads": 4, "pretraining_tp": 1, "rms_norm_eps": 1e-05, "rope_scaling": null, "rope_theta": 10000.0, "tie_word_embeddings": false, "torch_dtype": "float16", "use_cache": true, "vocab_size": 128256}
|
neuronxcc-2.15.141.0+d3cfc8ca/0_REGISTRY/0.0.26.dev0/inference/llama/llamafactory/tiny-random-Llama-3/e3f51e783923d10d0584.json
ADDED
@@ -0,0 +1 @@
|
|
|
|
|
1 |
+
{"architectures": ["LlamaForCausalLM"], "attention_bias": false, "attention_dropout": 0.0, "bos_token_id": 128000, "eos_token_id": 128009, "hidden_act": "silu", "hidden_size": 16, "initializer_range": 0.02, "intermediate_size": 64, "max_position_embeddings": 4096, "mlp_bias": false, "model_type": "llama", "neuron": {"auto_cast_type": "fp16", "batch_size": 2, "checkpoint_id": "llamafactory/tiny-random-Llama-3", "checkpoint_revision": "2d9d8c1112e9cd4b0d66bb612e09be7da7997b18", "compiler_type": "neuronx-cc", "compiler_version": "2.15.141.0+d3cfc8ca", "num_cores": 2, "sequence_length": 100, "task": "text-generation"}, "num_attention_heads": 4, "num_hidden_layers": 2, "num_key_value_heads": 4, "pretraining_tp": 1, "rms_norm_eps": 1e-05, "rope_scaling": null, "rope_theta": 10000.0, "tie_word_embeddings": false, "torch_dtype": "float16", "use_cache": true, "vocab_size": 128256}
|
neuronxcc-2.15.141.0+d3cfc8ca/0_REGISTRY/0.0.26.dev0/inference/llama/llamafactory/tiny-random-Llama-3/ecc12811687e476f6e4e.json
ADDED
@@ -0,0 +1 @@
|
|
|
|
|
1 |
+
{"architectures": ["LlamaForCausalLM"], "attention_bias": false, "attention_dropout": 0.0, "bos_token_id": 128000, "eos_token_id": 128009, "hidden_act": "silu", "hidden_size": 16, "initializer_range": 0.02, "intermediate_size": 64, "max_position_embeddings": 4096, "mlp_bias": false, "model_type": "llama", "neuron": {"auto_cast_type": "fp16", "batch_size": 1, "checkpoint_id": "llamafactory/tiny-random-Llama-3", "checkpoint_revision": "2d9d8c1112e9cd4b0d66bb612e09be7da7997b18", "compiler_type": "neuronx-cc", "compiler_version": "2.15.141.0+d3cfc8ca", "num_cores": 2, "sequence_length": 100, "task": "text-generation"}, "num_attention_heads": 4, "num_hidden_layers": 2, "num_key_value_heads": 4, "pretraining_tp": 1, "rms_norm_eps": 1e-05, "rope_scaling": null, "rope_theta": 10000.0, "tie_word_embeddings": false, "torch_dtype": "float16", "use_cache": true, "vocab_size": 128256}
|
neuronxcc-2.15.141.0+d3cfc8ca/0_REGISTRY/0.0.26.dev0/inference/mistral/dacorvo/tiny-random-MistralForCausalLM/2d8ff9c1b68b56c4928c.json
ADDED
@@ -0,0 +1 @@
|
|
|
|
|
1 |
+
{"architectures": ["MistralForCausalLM"], "attention_dropout": 0.0, "attention_probs_dropout_prob": 0.1, "bos_token_id": 1, "eos_token_id": 2, "head_dim": 8, "hidden_act": "gelu", "hidden_dropout_prob": 0.1, "hidden_size": 32, "initializer_range": 0.02, "intermediate_size": 37, "is_decoder": true, "max_position_embeddings": 512, "model_type": "mistral", "neuron": {"auto_cast_type": "fp16", "batch_size": 2, "checkpoint_id": "dacorvo/tiny-random-MistralForCausalLM", "checkpoint_revision": "81d453e3c8985649e9ee3d4c9378461029d1c73a", "compiler_type": "neuronx-cc", "compiler_version": "2.15.141.0+d3cfc8ca", "num_cores": 2, "sequence_length": 100, "task": "text-generation"}, "num_attention_heads": 4, "num_hidden_layers": 2, "num_key_value_heads": 2, "pad_token_id": 0, "rms_norm_eps": 1e-06, "rope_theta": 10000.0, "sliding_window": 4096, "tie_word_embeddings": false, "torch_dtype": "float32", "type_vocab_size": 16, "use_cache": true, "vocab_size": 32000}
|
neuronxcc-2.15.141.0+d3cfc8ca/0_REGISTRY/0.0.26.dev0/inference/mistral/dacorvo/tiny-random-MistralForCausalLM/9026f95ffe6ac8b787b3.json
ADDED
@@ -0,0 +1 @@
|
|
|
|
|
1 |
+
{"architectures": ["MistralForCausalLM"], "attention_dropout": 0.0, "attention_probs_dropout_prob": 0.1, "bos_token_id": 1, "eos_token_id": 2, "head_dim": 8, "hidden_act": "gelu", "hidden_dropout_prob": 0.1, "hidden_size": 32, "initializer_range": 0.02, "intermediate_size": 37, "is_decoder": true, "max_position_embeddings": 512, "model_type": "mistral", "neuron": {"auto_cast_type": "fp16", "batch_size": 1, "checkpoint_id": "dacorvo/tiny-random-MistralForCausalLM", "checkpoint_revision": "81d453e3c8985649e9ee3d4c9378461029d1c73a", "compiler_type": "neuronx-cc", "compiler_version": "2.15.141.0+d3cfc8ca", "num_cores": 2, "sequence_length": 100, "task": "text-generation"}, "num_attention_heads": 4, "num_hidden_layers": 2, "num_key_value_heads": 2, "pad_token_id": 0, "rms_norm_eps": 1e-06, "rope_theta": 10000.0, "sliding_window": 4096, "tie_word_embeddings": false, "torch_dtype": "float32", "type_vocab_size": 16, "use_cache": true, "vocab_size": 32000}
|
neuronxcc-2.15.141.0+d3cfc8ca/0_REGISTRY/0.0.26.dev0/inference/mistral/dacorvo/tiny-random-MistralForCausalLM/b00855af4661e6b2790f.json
ADDED
@@ -0,0 +1 @@
|
|
|
|
|
1 |
+
{"architectures": ["MistralForCausalLM"], "attention_dropout": 0.0, "attention_probs_dropout_prob": 0.1, "bos_token_id": 1, "eos_token_id": 2, "head_dim": 8, "hidden_act": "gelu", "hidden_dropout_prob": 0.1, "hidden_size": 32, "initializer_range": 0.02, "intermediate_size": 37, "is_decoder": true, "max_position_embeddings": 512, "model_type": "mistral", "neuron": {"auto_cast_type": "fp32", "batch_size": 1, "checkpoint_id": "dacorvo/tiny-random-MistralForCausalLM", "checkpoint_revision": "81d453e3c8985649e9ee3d4c9378461029d1c73a", "compiler_type": "neuronx-cc", "compiler_version": "2.15.141.0+d3cfc8ca", "num_cores": 2, "sequence_length": 100, "task": "text-generation"}, "num_attention_heads": 4, "num_hidden_layers": 2, "num_key_value_heads": 2, "pad_token_id": 0, "rms_norm_eps": 1e-06, "rope_theta": 10000.0, "sliding_window": 4096, "tie_word_embeddings": false, "torch_dtype": "float32", "type_vocab_size": 16, "use_cache": true, "vocab_size": 32000}
|
neuronxcc-2.15.141.0+d3cfc8ca/0_REGISTRY/0.0.26.dev0/inference/mixtral/dacorvo/Mixtral-tiny/25521949fb6b5d3d45d3.json
ADDED
@@ -0,0 +1 @@
|
|
|
|
|
1 |
+
{"architectures": ["MixtralForCausalLM"], "attention_dropout": 0.0, "bos_token_id": 1, "eos_token_id": 2, "hidden_act": "silu", "hidden_size": 1024, "initializer_range": 0.02, "intermediate_size": 3584, "max_position_embeddings": 1024, "model_type": "mixtral", "neuron": {"auto_cast_type": "fp32", "batch_size": 1, "checkpoint_id": "dacorvo/Mixtral-tiny", "checkpoint_revision": "c557ba205ddff6ea911f4719e0d543d6c08356b6", "compiler_type": "neuronx-cc", "compiler_version": "2.15.141.0+d3cfc8ca", "num_cores": 2, "sequence_length": 100, "task": "text-generation"}, "num_attention_heads": 32, "num_experts_per_tok": 2, "num_hidden_layers": 2, "num_key_value_heads": 8, "num_local_experts": 8, "output_router_logits": false, "rms_norm_eps": 1e-05, "rope_theta": 10000.0, "router_aux_loss_coef": 0.001, "router_jitter_noise": 0.0, "sliding_window": 4096, "tie_word_embeddings": false, "torch_dtype": "float16", "use_cache": true, "vocab_size": 32000}
|
neuronxcc-2.15.141.0+d3cfc8ca/0_REGISTRY/0.0.26.dev0/inference/mixtral/dacorvo/Mixtral-tiny/4d2eff2be165c67a9f78.json
ADDED
@@ -0,0 +1 @@
|
|
|
|
|
1 |
+
{"architectures": ["MixtralForCausalLM"], "attention_dropout": 0.0, "bos_token_id": 1, "eos_token_id": 2, "hidden_act": "silu", "hidden_size": 1024, "initializer_range": 0.02, "intermediate_size": 3584, "max_position_embeddings": 1024, "model_type": "mixtral", "neuron": {"auto_cast_type": "fp16", "batch_size": 2, "checkpoint_id": "dacorvo/Mixtral-tiny", "checkpoint_revision": "c557ba205ddff6ea911f4719e0d543d6c08356b6", "compiler_type": "neuronx-cc", "compiler_version": "2.15.141.0+d3cfc8ca", "num_cores": 2, "sequence_length": 100, "task": "text-generation"}, "num_attention_heads": 32, "num_experts_per_tok": 2, "num_hidden_layers": 2, "num_key_value_heads": 8, "num_local_experts": 8, "output_router_logits": false, "rms_norm_eps": 1e-05, "rope_theta": 10000.0, "router_aux_loss_coef": 0.001, "router_jitter_noise": 0.0, "sliding_window": 4096, "tie_word_embeddings": false, "torch_dtype": "float16", "use_cache": true, "vocab_size": 32000}
|
neuronxcc-2.15.141.0+d3cfc8ca/0_REGISTRY/0.0.26.dev0/inference/mixtral/dacorvo/Mixtral-tiny/79b28715925d5117dc06.json
ADDED
@@ -0,0 +1 @@
|
|
|
|
|
1 |
+
{"architectures": ["MixtralForCausalLM"], "attention_dropout": 0.0, "bos_token_id": 1, "eos_token_id": 2, "hidden_act": "silu", "hidden_size": 1024, "initializer_range": 0.02, "intermediate_size": 3584, "max_position_embeddings": 1024, "model_type": "mixtral", "neuron": {"auto_cast_type": "fp16", "batch_size": 1, "checkpoint_id": "dacorvo/Mixtral-tiny", "checkpoint_revision": "c557ba205ddff6ea911f4719e0d543d6c08356b6", "compiler_type": "neuronx-cc", "compiler_version": "2.15.141.0+d3cfc8ca", "num_cores": 2, "sequence_length": 100, "task": "text-generation"}, "num_attention_heads": 32, "num_experts_per_tok": 2, "num_hidden_layers": 2, "num_key_value_heads": 8, "num_local_experts": 8, "output_router_logits": false, "rms_norm_eps": 1e-05, "rope_theta": 10000.0, "router_aux_loss_coef": 0.001, "router_jitter_noise": 0.0, "sliding_window": 4096, "tie_word_embeddings": false, "torch_dtype": "float16", "use_cache": true, "vocab_size": 32000}
|
neuronxcc-2.15.141.0+d3cfc8ca/0_REGISTRY/0.0.26.dev0/inference/opt/hf-internal-testing/tiny-random-OPTForCausalLM/804ab3bb740f659d9fbe.json
ADDED
@@ -0,0 +1 @@
|
|
|
|
|
1 |
+
{"_remove_final_layer_norm": false, "activation_function": "relu", "architectures": ["OPTForCausalLM"], "attention_dropout": 0.1, "bos_token_id": 2, "do_layer_norm_before": true, "dropout": 0.1, "embed_dim": 16, "enable_bias": true, "eos_token_id": 2, "ffn_dim": 4, "hidden_size": 16, "init_std": 0.02, "is_decoder": true, "layer_norm_elementwise_affine": true, "layerdrop": 0.0, "max_position_embeddings": 100, "model_type": "opt", "neuron": {"auto_cast_type": "fp16", "batch_size": 2, "checkpoint_id": "hf-internal-testing/tiny-random-OPTForCausalLM", "checkpoint_revision": "190d1f4fc0011d2eaeaa05282e0fbd2445e4b11f", "compiler_type": "neuronx-cc", "compiler_version": "2.15.141.0+d3cfc8ca", "num_cores": 2, "sequence_length": 100, "task": "text-generation"}, "num_attention_heads": 4, "num_hidden_layers": 5, "pad_token_id": 1, "torch_dtype": "float32", "use_cache": true, "vocab_size": 50265, "word_embed_proj_dim": 16}
|
neuronxcc-2.15.141.0+d3cfc8ca/0_REGISTRY/0.0.26.dev0/inference/opt/hf-internal-testing/tiny-random-OPTForCausalLM/b39bdecd3bd4dfd67e74.json
ADDED
@@ -0,0 +1 @@
|
|
|
|
|
1 |
+
{"_remove_final_layer_norm": false, "activation_function": "relu", "architectures": ["OPTForCausalLM"], "attention_dropout": 0.1, "bos_token_id": 2, "do_layer_norm_before": true, "dropout": 0.1, "embed_dim": 16, "enable_bias": true, "eos_token_id": 2, "ffn_dim": 4, "hidden_size": 16, "init_std": 0.02, "is_decoder": true, "layer_norm_elementwise_affine": true, "layerdrop": 0.0, "max_position_embeddings": 100, "model_type": "opt", "neuron": {"auto_cast_type": "fp16", "batch_size": 1, "checkpoint_id": "hf-internal-testing/tiny-random-OPTForCausalLM", "checkpoint_revision": "190d1f4fc0011d2eaeaa05282e0fbd2445e4b11f", "compiler_type": "neuronx-cc", "compiler_version": "2.15.141.0+d3cfc8ca", "num_cores": 2, "sequence_length": 100, "task": "text-generation"}, "num_attention_heads": 4, "num_hidden_layers": 5, "pad_token_id": 1, "torch_dtype": "float32", "use_cache": true, "vocab_size": 50265, "word_embed_proj_dim": 16}
|
neuronxcc-2.15.141.0+d3cfc8ca/0_REGISTRY/0.0.26.dev0/inference/opt/hf-internal-testing/tiny-random-OPTForCausalLM/e3aab8719127da843d98.json
ADDED
@@ -0,0 +1 @@
|
|
|
|
|
1 |
+
{"_remove_final_layer_norm": false, "activation_function": "relu", "architectures": ["OPTForCausalLM"], "attention_dropout": 0.1, "bos_token_id": 2, "do_layer_norm_before": true, "dropout": 0.1, "embed_dim": 16, "enable_bias": true, "eos_token_id": 2, "ffn_dim": 4, "hidden_size": 16, "init_std": 0.02, "is_decoder": true, "layer_norm_elementwise_affine": true, "layerdrop": 0.0, "max_position_embeddings": 100, "model_type": "opt", "neuron": {"auto_cast_type": "fp32", "batch_size": 1, "checkpoint_id": "hf-internal-testing/tiny-random-OPTForCausalLM", "checkpoint_revision": "190d1f4fc0011d2eaeaa05282e0fbd2445e4b11f", "compiler_type": "neuronx-cc", "compiler_version": "2.15.141.0+d3cfc8ca", "num_cores": 2, "sequence_length": 100, "task": "text-generation"}, "num_attention_heads": 4, "num_hidden_layers": 5, "pad_token_id": 1, "torch_dtype": "float32", "use_cache": true, "vocab_size": 50265, "word_embed_proj_dim": 16}
|
neuronxcc-2.15.141.0+d3cfc8ca/MODULE_054a5de5e3be6e5388cf+39f12043/compile_flags.json
ADDED
@@ -0,0 +1 @@
|
|
|
|
|
1 |
+
["--model-type=transformer", "--auto-cast=none", "--execute-repetition=1"]
|
neuronxcc-2.15.141.0+d3cfc8ca/MODULE_054a5de5e3be6e5388cf+39f12043/model.done
ADDED
File without changes
|
neuronxcc-2.15.141.0+d3cfc8ca/MODULE_054a5de5e3be6e5388cf+39f12043/model.hlo_module.pb
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:6e95db68e81f671b1971bee14d85270114c1352b0d8b90be228518024b5081cb
|
3 |
+
size 52680
|
neuronxcc-2.15.141.0+d3cfc8ca/MODULE_054a5de5e3be6e5388cf+39f12043/model.neff
ADDED
Binary file (93.2 kB). View file
|
|
neuronxcc-2.15.141.0+d3cfc8ca/MODULE_12981c07402f0005952b+39f12043/compile_flags.json
ADDED
@@ -0,0 +1 @@
|
|
|
|
|
1 |
+
["--model-type=transformer", "--auto-cast=none", "--execute-repetition=1"]
|
neuronxcc-2.15.141.0+d3cfc8ca/MODULE_12981c07402f0005952b+39f12043/model.done
ADDED
File without changes
|
neuronxcc-2.15.141.0+d3cfc8ca/MODULE_12981c07402f0005952b+39f12043/model.hlo_module.pb
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:2c59cc0255ccb1ad43da9d8d9df741ddc5d71da55db74af8cd38030b625361ce
|
3 |
+
size 22544
|
neuronxcc-2.15.141.0+d3cfc8ca/MODULE_12981c07402f0005952b+39f12043/model.neff
ADDED
Binary file (144 kB). View file
|
|
neuronxcc-2.15.141.0+d3cfc8ca/MODULE_18903ade5ea61da2ce44+39f12043/compile_flags.json
ADDED
@@ -0,0 +1 @@
|
|
|
|
|
1 |
+
["--model-type=transformer", "--auto-cast=none", "--execute-repetition=1"]
|
neuronxcc-2.15.141.0+d3cfc8ca/MODULE_18903ade5ea61da2ce44+39f12043/model.done
ADDED
File without changes
|
neuronxcc-2.15.141.0+d3cfc8ca/MODULE_18903ade5ea61da2ce44+39f12043/model.hlo_module.pb
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:d90ea63f393c8b88475d8852b1033305e6461e0441610873fec4dedd070fd948
|
3 |
+
size 52301
|
neuronxcc-2.15.141.0+d3cfc8ca/MODULE_18903ade5ea61da2ce44+39f12043/model.neff
ADDED
Binary file (124 kB). View file
|
|
neuronxcc-2.15.141.0+d3cfc8ca/MODULE_22991d85a6fe7d79cfb5+39f12043/compile_flags.json
ADDED
@@ -0,0 +1 @@
|
|
|
|
|
1 |
+
["--model-type=transformer", "--auto-cast=none", "--execute-repetition=1"]
|
neuronxcc-2.15.141.0+d3cfc8ca/MODULE_22991d85a6fe7d79cfb5+39f12043/model.done
ADDED
File without changes
|
neuronxcc-2.15.141.0+d3cfc8ca/MODULE_22991d85a6fe7d79cfb5+39f12043/model.hlo_module.pb
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:4fd85291740379776513f453dbcace08a8fc212945914e9aacbfe9695a6e396f
|
3 |
+
size 130974
|
neuronxcc-2.15.141.0+d3cfc8ca/MODULE_22991d85a6fe7d79cfb5+39f12043/model.neff
ADDED
Binary file (503 kB). View file
|
|
neuronxcc-2.15.141.0+d3cfc8ca/MODULE_231f2efdd8e57cd0c9ef+39f12043/compile_flags.json
ADDED
@@ -0,0 +1 @@
|
|
|
|
|
1 |
+
["--model-type=transformer", "--auto-cast=none", "--execute-repetition=1"]
|
neuronxcc-2.15.141.0+d3cfc8ca/MODULE_231f2efdd8e57cd0c9ef+39f12043/model.done
ADDED
File without changes
|
neuronxcc-2.15.141.0+d3cfc8ca/MODULE_231f2efdd8e57cd0c9ef+39f12043/model.hlo_module.pb
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:ae568cf203687d0bf800521fe6d0c581b434460ea802f3567cd4ca965654ca74
|
3 |
+
size 45613
|
neuronxcc-2.15.141.0+d3cfc8ca/MODULE_231f2efdd8e57cd0c9ef+39f12043/model.neff
ADDED
Binary file (421 kB). View file
|
|