Synchronizing local compiler cache.
Browse filesThis view is limited to 50 files because it contains too many changes.
See raw diff
- .gitattributes +14 -0
- neuronxcc-2.15.128.0+56dc5a86/0_REGISTRY/0.0.25.dev0/inference/bloom/hf-internal-testing/tiny-random-BloomForCausalLM/12469cfbaa2be7979be9.json +1 -0
- neuronxcc-2.15.128.0+56dc5a86/0_REGISTRY/0.0.25.dev0/inference/bloom/hf-internal-testing/tiny-random-BloomForCausalLM/3047f14e5cd104f6da88.json +1 -0
- neuronxcc-2.15.128.0+56dc5a86/0_REGISTRY/0.0.25.dev0/inference/bloom/hf-internal-testing/tiny-random-BloomForCausalLM/cbf5e07921ee14bd6483.json +1 -0
- neuronxcc-2.15.128.0+56dc5a86/0_REGISTRY/0.0.25.dev0/inference/gpt2/gpt2/36aa91009c1c430c8dcf.json +1 -0
- neuronxcc-2.15.128.0+56dc5a86/0_REGISTRY/0.0.25.dev0/inference/gpt2/gpt2/98f807ee88912bf4b9d8.json +1 -0
- neuronxcc-2.15.128.0+56dc5a86/0_REGISTRY/0.0.25.dev0/inference/gpt2/hf-internal-testing/tiny-random-gpt2/0b5b5348467d20a868da.json +1 -0
- neuronxcc-2.15.128.0+56dc5a86/0_REGISTRY/0.0.25.dev0/inference/gpt2/hf-internal-testing/tiny-random-gpt2/3a30f319d7cae1f3076c.json +1 -0
- neuronxcc-2.15.128.0+56dc5a86/0_REGISTRY/0.0.25.dev0/inference/gpt2/hf-internal-testing/tiny-random-gpt2/6b3192766c61d426df54.json +1 -0
- neuronxcc-2.15.128.0+56dc5a86/0_REGISTRY/0.0.25.dev0/inference/gpt2/hf-internal-testing/tiny-random-gpt2/74bc6efd0eb46f234bef.json +1 -0
- neuronxcc-2.15.128.0+56dc5a86/0_REGISTRY/0.0.25.dev0/inference/gpt2/hf-internal-testing/tiny-random-gpt2/8a6582d10bf46ea7004e.json +1 -0
- neuronxcc-2.15.128.0+56dc5a86/0_REGISTRY/0.0.25.dev0/inference/llama/NousResearch/Hermes-2-Theta-Llama-3-8B/eda26b3891458ca144a4.json +1 -0
- neuronxcc-2.15.128.0+56dc5a86/0_REGISTRY/0.0.25.dev0/inference/llama/dacorvo/tiny-random-llama/08720e1cfeb1befa20c2.json +1 -0
- neuronxcc-2.15.128.0+56dc5a86/0_REGISTRY/0.0.25.dev0/inference/llama/dacorvo/tiny-random-llama/5ba8a7533020030cff1d.json +1 -0
- neuronxcc-2.15.128.0+56dc5a86/0_REGISTRY/0.0.25.dev0/inference/llama/dacorvo/tiny-random-llama/e7337f4af7481c9d827e.json +1 -0
- neuronxcc-2.15.128.0+56dc5a86/0_REGISTRY/0.0.25.dev0/inference/mistral/dacorvo/tiny-random-MistralForCausalLM/3bdcb3d6d5fb4b0b7097.json +1 -0
- neuronxcc-2.15.128.0+56dc5a86/0_REGISTRY/0.0.25.dev0/inference/mistral/dacorvo/tiny-random-MistralForCausalLM/3d676d12f1566545d6b0.json +1 -0
- neuronxcc-2.15.128.0+56dc5a86/0_REGISTRY/0.0.25.dev0/inference/mistral/dacorvo/tiny-random-MistralForCausalLM/ad4d8b389573be75fbee.json +1 -0
- neuronxcc-2.15.128.0+56dc5a86/0_REGISTRY/0.0.25.dev0/inference/mistral/optimum/mistral-1.1b-testing/081b187113a5f417a9e0.json +1 -0
- neuronxcc-2.15.128.0+56dc5a86/0_REGISTRY/0.0.25.dev0/inference/mixtral/dacorvo/Mixtral-tiny/3a688b7dcd45d9a80b14.json +1 -0
- neuronxcc-2.15.128.0+56dc5a86/0_REGISTRY/0.0.25.dev0/inference/mixtral/dacorvo/Mixtral-tiny/89980755d7c8bc1b31b0.json +1 -0
- neuronxcc-2.15.128.0+56dc5a86/0_REGISTRY/0.0.25.dev0/inference/mixtral/dacorvo/Mixtral-tiny/9f85ef35436cc6a20682.json +1 -0
- neuronxcc-2.15.128.0+56dc5a86/0_REGISTRY/0.0.25.dev0/inference/mixtral/dacorvo/Mixtral-tiny/f1477b1f14a6669df1a7.json +1 -0
- neuronxcc-2.15.128.0+56dc5a86/0_REGISTRY/0.0.25.dev0/inference/opt/hf-internal-testing/tiny-random-OPTForCausalLM/63ef91a527f73ba3b87e.json +1 -0
- neuronxcc-2.15.128.0+56dc5a86/0_REGISTRY/0.0.25.dev0/inference/opt/hf-internal-testing/tiny-random-OPTForCausalLM/aa9fef08c5f79c9217fb.json +1 -0
- neuronxcc-2.15.128.0+56dc5a86/0_REGISTRY/0.0.25.dev0/inference/opt/hf-internal-testing/tiny-random-OPTForCausalLM/fe9dc58f7d028523a0a6.json +1 -0
- neuronxcc-2.15.128.0+56dc5a86/0_REGISTRY/0.0.25.dev0/inference/t5/hf-internal-testing/tiny-random-t5/af6cce39edb0ba1725d3.json +1 -0
- neuronxcc-2.15.128.0+56dc5a86/MODULE_00a895c500f614d5e448+39f12043/compile_flags.json +1 -0
- neuronxcc-2.15.128.0+56dc5a86/MODULE_00a895c500f614d5e448+39f12043/model.done +0 -0
- neuronxcc-2.15.128.0+56dc5a86/MODULE_00a895c500f614d5e448+39f12043/model.hlo_module.pb +3 -0
- neuronxcc-2.15.128.0+56dc5a86/MODULE_00a895c500f614d5e448+39f12043/model.neff +0 -0
- neuronxcc-2.15.128.0+56dc5a86/MODULE_0270f34f54c498cedeec+39f12043/compile_flags.json +1 -0
- neuronxcc-2.15.128.0+56dc5a86/MODULE_0270f34f54c498cedeec+39f12043/model.done +0 -0
- neuronxcc-2.15.128.0+56dc5a86/MODULE_0270f34f54c498cedeec+39f12043/model.hlo_module.pb +3 -0
- neuronxcc-2.15.128.0+56dc5a86/MODULE_0270f34f54c498cedeec+39f12043/model.neff +0 -0
- neuronxcc-2.15.128.0+56dc5a86/MODULE_05d01a658506892e6bcf+39f12043/compile_flags.json +1 -0
- neuronxcc-2.15.128.0+56dc5a86/MODULE_05d01a658506892e6bcf+39f12043/model.done +0 -0
- neuronxcc-2.15.128.0+56dc5a86/MODULE_05d01a658506892e6bcf+39f12043/model.hlo_module.pb +3 -0
- neuronxcc-2.15.128.0+56dc5a86/MODULE_05d01a658506892e6bcf+39f12043/model.neff +0 -0
- neuronxcc-2.15.128.0+56dc5a86/MODULE_0d7bde45dbf2d5c06852+39f12043/compile_flags.json +1 -0
- neuronxcc-2.15.128.0+56dc5a86/MODULE_0d7bde45dbf2d5c06852+39f12043/model.done +0 -0
- neuronxcc-2.15.128.0+56dc5a86/MODULE_0d7bde45dbf2d5c06852+39f12043/model.hlo_module.pb +3 -0
- neuronxcc-2.15.128.0+56dc5a86/MODULE_0d7bde45dbf2d5c06852+39f12043/model.neff +0 -0
- neuronxcc-2.15.128.0+56dc5a86/MODULE_0ea3601cf80a3b68f5bb/feature_extractor/preprocessor_config.json +27 -0
- neuronxcc-2.15.128.0+56dc5a86/MODULE_0ea3601cf80a3b68f5bb/model_index.json +38 -0
- neuronxcc-2.15.128.0+56dc5a86/MODULE_0ea3601cf80a3b68f5bb/scheduler/scheduler_config.json +15 -0
- neuronxcc-2.15.128.0+56dc5a86/MODULE_0ea3601cf80a3b68f5bb/text_encoder/config.json +46 -0
- neuronxcc-2.15.128.0+56dc5a86/MODULE_0ea3601cf80a3b68f5bb/text_encoder/model.neuron +0 -0
- neuronxcc-2.15.128.0+56dc5a86/MODULE_0ea3601cf80a3b68f5bb/tokenizer/merges.txt +647 -0
- neuronxcc-2.15.128.0+56dc5a86/MODULE_0ea3601cf80a3b68f5bb/tokenizer/special_tokens_map.json +24 -0
.gitattributes
CHANGED
@@ -3874,3 +3874,17 @@ neuronxcc-2.14.227.0+2d4f85be/MODULE_3b2bf60b825c2327d04d/text_encoder_2/model.n
|
|
3874 |
neuronxcc-2.14.227.0+2d4f85be/MODULE_3b2bf60b825c2327d04d/unet/model.neuron filter=lfs diff=lfs merge=lfs -text
|
3875 |
neuronxcc-2.14.227.0+2d4f85be/MODULE_3b2bf60b825c2327d04d/vae_decoder/model.neuron filter=lfs diff=lfs merge=lfs -text
|
3876 |
neuronxcc-2.14.227.0+2d4f85be/MODULE_3b2bf60b825c2327d04d/vae_encoder/model.neuron filter=lfs diff=lfs merge=lfs -text
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
3874 |
neuronxcc-2.14.227.0+2d4f85be/MODULE_3b2bf60b825c2327d04d/unet/model.neuron filter=lfs diff=lfs merge=lfs -text
|
3875 |
neuronxcc-2.14.227.0+2d4f85be/MODULE_3b2bf60b825c2327d04d/vae_decoder/model.neuron filter=lfs diff=lfs merge=lfs -text
|
3876 |
neuronxcc-2.14.227.0+2d4f85be/MODULE_3b2bf60b825c2327d04d/vae_encoder/model.neuron filter=lfs diff=lfs merge=lfs -text
|
3877 |
+
neuronxcc-2.15.128.0+56dc5a86/MODULE_0ea3601cf80a3b68f5bb/unet/model.neuron filter=lfs diff=lfs merge=lfs -text
|
3878 |
+
neuronxcc-2.15.128.0+56dc5a86/MODULE_0ea3601cf80a3b68f5bb/vae_decoder/model.neuron filter=lfs diff=lfs merge=lfs -text
|
3879 |
+
neuronxcc-2.15.128.0+56dc5a86/MODULE_0ea3601cf80a3b68f5bb/vae_encoder/model.neuron filter=lfs diff=lfs merge=lfs -text
|
3880 |
+
neuronxcc-2.15.128.0+56dc5a86/MODULE_1817e58bba6976294b3a/unet/model.neuron filter=lfs diff=lfs merge=lfs -text
|
3881 |
+
neuronxcc-2.15.128.0+56dc5a86/MODULE_1817e58bba6976294b3a/vae_decoder/model.neuron filter=lfs diff=lfs merge=lfs -text
|
3882 |
+
neuronxcc-2.15.128.0+56dc5a86/MODULE_1817e58bba6976294b3a/vae_encoder/model.neuron filter=lfs diff=lfs merge=lfs -text
|
3883 |
+
neuronxcc-2.15.128.0+56dc5a86/MODULE_3a7777a73798af802ac1+39f12043/model.neff filter=lfs diff=lfs merge=lfs -text
|
3884 |
+
neuronxcc-2.15.128.0+56dc5a86/MODULE_3da92ce16fff3bc522f8+39f12043/model.neff filter=lfs diff=lfs merge=lfs -text
|
3885 |
+
neuronxcc-2.15.128.0+56dc5a86/MODULE_3e0ab9335d16a5ec49ef+39f12043/model.neff filter=lfs diff=lfs merge=lfs -text
|
3886 |
+
neuronxcc-2.15.128.0+56dc5a86/MODULE_8a0f4078fcafddeba45a/unet/model.neuron filter=lfs diff=lfs merge=lfs -text
|
3887 |
+
neuronxcc-2.15.128.0+56dc5a86/MODULE_8a0f4078fcafddeba45a/vae_decoder/model.neuron filter=lfs diff=lfs merge=lfs -text
|
3888 |
+
neuronxcc-2.15.128.0+56dc5a86/MODULE_8a0f4078fcafddeba45a/vae_encoder/model.neuron filter=lfs diff=lfs merge=lfs -text
|
3889 |
+
neuronxcc-2.15.128.0+56dc5a86/MODULE_d1618c04ba6ca3abe828+39f12043/model.neff filter=lfs diff=lfs merge=lfs -text
|
3890 |
+
neuronxcc-2.15.128.0+56dc5a86/MODULE_d36186ef4c022d90fe8c+39f12043/model.neff filter=lfs diff=lfs merge=lfs -text
|
neuronxcc-2.15.128.0+56dc5a86/0_REGISTRY/0.0.25.dev0/inference/bloom/hf-internal-testing/tiny-random-BloomForCausalLM/12469cfbaa2be7979be9.json
ADDED
@@ -0,0 +1 @@
|
|
|
|
|
1 |
+
{"apply_residual_connection_post_layernorm": false, "architectures": ["BloomForCausalLM"], "attention_dropout": 0.1, "bos_token_id": 1, "dtype": "float32", "eos_token_id": 2, "gradient_checkpointing": false, "hidden_dropout": 0.1, "hidden_size": 32, "id2label": {"0": "LABEL_0", "1": "LABEL_1", "2": "LABEL_2"}, "initializer_range": 0.02, "is_decoder": true, "label2id": {"LABEL_0": 0, "LABEL_1": 1, "LABEL_2": 2}, "layer_norm_epsilon": 1e-05, "model_type": "bloom", "n_head": 4, "n_layer": 5, "n_positions": 512, "neuron": {"auto_cast_type": "fp16", "batch_size": 1, "checkpoint_id": "hf-internal-testing/tiny-random-BloomForCausalLM", "checkpoint_revision": "92b07e9b7b4f986fa7c54e2ac3b9201b4ba5212e", "compiler_type": "neuronx-cc", "compiler_version": "2.15.128.0+56dc5a86", "num_cores": 2, "sequence_length": 100, "task": "text-generation"}, "pad_token_id": 3, "pretraining_tp": 1, "seq_length": 7, "slow_but_exact": true, "torch_dtype": "float32", "type_vocab_size": 16, "use_cache": true, "vocab_size": 1024}
|
neuronxcc-2.15.128.0+56dc5a86/0_REGISTRY/0.0.25.dev0/inference/bloom/hf-internal-testing/tiny-random-BloomForCausalLM/3047f14e5cd104f6da88.json
ADDED
@@ -0,0 +1 @@
|
|
|
|
|
1 |
+
{"apply_residual_connection_post_layernorm": false, "architectures": ["BloomForCausalLM"], "attention_dropout": 0.1, "bos_token_id": 1, "dtype": "float32", "eos_token_id": 2, "gradient_checkpointing": false, "hidden_dropout": 0.1, "hidden_size": 32, "id2label": {"0": "LABEL_0", "1": "LABEL_1", "2": "LABEL_2"}, "initializer_range": 0.02, "is_decoder": true, "label2id": {"LABEL_0": 0, "LABEL_1": 1, "LABEL_2": 2}, "layer_norm_epsilon": 1e-05, "model_type": "bloom", "n_head": 4, "n_layer": 5, "n_positions": 512, "neuron": {"auto_cast_type": "fp32", "batch_size": 1, "checkpoint_id": "hf-internal-testing/tiny-random-BloomForCausalLM", "checkpoint_revision": "92b07e9b7b4f986fa7c54e2ac3b9201b4ba5212e", "compiler_type": "neuronx-cc", "compiler_version": "2.15.128.0+56dc5a86", "num_cores": 2, "sequence_length": 100, "task": "text-generation"}, "pad_token_id": 3, "pretraining_tp": 1, "seq_length": 7, "slow_but_exact": true, "torch_dtype": "float32", "type_vocab_size": 16, "use_cache": true, "vocab_size": 1024}
|
neuronxcc-2.15.128.0+56dc5a86/0_REGISTRY/0.0.25.dev0/inference/bloom/hf-internal-testing/tiny-random-BloomForCausalLM/cbf5e07921ee14bd6483.json
ADDED
@@ -0,0 +1 @@
|
|
|
|
|
1 |
+
{"apply_residual_connection_post_layernorm": false, "architectures": ["BloomForCausalLM"], "attention_dropout": 0.1, "bos_token_id": 1, "dtype": "float32", "eos_token_id": 2, "gradient_checkpointing": false, "hidden_dropout": 0.1, "hidden_size": 32, "id2label": {"0": "LABEL_0", "1": "LABEL_1", "2": "LABEL_2"}, "initializer_range": 0.02, "is_decoder": true, "label2id": {"LABEL_0": 0, "LABEL_1": 1, "LABEL_2": 2}, "layer_norm_epsilon": 1e-05, "model_type": "bloom", "n_head": 4, "n_layer": 5, "n_positions": 512, "neuron": {"auto_cast_type": "fp16", "batch_size": 2, "checkpoint_id": "hf-internal-testing/tiny-random-BloomForCausalLM", "checkpoint_revision": "92b07e9b7b4f986fa7c54e2ac3b9201b4ba5212e", "compiler_type": "neuronx-cc", "compiler_version": "2.15.128.0+56dc5a86", "num_cores": 2, "sequence_length": 100, "task": "text-generation"}, "pad_token_id": 3, "pretraining_tp": 1, "seq_length": 7, "slow_but_exact": true, "torch_dtype": "float32", "type_vocab_size": 16, "use_cache": true, "vocab_size": 1024}
|
neuronxcc-2.15.128.0+56dc5a86/0_REGISTRY/0.0.25.dev0/inference/gpt2/gpt2/36aa91009c1c430c8dcf.json
ADDED
@@ -0,0 +1 @@
|
|
|
|
|
1 |
+
{"activation_function": "gelu_new", "architectures": ["GPT2LMHeadModel"], "attn_pdrop": 0.1, "bos_token_id": 50256, "embd_pdrop": 0.1, "eos_token_id": 50256, "initializer_range": 0.02, "layer_norm_epsilon": 1e-05, "model_type": "gpt2", "n_ctx": 1024, "n_embd": 768, "n_head": 12, "n_inner": null, "n_layer": 12, "n_positions": 1024, "neuron": {"auto_cast_type": "fp32", "batch_size": 1, "checkpoint_id": "gpt2", "checkpoint_revision": "607a30d783dfa663caf39e06633721c8d4cfcd7e", "compiler_type": "neuronx-cc", "compiler_version": "2.15.128.0+56dc5a86", "num_cores": 2, "sequence_length": 128, "task": "text-generation"}, "reorder_and_upcast_attn": false, "resid_pdrop": 0.1, "scale_attn_by_inverse_layer_idx": false, "scale_attn_weights": true, "summary_activation": null, "summary_first_dropout": 0.1, "summary_proj_to_labels": true, "summary_type": "cls_index", "summary_use_proj": true, "task_specific_params": {"text-generation": {"do_sample": true, "max_length": 50}}, "use_cache": true, "vocab_size": 50257}
|
neuronxcc-2.15.128.0+56dc5a86/0_REGISTRY/0.0.25.dev0/inference/gpt2/gpt2/98f807ee88912bf4b9d8.json
ADDED
@@ -0,0 +1 @@
|
|
|
|
|
1 |
+
{"activation_function": "gelu_new", "architectures": ["GPT2LMHeadModel"], "attn_pdrop": 0.1, "bos_token_id": 50256, "embd_pdrop": 0.1, "eos_token_id": 50256, "initializer_range": 0.02, "layer_norm_epsilon": 1e-05, "model_type": "gpt2", "n_ctx": 1024, "n_embd": 768, "n_head": 12, "n_inner": null, "n_layer": 12, "n_positions": 1024, "neuron": {"auto_cast_type": "fp16", "batch_size": 4, "checkpoint_id": "gpt2", "checkpoint_revision": "607a30d783dfa663caf39e06633721c8d4cfcd7e", "compiler_type": "neuronx-cc", "compiler_version": "2.15.128.0+56dc5a86", "num_cores": 2, "sequence_length": 1024, "task": "text-generation"}, "reorder_and_upcast_attn": false, "resid_pdrop": 0.1, "scale_attn_by_inverse_layer_idx": false, "scale_attn_weights": true, "summary_activation": null, "summary_first_dropout": 0.1, "summary_proj_to_labels": true, "summary_type": "cls_index", "summary_use_proj": true, "task_specific_params": {"text-generation": {"do_sample": true, "max_length": 50}}, "use_cache": true, "vocab_size": 50257}
|
neuronxcc-2.15.128.0+56dc5a86/0_REGISTRY/0.0.25.dev0/inference/gpt2/hf-internal-testing/tiny-random-gpt2/0b5b5348467d20a868da.json
ADDED
@@ -0,0 +1 @@
|
|
|
|
|
1 |
+
{"activation_function": "gelu_new", "attention_probs_dropout_prob": 0.1, "attn_pdrop": 0.1, "bos_token_id": 98, "embd_pdrop": 0.1, "eos_token_id": 98, "gradient_checkpointing": false, "hidden_act": "gelu", "hidden_dropout_prob": 0.1, "initializer_range": 0.02, "intermediate_size": 37, "layer_norm_epsilon": 1e-05, "model_type": "gpt2", "n_ctx": 512, "n_embd": 32, "n_head": 4, "n_inner": null, "n_layer": 5, "n_positions": 512, "neuron": {"auto_cast_type": "fp16", "batch_size": 2, "checkpoint_id": "hf-internal-testing/tiny-random-gpt2", "checkpoint_revision": "71034c5d8bde858ff824298bdedc65515b97d2b9", "compiler_type": "neuronx-cc", "compiler_version": "2.15.128.0+56dc5a86", "num_cores": 2, "sequence_length": 100, "task": "text-generation"}, "pad_token_id": 98, "reorder_and_upcast_attn": false, "resid_pdrop": 0.1, "scale_attn_by_inverse_layer_idx": false, "scale_attn_weights": true, "summary_activation": null, "summary_first_dropout": 0.1, "summary_proj_to_labels": true, "summary_type": "cls_index", "summary_use_proj": true, "type_vocab_size": 16, "use_cache": true, "vocab_size": 1000}
|
neuronxcc-2.15.128.0+56dc5a86/0_REGISTRY/0.0.25.dev0/inference/gpt2/hf-internal-testing/tiny-random-gpt2/3a30f319d7cae1f3076c.json
ADDED
@@ -0,0 +1 @@
|
|
|
|
|
1 |
+
{"activation_function": "gelu_new", "attention_probs_dropout_prob": 0.1, "attn_pdrop": 0.1, "bos_token_id": 98, "embd_pdrop": 0.1, "eos_token_id": 98, "gradient_checkpointing": false, "hidden_act": "gelu", "hidden_dropout_prob": 0.1, "initializer_range": 0.02, "intermediate_size": 37, "layer_norm_epsilon": 1e-05, "model_type": "gpt2", "n_ctx": 512, "n_embd": 32, "n_head": 4, "n_inner": null, "n_layer": 5, "n_positions": 512, "neuron": {"auto_cast_type": "fp32", "batch_size": 1, "checkpoint_id": "hf-internal-testing/tiny-random-gpt2", "checkpoint_revision": "71034c5d8bde858ff824298bdedc65515b97d2b9", "compiler_type": "neuronx-cc", "compiler_version": "2.15.128.0+56dc5a86", "num_cores": 2, "sequence_length": 100, "task": "text-generation"}, "pad_token_id": 98, "reorder_and_upcast_attn": false, "resid_pdrop": 0.1, "scale_attn_by_inverse_layer_idx": false, "scale_attn_weights": true, "summary_activation": null, "summary_first_dropout": 0.1, "summary_proj_to_labels": true, "summary_type": "cls_index", "summary_use_proj": true, "type_vocab_size": 16, "use_cache": true, "vocab_size": 1000}
|
neuronxcc-2.15.128.0+56dc5a86/0_REGISTRY/0.0.25.dev0/inference/gpt2/hf-internal-testing/tiny-random-gpt2/6b3192766c61d426df54.json
ADDED
@@ -0,0 +1 @@
|
|
|
|
|
1 |
+
{"activation_function": "gelu_new", "attention_probs_dropout_prob": 0.1, "attn_pdrop": 0.1, "bos_token_id": 98, "embd_pdrop": 0.1, "eos_token_id": 98, "gradient_checkpointing": false, "hidden_act": "gelu", "hidden_dropout_prob": 0.1, "initializer_range": 0.02, "intermediate_size": 37, "layer_norm_epsilon": 1e-05, "model_type": "gpt2", "n_ctx": 512, "n_embd": 32, "n_head": 4, "n_inner": null, "n_layer": 5, "n_positions": 512, "neuron": {"auto_cast_type": "fp32", "batch_size": 1, "checkpoint_id": "hf-internal-testing/tiny-random-gpt2", "checkpoint_revision": "71034c5d8bde858ff824298bdedc65515b97d2b9", "compiler_type": "neuronx-cc", "compiler_version": "2.15.128.0+56dc5a86", "num_cores": 24, "sequence_length": 512, "task": "text-generation"}, "pad_token_id": 98, "reorder_and_upcast_attn": false, "resid_pdrop": 0.1, "scale_attn_by_inverse_layer_idx": false, "scale_attn_weights": true, "summary_activation": null, "summary_first_dropout": 0.1, "summary_proj_to_labels": true, "summary_type": "cls_index", "summary_use_proj": true, "type_vocab_size": 16, "use_cache": true, "vocab_size": 1000}
|
neuronxcc-2.15.128.0+56dc5a86/0_REGISTRY/0.0.25.dev0/inference/gpt2/hf-internal-testing/tiny-random-gpt2/74bc6efd0eb46f234bef.json
ADDED
@@ -0,0 +1 @@
|
|
|
|
|
1 |
+
{"activation_function": "gelu_new", "attention_probs_dropout_prob": 0.1, "attn_pdrop": 0.1, "bos_token_id": 98, "embd_pdrop": 0.1, "eos_token_id": 98, "gradient_checkpointing": false, "hidden_act": "gelu", "hidden_dropout_prob": 0.1, "initializer_range": 0.02, "intermediate_size": 37, "layer_norm_epsilon": 1e-05, "model_type": "gpt2", "n_ctx": 512, "n_embd": 32, "n_head": 4, "n_inner": null, "n_layer": 5, "n_positions": 512, "neuron": {"auto_cast_type": "fp32", "batch_size": 1, "checkpoint_id": "hf-internal-testing/tiny-random-gpt2", "checkpoint_revision": "71034c5d8bde858ff824298bdedc65515b97d2b9", "compiler_type": "neuronx-cc", "compiler_version": "2.15.128.0+56dc5a86", "num_cores": 2, "sequence_length": 512, "task": "text-generation"}, "pad_token_id": 98, "reorder_and_upcast_attn": false, "resid_pdrop": 0.1, "scale_attn_by_inverse_layer_idx": false, "scale_attn_weights": true, "summary_activation": null, "summary_first_dropout": 0.1, "summary_proj_to_labels": true, "summary_type": "cls_index", "summary_use_proj": true, "type_vocab_size": 16, "use_cache": true, "vocab_size": 1000}
|
neuronxcc-2.15.128.0+56dc5a86/0_REGISTRY/0.0.25.dev0/inference/gpt2/hf-internal-testing/tiny-random-gpt2/8a6582d10bf46ea7004e.json
ADDED
@@ -0,0 +1 @@
|
|
|
|
|
1 |
+
{"activation_function": "gelu_new", "attention_probs_dropout_prob": 0.1, "attn_pdrop": 0.1, "bos_token_id": 98, "embd_pdrop": 0.1, "eos_token_id": 98, "gradient_checkpointing": false, "hidden_act": "gelu", "hidden_dropout_prob": 0.1, "initializer_range": 0.02, "intermediate_size": 37, "layer_norm_epsilon": 1e-05, "model_type": "gpt2", "n_ctx": 512, "n_embd": 32, "n_head": 4, "n_inner": null, "n_layer": 5, "n_positions": 512, "neuron": {"auto_cast_type": "fp16", "batch_size": 1, "checkpoint_id": "hf-internal-testing/tiny-random-gpt2", "checkpoint_revision": "71034c5d8bde858ff824298bdedc65515b97d2b9", "compiler_type": "neuronx-cc", "compiler_version": "2.15.128.0+56dc5a86", "num_cores": 2, "sequence_length": 100, "task": "text-generation"}, "pad_token_id": 98, "reorder_and_upcast_attn": false, "resid_pdrop": 0.1, "scale_attn_by_inverse_layer_idx": false, "scale_attn_weights": true, "summary_activation": null, "summary_first_dropout": 0.1, "summary_proj_to_labels": true, "summary_type": "cls_index", "summary_use_proj": true, "type_vocab_size": 16, "use_cache": true, "vocab_size": 1000}
|
neuronxcc-2.15.128.0+56dc5a86/0_REGISTRY/0.0.25.dev0/inference/llama/NousResearch/Hermes-2-Theta-Llama-3-8B/eda26b3891458ca144a4.json
ADDED
@@ -0,0 +1 @@
|
|
|
|
|
1 |
+
{"architectures": ["LlamaForCausalLM"], "attention_bias": false, "attention_dropout": 0.0, "bos_token_id": 128000, "eos_token_id": 128003, "hidden_act": "silu", "hidden_size": 4096, "initializer_range": 0.02, "intermediate_size": 14336, "max_position_embeddings": 8192, "mlp_bias": false, "model_type": "llama", "neuron": {"auto_cast_type": "fp16", "batch_size": 4, "checkpoint_id": "NousResearch/Hermes-2-Theta-Llama-3-8B", "checkpoint_revision": "d62e0c7237c7b851e8d9ae9277f9f107d174542c", "compiler_type": "neuronx-cc", "compiler_version": "2.15.128.0+56dc5a86", "num_cores": 2, "sequence_length": 4096, "task": "text-generation"}, "num_attention_heads": 32, "num_hidden_layers": 32, "num_key_value_heads": 8, "pretraining_tp": 1, "rms_norm_eps": 1e-05, "rope_scaling": null, "rope_theta": 500000.0, "tie_word_embeddings": false, "torch_dtype": "bfloat16", "use_cache": true, "vocab_size": 128256}
|
neuronxcc-2.15.128.0+56dc5a86/0_REGISTRY/0.0.25.dev0/inference/llama/dacorvo/tiny-random-llama/08720e1cfeb1befa20c2.json
ADDED
@@ -0,0 +1 @@
|
|
|
|
|
1 |
+
{"architectures": ["LlamaForCausalLM"], "attention_bias": false, "attention_dropout": 0.0, "bos_token_id": 1, "eos_token_id": 2, "hidden_act": "silu", "hidden_size": 128, "initializer_range": 0.02, "intermediate_size": 256, "max_position_embeddings": 512, "mlp_bias": false, "model_type": "llama", "neuron": {"auto_cast_type": "fp16", "batch_size": 2, "checkpoint_id": "dacorvo/tiny-random-llama", "checkpoint_revision": "7fdafd2fe6a2d31c6abb72ae60db606d8bb23196", "compiler_type": "neuronx-cc", "compiler_version": "2.15.128.0+56dc5a86", "num_cores": 2, "sequence_length": 100, "task": "text-generation"}, "num_attention_heads": 1, "num_hidden_layers": 1, "num_key_value_heads": 1, "pretraining_tp": 1, "rms_norm_eps": 1e-06, "rope_scaling": null, "rope_theta": 10000.0, "tie_word_embeddings": false, "torch_dtype": "float32", "use_cache": true, "vocab_size": 32000}
|
neuronxcc-2.15.128.0+56dc5a86/0_REGISTRY/0.0.25.dev0/inference/llama/dacorvo/tiny-random-llama/5ba8a7533020030cff1d.json
ADDED
@@ -0,0 +1 @@
|
|
|
|
|
1 |
+
{"architectures": ["LlamaForCausalLM"], "attention_bias": false, "attention_dropout": 0.0, "bos_token_id": 1, "eos_token_id": 2, "hidden_act": "silu", "hidden_size": 128, "initializer_range": 0.02, "intermediate_size": 256, "max_position_embeddings": 512, "mlp_bias": false, "model_type": "llama", "neuron": {"auto_cast_type": "fp16", "batch_size": 1, "checkpoint_id": "dacorvo/tiny-random-llama", "checkpoint_revision": "7fdafd2fe6a2d31c6abb72ae60db606d8bb23196", "compiler_type": "neuronx-cc", "compiler_version": "2.15.128.0+56dc5a86", "num_cores": 2, "sequence_length": 100, "task": "text-generation"}, "num_attention_heads": 1, "num_hidden_layers": 1, "num_key_value_heads": 1, "pretraining_tp": 1, "rms_norm_eps": 1e-06, "rope_scaling": null, "rope_theta": 10000.0, "tie_word_embeddings": false, "torch_dtype": "float32", "use_cache": true, "vocab_size": 32000}
|
neuronxcc-2.15.128.0+56dc5a86/0_REGISTRY/0.0.25.dev0/inference/llama/dacorvo/tiny-random-llama/e7337f4af7481c9d827e.json
ADDED
@@ -0,0 +1 @@
|
|
|
|
|
1 |
+
{"architectures": ["LlamaForCausalLM"], "attention_bias": false, "attention_dropout": 0.0, "bos_token_id": 1, "eos_token_id": 2, "hidden_act": "silu", "hidden_size": 128, "initializer_range": 0.02, "intermediate_size": 256, "max_position_embeddings": 512, "mlp_bias": false, "model_type": "llama", "neuron": {"auto_cast_type": "fp32", "batch_size": 1, "checkpoint_id": "dacorvo/tiny-random-llama", "checkpoint_revision": "7fdafd2fe6a2d31c6abb72ae60db606d8bb23196", "compiler_type": "neuronx-cc", "compiler_version": "2.15.128.0+56dc5a86", "num_cores": 2, "sequence_length": 100, "task": "text-generation"}, "num_attention_heads": 1, "num_hidden_layers": 1, "num_key_value_heads": 1, "pretraining_tp": 1, "rms_norm_eps": 1e-06, "rope_scaling": null, "rope_theta": 10000.0, "tie_word_embeddings": false, "torch_dtype": "float32", "use_cache": true, "vocab_size": 32000}
|
neuronxcc-2.15.128.0+56dc5a86/0_REGISTRY/0.0.25.dev0/inference/mistral/dacorvo/tiny-random-MistralForCausalLM/3bdcb3d6d5fb4b0b7097.json
ADDED
@@ -0,0 +1 @@
|
|
|
|
|
1 |
+
{"architectures": ["MistralForCausalLM"], "attention_dropout": 0.0, "attention_probs_dropout_prob": 0.1, "bos_token_id": 1, "eos_token_id": 2, "head_dim": 8, "hidden_act": "gelu", "hidden_dropout_prob": 0.1, "hidden_size": 32, "initializer_range": 0.02, "intermediate_size": 37, "is_decoder": true, "max_position_embeddings": 512, "model_type": "mistral", "neuron": {"auto_cast_type": "fp32", "batch_size": 1, "checkpoint_id": "dacorvo/tiny-random-MistralForCausalLM", "checkpoint_revision": "81d453e3c8985649e9ee3d4c9378461029d1c73a", "compiler_type": "neuronx-cc", "compiler_version": "2.15.128.0+56dc5a86", "num_cores": 2, "sequence_length": 100, "task": "text-generation"}, "num_attention_heads": 4, "num_hidden_layers": 2, "num_key_value_heads": 2, "pad_token_id": 0, "rms_norm_eps": 1e-06, "rope_theta": 10000.0, "sliding_window": 4096, "tie_word_embeddings": false, "torch_dtype": "float32", "type_vocab_size": 16, "use_cache": true, "vocab_size": 32000}
|
neuronxcc-2.15.128.0+56dc5a86/0_REGISTRY/0.0.25.dev0/inference/mistral/dacorvo/tiny-random-MistralForCausalLM/3d676d12f1566545d6b0.json
ADDED
@@ -0,0 +1 @@
|
|
|
|
|
1 |
+
{"architectures": ["MistralForCausalLM"], "attention_dropout": 0.0, "attention_probs_dropout_prob": 0.1, "bos_token_id": 1, "eos_token_id": 2, "head_dim": 8, "hidden_act": "gelu", "hidden_dropout_prob": 0.1, "hidden_size": 32, "initializer_range": 0.02, "intermediate_size": 37, "is_decoder": true, "max_position_embeddings": 512, "model_type": "mistral", "neuron": {"auto_cast_type": "fp16", "batch_size": 2, "checkpoint_id": "dacorvo/tiny-random-MistralForCausalLM", "checkpoint_revision": "81d453e3c8985649e9ee3d4c9378461029d1c73a", "compiler_type": "neuronx-cc", "compiler_version": "2.15.128.0+56dc5a86", "num_cores": 2, "sequence_length": 100, "task": "text-generation"}, "num_attention_heads": 4, "num_hidden_layers": 2, "num_key_value_heads": 2, "pad_token_id": 0, "rms_norm_eps": 1e-06, "rope_theta": 10000.0, "sliding_window": 4096, "tie_word_embeddings": false, "torch_dtype": "float32", "type_vocab_size": 16, "use_cache": true, "vocab_size": 32000}
|
neuronxcc-2.15.128.0+56dc5a86/0_REGISTRY/0.0.25.dev0/inference/mistral/dacorvo/tiny-random-MistralForCausalLM/ad4d8b389573be75fbee.json
ADDED
@@ -0,0 +1 @@
|
|
|
|
|
1 |
+
{"architectures": ["MistralForCausalLM"], "attention_dropout": 0.0, "attention_probs_dropout_prob": 0.1, "bos_token_id": 1, "eos_token_id": 2, "head_dim": 8, "hidden_act": "gelu", "hidden_dropout_prob": 0.1, "hidden_size": 32, "initializer_range": 0.02, "intermediate_size": 37, "is_decoder": true, "max_position_embeddings": 512, "model_type": "mistral", "neuron": {"auto_cast_type": "fp16", "batch_size": 1, "checkpoint_id": "dacorvo/tiny-random-MistralForCausalLM", "checkpoint_revision": "81d453e3c8985649e9ee3d4c9378461029d1c73a", "compiler_type": "neuronx-cc", "compiler_version": "2.15.128.0+56dc5a86", "num_cores": 2, "sequence_length": 100, "task": "text-generation"}, "num_attention_heads": 4, "num_hidden_layers": 2, "num_key_value_heads": 2, "pad_token_id": 0, "rms_norm_eps": 1e-06, "rope_theta": 10000.0, "sliding_window": 4096, "tie_word_embeddings": false, "torch_dtype": "float32", "type_vocab_size": 16, "use_cache": true, "vocab_size": 32000}
|
neuronxcc-2.15.128.0+56dc5a86/0_REGISTRY/0.0.25.dev0/inference/mistral/optimum/mistral-1.1b-testing/081b187113a5f417a9e0.json
ADDED
@@ -0,0 +1 @@
|
|
|
|
|
1 |
+
{"architectures": ["MistralForCausalLM"], "attention_dropout": 0.0, "bos_token_id": 1, "eos_token_id": 2, "head_dim": 64, "hidden_act": "silu", "hidden_size": 2048, "initializer_range": 0.02, "intermediate_size": 5632, "max_position_embeddings": 32768, "model_type": "mistral", "neuron": {"auto_cast_type": "bf16", "batch_size": 4, "checkpoint_id": "optimum/mistral-1.1b-testing", "checkpoint_revision": "ce03bc8d47dbd2c173ff65f3a8de1325ba724195", "compiler_type": "neuronx-cc", "compiler_version": "2.15.128.0+56dc5a86", "num_cores": 2, "sequence_length": 4096, "task": "text-generation"}, "num_attention_heads": 32, "num_hidden_layers": 22, "num_key_value_heads": 4, "rms_norm_eps": 1e-05, "rope_theta": 1000000.0, "sliding_window": null, "tie_word_embeddings": false, "torch_dtype": "bfloat16", "use_cache": true, "vocab_size": 32000}
|
neuronxcc-2.15.128.0+56dc5a86/0_REGISTRY/0.0.25.dev0/inference/mixtral/dacorvo/Mixtral-tiny/3a688b7dcd45d9a80b14.json
ADDED
@@ -0,0 +1 @@
|
|
|
|
|
1 |
+
{"architectures": ["MixtralForCausalLM"], "attention_dropout": 0.0, "bos_token_id": 1, "eos_token_id": 2, "hidden_act": "silu", "hidden_size": 1024, "initializer_range": 0.02, "intermediate_size": 3584, "max_position_embeddings": 1024, "model_type": "mixtral", "neuron": {"auto_cast_type": "fp32", "batch_size": 1, "checkpoint_id": "dacorvo/Mixtral-tiny", "checkpoint_revision": "c557ba205ddff6ea911f4719e0d543d6c08356b6", "compiler_type": "neuronx-cc", "compiler_version": "2.15.128.0+56dc5a86", "num_cores": 2, "sequence_length": 100, "task": "text-generation"}, "num_attention_heads": 32, "num_experts_per_tok": 2, "num_hidden_layers": 2, "num_key_value_heads": 8, "num_local_experts": 8, "output_router_logits": false, "rms_norm_eps": 1e-05, "rope_theta": 10000.0, "router_aux_loss_coef": 0.001, "router_jitter_noise": 0.0, "sliding_window": 4096, "tie_word_embeddings": false, "torch_dtype": "float16", "use_cache": true, "vocab_size": 32000}
|
neuronxcc-2.15.128.0+56dc5a86/0_REGISTRY/0.0.25.dev0/inference/mixtral/dacorvo/Mixtral-tiny/89980755d7c8bc1b31b0.json
ADDED
@@ -0,0 +1 @@
|
|
|
|
|
1 |
+
{"architectures": ["MixtralForCausalLM"], "attention_dropout": 0.0, "bos_token_id": 1, "eos_token_id": 2, "hidden_act": "silu", "hidden_size": 1024, "initializer_range": 0.02, "intermediate_size": 3584, "max_position_embeddings": 1024, "model_type": "mixtral", "neuron": {"auto_cast_type": "fp16", "batch_size": 4, "checkpoint_id": "dacorvo/Mixtral-tiny", "checkpoint_revision": "c557ba205ddff6ea911f4719e0d543d6c08356b6", "compiler_type": "neuronx-cc", "compiler_version": "2.15.128.0+56dc5a86", "num_cores": 2, "sequence_length": 1024, "task": "text-generation"}, "num_attention_heads": 32, "num_experts_per_tok": 2, "num_hidden_layers": 2, "num_key_value_heads": 8, "num_local_experts": 8, "output_router_logits": false, "rms_norm_eps": 1e-05, "rope_theta": 10000.0, "router_aux_loss_coef": 0.001, "router_jitter_noise": 0.0, "sliding_window": 4096, "tie_word_embeddings": false, "torch_dtype": "float16", "use_cache": true, "vocab_size": 32000}
|
neuronxcc-2.15.128.0+56dc5a86/0_REGISTRY/0.0.25.dev0/inference/mixtral/dacorvo/Mixtral-tiny/9f85ef35436cc6a20682.json
ADDED
@@ -0,0 +1 @@
|
|
|
|
|
1 |
+
{"architectures": ["MixtralForCausalLM"], "attention_dropout": 0.0, "bos_token_id": 1, "eos_token_id": 2, "hidden_act": "silu", "hidden_size": 1024, "initializer_range": 0.02, "intermediate_size": 3584, "max_position_embeddings": 1024, "model_type": "mixtral", "neuron": {"auto_cast_type": "fp16", "batch_size": 1, "checkpoint_id": "dacorvo/Mixtral-tiny", "checkpoint_revision": "c557ba205ddff6ea911f4719e0d543d6c08356b6", "compiler_type": "neuronx-cc", "compiler_version": "2.15.128.0+56dc5a86", "num_cores": 2, "sequence_length": 100, "task": "text-generation"}, "num_attention_heads": 32, "num_experts_per_tok": 2, "num_hidden_layers": 2, "num_key_value_heads": 8, "num_local_experts": 8, "output_router_logits": false, "rms_norm_eps": 1e-05, "rope_theta": 10000.0, "router_aux_loss_coef": 0.001, "router_jitter_noise": 0.0, "sliding_window": 4096, "tie_word_embeddings": false, "torch_dtype": "float16", "use_cache": true, "vocab_size": 32000}
|
neuronxcc-2.15.128.0+56dc5a86/0_REGISTRY/0.0.25.dev0/inference/mixtral/dacorvo/Mixtral-tiny/f1477b1f14a6669df1a7.json
ADDED
@@ -0,0 +1 @@
|
|
|
|
|
1 |
+
{"architectures": ["MixtralForCausalLM"], "attention_dropout": 0.0, "bos_token_id": 1, "eos_token_id": 2, "hidden_act": "silu", "hidden_size": 1024, "initializer_range": 0.02, "intermediate_size": 3584, "max_position_embeddings": 1024, "model_type": "mixtral", "neuron": {"auto_cast_type": "fp16", "batch_size": 2, "checkpoint_id": "dacorvo/Mixtral-tiny", "checkpoint_revision": "c557ba205ddff6ea911f4719e0d543d6c08356b6", "compiler_type": "neuronx-cc", "compiler_version": "2.15.128.0+56dc5a86", "num_cores": 2, "sequence_length": 100, "task": "text-generation"}, "num_attention_heads": 32, "num_experts_per_tok": 2, "num_hidden_layers": 2, "num_key_value_heads": 8, "num_local_experts": 8, "output_router_logits": false, "rms_norm_eps": 1e-05, "rope_theta": 10000.0, "router_aux_loss_coef": 0.001, "router_jitter_noise": 0.0, "sliding_window": 4096, "tie_word_embeddings": false, "torch_dtype": "float16", "use_cache": true, "vocab_size": 32000}
|
neuronxcc-2.15.128.0+56dc5a86/0_REGISTRY/0.0.25.dev0/inference/opt/hf-internal-testing/tiny-random-OPTForCausalLM/63ef91a527f73ba3b87e.json
ADDED
@@ -0,0 +1 @@
|
|
|
|
|
1 |
+
{"_remove_final_layer_norm": false, "activation_function": "relu", "architectures": ["OPTForCausalLM"], "attention_dropout": 0.1, "bos_token_id": 2, "do_layer_norm_before": true, "dropout": 0.1, "embed_dim": 16, "enable_bias": true, "eos_token_id": 2, "ffn_dim": 4, "hidden_size": 16, "init_std": 0.02, "is_decoder": true, "layer_norm_elementwise_affine": true, "layerdrop": 0.0, "max_position_embeddings": 100, "model_type": "opt", "neuron": {"auto_cast_type": "fp16", "batch_size": 1, "checkpoint_id": "hf-internal-testing/tiny-random-OPTForCausalLM", "checkpoint_revision": "190d1f4fc0011d2eaeaa05282e0fbd2445e4b11f", "compiler_type": "neuronx-cc", "compiler_version": "2.15.128.0+56dc5a86", "num_cores": 2, "sequence_length": 100, "task": "text-generation"}, "num_attention_heads": 4, "num_hidden_layers": 5, "pad_token_id": 1, "torch_dtype": "float32", "use_cache": true, "vocab_size": 50265, "word_embed_proj_dim": 16}
|
neuronxcc-2.15.128.0+56dc5a86/0_REGISTRY/0.0.25.dev0/inference/opt/hf-internal-testing/tiny-random-OPTForCausalLM/aa9fef08c5f79c9217fb.json
ADDED
@@ -0,0 +1 @@
|
|
|
|
|
1 |
+
{"_remove_final_layer_norm": false, "activation_function": "relu", "architectures": ["OPTForCausalLM"], "attention_dropout": 0.1, "bos_token_id": 2, "do_layer_norm_before": true, "dropout": 0.1, "embed_dim": 16, "enable_bias": true, "eos_token_id": 2, "ffn_dim": 4, "hidden_size": 16, "init_std": 0.02, "is_decoder": true, "layer_norm_elementwise_affine": true, "layerdrop": 0.0, "max_position_embeddings": 100, "model_type": "opt", "neuron": {"auto_cast_type": "fp32", "batch_size": 1, "checkpoint_id": "hf-internal-testing/tiny-random-OPTForCausalLM", "checkpoint_revision": "190d1f4fc0011d2eaeaa05282e0fbd2445e4b11f", "compiler_type": "neuronx-cc", "compiler_version": "2.15.128.0+56dc5a86", "num_cores": 2, "sequence_length": 100, "task": "text-generation"}, "num_attention_heads": 4, "num_hidden_layers": 5, "pad_token_id": 1, "torch_dtype": "float32", "use_cache": true, "vocab_size": 50265, "word_embed_proj_dim": 16}
|
neuronxcc-2.15.128.0+56dc5a86/0_REGISTRY/0.0.25.dev0/inference/opt/hf-internal-testing/tiny-random-OPTForCausalLM/fe9dc58f7d028523a0a6.json
ADDED
@@ -0,0 +1 @@
|
|
|
|
|
1 |
+
{"_remove_final_layer_norm": false, "activation_function": "relu", "architectures": ["OPTForCausalLM"], "attention_dropout": 0.1, "bos_token_id": 2, "do_layer_norm_before": true, "dropout": 0.1, "embed_dim": 16, "enable_bias": true, "eos_token_id": 2, "ffn_dim": 4, "hidden_size": 16, "init_std": 0.02, "is_decoder": true, "layer_norm_elementwise_affine": true, "layerdrop": 0.0, "max_position_embeddings": 100, "model_type": "opt", "neuron": {"auto_cast_type": "fp16", "batch_size": 2, "checkpoint_id": "hf-internal-testing/tiny-random-OPTForCausalLM", "checkpoint_revision": "190d1f4fc0011d2eaeaa05282e0fbd2445e4b11f", "compiler_type": "neuronx-cc", "compiler_version": "2.15.128.0+56dc5a86", "num_cores": 2, "sequence_length": 100, "task": "text-generation"}, "num_attention_heads": 4, "num_hidden_layers": 5, "pad_token_id": 1, "torch_dtype": "float32", "use_cache": true, "vocab_size": 50265, "word_embed_proj_dim": 16}
|
neuronxcc-2.15.128.0+56dc5a86/0_REGISTRY/0.0.25.dev0/inference/t5/hf-internal-testing/tiny-random-t5/af6cce39edb0ba1725d3.json
ADDED
@@ -0,0 +1 @@
|
|
|
|
|
1 |
+
{"decoder": {"classifier_dropout": 0.0, "d_ff": 37, "d_kv": 8, "d_model": 32, "decoder_start_token_id": 0, "dense_act_fn": "relu", "dropout_rate": 0.1, "feed_forward_proj": "relu", "gradient_checkpointing": false, "initializer_factor": 0.002, "is_encoder_decoder": true, "is_gated_act": false, "layer_norm_epsilon": 1e-06, "model_type": "t5", "neuron": {"compiler_type": "neuronx-cc", "compiler_version": "2.15.128.0+56dc5a86", "dynamic_batch_size": false, "inline_weights_to_neff": true, "optlevel": "2", "output_attentions": true, "output_hidden_states": true, "static_batch_size": 1, "static_num_beams": 4, "static_sequence_length": 18}, "num_decoder_layers": 5, "num_heads": 4, "num_layers": 5, "relative_attention_max_distance": 128, "relative_attention_num_buckets": 8, "task": "text2text-generation", "use_cache": true, "vocab_size": 1103}, "encoder": {"classifier_dropout": 0.0, "d_ff": 37, "d_kv": 8, "d_model": 32, "decoder_start_token_id": 0, "dense_act_fn": "relu", "dropout_rate": 0.1, "feed_forward_proj": "relu", "gradient_checkpointing": false, "initializer_factor": 0.002, "is_encoder_decoder": true, "is_gated_act": false, "layer_norm_epsilon": 1e-06, "model_type": "t5", "neuron": {"compiler_type": "neuronx-cc", "compiler_version": "2.15.128.0+56dc5a86", "dynamic_batch_size": false, "inline_weights_to_neff": true, "optlevel": "2", "output_attentions": true, "output_hidden_states": true, "static_batch_size": 1, "static_num_beams": 4, "static_sequence_length": 18}, "num_decoder_layers": 5, "num_heads": 4, "num_layers": 5, "relative_attention_max_distance": 128, "relative_attention_num_buckets": 8, "task": "text2text-generation", "use_cache": true, "vocab_size": 1103}, "model_type": "t5"}
|
neuronxcc-2.15.128.0+56dc5a86/MODULE_00a895c500f614d5e448+39f12043/compile_flags.json
ADDED
@@ -0,0 +1 @@
|
|
|
|
|
1 |
+
["--model-type=transformer", "--auto-cast=none", "--execute-repetition=1"]
|
neuronxcc-2.15.128.0+56dc5a86/MODULE_00a895c500f614d5e448+39f12043/model.done
ADDED
File without changes
|
neuronxcc-2.15.128.0+56dc5a86/MODULE_00a895c500f614d5e448+39f12043/model.hlo_module.pb
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:2fd039f830e38474b35ffa7875e575823b49051252f489fa859a10027c0b0176
|
3 |
+
size 56998
|
neuronxcc-2.15.128.0+56dc5a86/MODULE_00a895c500f614d5e448+39f12043/model.neff
ADDED
Binary file (124 kB). View file
|
|
neuronxcc-2.15.128.0+56dc5a86/MODULE_0270f34f54c498cedeec+39f12043/compile_flags.json
ADDED
@@ -0,0 +1 @@
|
|
|
|
|
1 |
+
["--model-type=transformer", "--auto-cast=none", "--execute-repetition=1"]
|
neuronxcc-2.15.128.0+56dc5a86/MODULE_0270f34f54c498cedeec+39f12043/model.done
ADDED
File without changes
|
neuronxcc-2.15.128.0+56dc5a86/MODULE_0270f34f54c498cedeec+39f12043/model.hlo_module.pb
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:55232d1f8a7bf3216037d725f1d9f971826f44b3f0bfe700dba381587050e2ea
|
3 |
+
size 46555
|
neuronxcc-2.15.128.0+56dc5a86/MODULE_0270f34f54c498cedeec+39f12043/model.neff
ADDED
Binary file (400 kB). View file
|
|
neuronxcc-2.15.128.0+56dc5a86/MODULE_05d01a658506892e6bcf+39f12043/compile_flags.json
ADDED
@@ -0,0 +1 @@
|
|
|
|
|
1 |
+
["--model-type=transformer", "--auto-cast=none", "--execute-repetition=1"]
|
neuronxcc-2.15.128.0+56dc5a86/MODULE_05d01a658506892e6bcf+39f12043/model.done
ADDED
File without changes
|
neuronxcc-2.15.128.0+56dc5a86/MODULE_05d01a658506892e6bcf+39f12043/model.hlo_module.pb
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:19bad7c7a1aca39e7acb9d1644e06f34534612f3f9d8da135a7ac71b567941c4
|
3 |
+
size 142880
|
neuronxcc-2.15.128.0+56dc5a86/MODULE_05d01a658506892e6bcf+39f12043/model.neff
ADDED
Binary file (646 kB). View file
|
|
neuronxcc-2.15.128.0+56dc5a86/MODULE_0d7bde45dbf2d5c06852+39f12043/compile_flags.json
ADDED
@@ -0,0 +1 @@
|
|
|
|
|
1 |
+
["--model-type=transformer", "--auto-cast=none", "--execute-repetition=1"]
|
neuronxcc-2.15.128.0+56dc5a86/MODULE_0d7bde45dbf2d5c06852+39f12043/model.done
ADDED
File without changes
|
neuronxcc-2.15.128.0+56dc5a86/MODULE_0d7bde45dbf2d5c06852+39f12043/model.hlo_module.pb
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:e8a4ae1405062b3adda459c6aa4a11045fe5add4801765e60a77d687a3c26e78
|
3 |
+
size 46216
|
neuronxcc-2.15.128.0+56dc5a86/MODULE_0d7bde45dbf2d5c06852+39f12043/model.neff
ADDED
Binary file (441 kB). View file
|
|
neuronxcc-2.15.128.0+56dc5a86/MODULE_0ea3601cf80a3b68f5bb/feature_extractor/preprocessor_config.json
ADDED
@@ -0,0 +1,27 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
{
|
2 |
+
"crop_size": {
|
3 |
+
"height": 224,
|
4 |
+
"width": 224
|
5 |
+
},
|
6 |
+
"do_center_crop": true,
|
7 |
+
"do_convert_rgb": true,
|
8 |
+
"do_normalize": true,
|
9 |
+
"do_rescale": true,
|
10 |
+
"do_resize": true,
|
11 |
+
"image_mean": [
|
12 |
+
0.48145466,
|
13 |
+
0.4578275,
|
14 |
+
0.40821073
|
15 |
+
],
|
16 |
+
"image_processor_type": "CLIPImageProcessor",
|
17 |
+
"image_std": [
|
18 |
+
0.26862954,
|
19 |
+
0.26130258,
|
20 |
+
0.27577711
|
21 |
+
],
|
22 |
+
"resample": 3,
|
23 |
+
"rescale_factor": 0.00392156862745098,
|
24 |
+
"size": {
|
25 |
+
"shortest_edge": 224
|
26 |
+
}
|
27 |
+
}
|
neuronxcc-2.15.128.0+56dc5a86/MODULE_0ea3601cf80a3b68f5bb/model_index.json
ADDED
@@ -0,0 +1,38 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
{
|
2 |
+
"_class_name": "StableDiffusionPipeline",
|
3 |
+
"_diffusers_version": "0.28.2",
|
4 |
+
"_name_or_path": "hf-internal-testing/tiny-stable-diffusion-torch",
|
5 |
+
"feature_extractor": [
|
6 |
+
"transformers",
|
7 |
+
"CLIPImageProcessor"
|
8 |
+
],
|
9 |
+
"image_encoder": [
|
10 |
+
null,
|
11 |
+
null
|
12 |
+
],
|
13 |
+
"requires_safety_checker": true,
|
14 |
+
"safety_checker": [
|
15 |
+
null,
|
16 |
+
null
|
17 |
+
],
|
18 |
+
"scheduler": [
|
19 |
+
"diffusers",
|
20 |
+
"PNDMScheduler"
|
21 |
+
],
|
22 |
+
"text_encoder": [
|
23 |
+
"transformers",
|
24 |
+
"CLIPTextModel"
|
25 |
+
],
|
26 |
+
"tokenizer": [
|
27 |
+
"transformers",
|
28 |
+
"CLIPTokenizer"
|
29 |
+
],
|
30 |
+
"unet": [
|
31 |
+
"diffusers",
|
32 |
+
"UNet2DConditionModel"
|
33 |
+
],
|
34 |
+
"vae": [
|
35 |
+
"diffusers",
|
36 |
+
"AutoencoderKL"
|
37 |
+
]
|
38 |
+
}
|
neuronxcc-2.15.128.0+56dc5a86/MODULE_0ea3601cf80a3b68f5bb/scheduler/scheduler_config.json
ADDED
@@ -0,0 +1,15 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
{
|
2 |
+
"_class_name": "PNDMScheduler",
|
3 |
+
"_diffusers_version": "0.28.2",
|
4 |
+
"beta_end": 0.012,
|
5 |
+
"beta_schedule": "scaled_linear",
|
6 |
+
"beta_start": 0.00085,
|
7 |
+
"clip_sample": false,
|
8 |
+
"num_train_timesteps": 1000,
|
9 |
+
"prediction_type": "epsilon",
|
10 |
+
"set_alpha_to_one": false,
|
11 |
+
"skip_prk_steps": true,
|
12 |
+
"steps_offset": 1,
|
13 |
+
"timestep_spacing": "leading",
|
14 |
+
"trained_betas": null
|
15 |
+
}
|
neuronxcc-2.15.128.0+56dc5a86/MODULE_0ea3601cf80a3b68f5bb/text_encoder/config.json
ADDED
@@ -0,0 +1,46 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
{
|
2 |
+
"_name_or_path": "/home/ubuntu/.cache/huggingface/hub/models--hf-internal-testing--tiny-stable-diffusion-torch/snapshots/a88cdfbd91f96ec7f61eb7484b652ff0f4ee701d/text_encoder",
|
3 |
+
"architectures": [
|
4 |
+
"CLIPTextModel"
|
5 |
+
],
|
6 |
+
"attention_dropout": 0.0,
|
7 |
+
"bos_token_id": 0,
|
8 |
+
"dropout": 0.0,
|
9 |
+
"eos_token_id": 2,
|
10 |
+
"hidden_act": "quick_gelu",
|
11 |
+
"hidden_size": 32,
|
12 |
+
"initializer_factor": 1.0,
|
13 |
+
"initializer_range": 0.02,
|
14 |
+
"intermediate_size": 37,
|
15 |
+
"layer_norm_eps": 1e-05,
|
16 |
+
"max_position_embeddings": 77,
|
17 |
+
"model_type": "clip_text_model",
|
18 |
+
"neuron": {
|
19 |
+
"compiler_type": "neuronx-cc",
|
20 |
+
"compiler_version": "2.15.128.0+56dc5a86",
|
21 |
+
"dynamic_batch_size": false,
|
22 |
+
"inline_weights_to_neff": true,
|
23 |
+
"input_names": [
|
24 |
+
"input_ids"
|
25 |
+
],
|
26 |
+
"model_type": "clip-text-model",
|
27 |
+
"optlevel": "2",
|
28 |
+
"output_attentions": false,
|
29 |
+
"output_hidden_states": false,
|
30 |
+
"output_names": [
|
31 |
+
"last_hidden_state",
|
32 |
+
"pooler_output"
|
33 |
+
],
|
34 |
+
"static_batch_size": 1,
|
35 |
+
"static_sequence_length": 77
|
36 |
+
},
|
37 |
+
"num_attention_heads": 4,
|
38 |
+
"num_hidden_layers": 5,
|
39 |
+
"pad_token_id": 1,
|
40 |
+
"projection_dim": 512,
|
41 |
+
"task": "feature-extraction",
|
42 |
+
"torch_dtype": "float32",
|
43 |
+
"torchscript": true,
|
44 |
+
"transformers_version": "4.43.2",
|
45 |
+
"vocab_size": 1000
|
46 |
+
}
|
neuronxcc-2.15.128.0+56dc5a86/MODULE_0ea3601cf80a3b68f5bb/text_encoder/model.neuron
ADDED
Binary file (826 kB). View file
|
|
neuronxcc-2.15.128.0+56dc5a86/MODULE_0ea3601cf80a3b68f5bb/tokenizer/merges.txt
ADDED
@@ -0,0 +1,647 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
#version: 0.2
|
2 |
+
Ġ t
|
3 |
+
Ġt h
|
4 |
+
Ġ a
|
5 |
+
Ġth e</w>
|
6 |
+
i n
|
7 |
+
Ġ o
|
8 |
+
Ġ ,</w>
|
9 |
+
Ġ s
|
10 |
+
e d</w>
|
11 |
+
Ġ w
|
12 |
+
e r
|
13 |
+
Ġ .</w>
|
14 |
+
Ġ i
|
15 |
+
r e
|
16 |
+
Ġ c
|
17 |
+
n d</w>
|
18 |
+
Ġ f
|
19 |
+
Ġ b
|
20 |
+
a t
|
21 |
+
Ġo f</w>
|
22 |
+
e r</w>
|
23 |
+
e n
|
24 |
+
a r
|
25 |
+
o r
|
26 |
+
i t
|
27 |
+
Ġ p
|
28 |
+
Ġ h
|
29 |
+
Ġa nd</w>
|
30 |
+
o n
|
31 |
+
in g</w>
|
32 |
+
a n
|
33 |
+
r o
|
34 |
+
Ġ m
|
35 |
+
Ġ d
|
36 |
+
e s</w>
|
37 |
+
Ġi n</w>
|
38 |
+
o n</w>
|
39 |
+
Ġt o</w>
|
40 |
+
o u
|
41 |
+
i s
|
42 |
+
Ġ a</w>
|
43 |
+
i c
|
44 |
+
Ġ T
|
45 |
+
a l
|
46 |
+
Ġ l
|
47 |
+
Ġ =</w>
|
48 |
+
Ġ re
|
49 |
+
Ġ "</w>
|
50 |
+
e s
|
51 |
+
Ġ S
|
52 |
+
a s</w>
|
53 |
+
a l</w>
|
54 |
+
i l
|
55 |
+
e l
|
56 |
+
i on</w>
|
57 |
+
Ġ A
|
58 |
+
Ġ C
|
59 |
+
Ġ 1
|
60 |
+
Ġ Ċ</w>
|
61 |
+
u r
|
62 |
+
ĠT h
|
63 |
+
Ġ n
|
64 |
+
a s
|
65 |
+
Ġ @
|
66 |
+
e c
|
67 |
+
o m
|
68 |
+
a c
|
69 |
+
Ġ e
|
70 |
+
Ġw as</w>
|
71 |
+
Ġ M
|
72 |
+
o r</w>
|
73 |
+
a n</w>
|
74 |
+
a m
|
75 |
+
e n</w>
|
76 |
+
o l
|
77 |
+
Ġ in
|
78 |
+
Ġ g
|
79 |
+
Ġ '</w>
|
80 |
+
Ġ B
|
81 |
+
l y</w>
|
82 |
+
a t</w>
|
83 |
+
i v
|
84 |
+
t s</w>
|
85 |
+
ĠTh e</w>
|
86 |
+
u s
|
87 |
+
- @</w>
|
88 |
+
Ġ@ -@</w>
|
89 |
+
i s</w>
|
90 |
+
Ġ I
|
91 |
+
Ġw h
|
92 |
+
i g
|
93 |
+
Ġ H
|
94 |
+
Ġs t
|
95 |
+
o s
|
96 |
+
u n
|
97 |
+
t h
|
98 |
+
Ġ P
|
99 |
+
Ġw it
|
100 |
+
Ġth at</w>
|
101 |
+
i r
|
102 |
+
Ġa s</w>
|
103 |
+
e m
|
104 |
+
Ġo n</w>
|
105 |
+
r a
|
106 |
+
Ġf or</w>
|
107 |
+
Ġ R
|
108 |
+
e t
|
109 |
+
o w
|
110 |
+
Ġ 2
|
111 |
+
i d
|
112 |
+
Ġ D
|
113 |
+
l e</w>
|
114 |
+
Ġwit h</w>
|
115 |
+
l a
|
116 |
+
en t</w>
|
117 |
+
i m
|
118 |
+
Ġ F
|
119 |
+
e a
|
120 |
+
i on
|
121 |
+
Ġb y</w>
|
122 |
+
Ġ )</w>
|
123 |
+
Ġ (</w>
|
124 |
+
Ġa l
|
125 |
+
Ġc on
|
126 |
+
en t
|
127 |
+
Ġ W
|
128 |
+
Ġi s</w>
|
129 |
+
er e</w>
|
130 |
+
Ġ G
|
131 |
+
Ġ N
|
132 |
+
Ġ L
|
133 |
+
Ġh a
|
134 |
+
er s</w>
|
135 |
+
r i
|
136 |
+
t h</w>
|
137 |
+
t ed</w>
|
138 |
+
u c
|
139 |
+
Ġ J
|
140 |
+
Ġ1 9
|
141 |
+
e v
|
142 |
+
u l
|
143 |
+
Ġ v
|
144 |
+
c e</w>
|
145 |
+
at ion</w>
|
146 |
+
ro m</w>
|
147 |
+
Ġb e
|
148 |
+
Ġ E
|
149 |
+
i n</w>
|
150 |
+
Ġth e
|
151 |
+
Ġf rom</w>
|
152 |
+
Ġ O
|
153 |
+
t er</w>
|
154 |
+
Ġp ro
|
155 |
+
Ġa r
|
156 |
+
a d
|
157 |
+
Ġc om
|
158 |
+
i c</w>
|
159 |
+
a g
|
160 |
+
Ġh is</w>
|
161 |
+
Ġs h
|
162 |
+
Ġa t</w>
|
163 |
+
o v
|
164 |
+
i es</w>
|
165 |
+
o o
|
166 |
+
p p
|
167 |
+
s t
|
168 |
+
c h
|
169 |
+
Ġ r
|
170 |
+
Ġ2 0
|
171 |
+
a y</w>
|
172 |
+
i f
|
173 |
+
Ġw ere</w>
|
174 |
+
Ġc h
|
175 |
+
u t</w>
|
176 |
+
s t</w>
|
177 |
+
u t
|
178 |
+
d s</w>
|
179 |
+
o p
|
180 |
+
u m
|
181 |
+
Ġi t</w>
|
182 |
+
o c
|
183 |
+
t er
|
184 |
+
l e
|
185 |
+
ig h
|
186 |
+
u d
|
187 |
+
Ġe x
|
188 |
+
ion s</w>
|
189 |
+
at e</w>
|
190 |
+
it y</w>
|
191 |
+
at ed</w>
|
192 |
+
Ġ un
|
193 |
+
e p
|
194 |
+
q u
|
195 |
+
Ġn o
|
196 |
+
Ġ K
|
197 |
+
iv e</w>
|
198 |
+
is t
|
199 |
+
Ġo n
|
200 |
+
am e</w>
|
201 |
+
ou n
|
202 |
+
i r</w>
|
203 |
+
a b
|
204 |
+
Ġ â
|
205 |
+
in g
|
206 |
+
Ġh e</w>
|
207 |
+
l d</w>
|
208 |
+
u g
|
209 |
+
ic h</w>
|
210 |
+
Ġa n</w>
|
211 |
+
e d
|
212 |
+
Ġ k
|
213 |
+
Ġâ Ģ
|
214 |
+
Ġha d</w>
|
215 |
+
v e</w>
|
216 |
+
a in
|
217 |
+
Ġs e
|
218 |
+
t ion</w>
|
219 |
+
or e</w>
|
220 |
+
re s
|
221 |
+
Ġwh ich</w>
|
222 |
+
ĠI n</w>
|
223 |
+
o d
|
224 |
+
th er</w>
|
225 |
+
a k
|
226 |
+
Ġs p
|
227 |
+
a r</w>
|
228 |
+
Ġ y
|
229 |
+
ĠC h
|
230 |
+
on g</w>
|
231 |
+
Ġa c
|
232 |
+
es t</w>
|
233 |
+
Ġ U
|
234 |
+
a p
|
235 |
+
f f
|
236 |
+
al ly</w>
|
237 |
+
r it
|
238 |
+
ĠS t
|
239 |
+
u b
|
240 |
+
g e</w>
|
241 |
+
b er</w>
|
242 |
+
e t</w>
|
243 |
+
Ġb e</w>
|
244 |
+
e ar
|
245 |
+
Ġre c
|
246 |
+
er s
|
247 |
+
Ġf ir
|
248 |
+
o t
|
249 |
+
Ġar e</w>
|
250 |
+
Ġa n
|
251 |
+
c h</w>
|
252 |
+
o g
|
253 |
+
i a</w>
|
254 |
+
es t
|
255 |
+
in e</w>
|
256 |
+
il l
|
257 |
+
an d
|
258 |
+
e l</w>
|
259 |
+
ar y</w>
|
260 |
+
e w</w>
|
261 |
+
i d</w>
|
262 |
+
Ġf or
|
263 |
+
Ġ ;</w>
|
264 |
+
Ġcom p
|
265 |
+
Ġ V
|
266 |
+
Ġin c
|
267 |
+
t r
|
268 |
+
Ġ20 0
|
269 |
+
Ġthe ir</w>
|
270 |
+
u s</w>
|
271 |
+
Ġb ut</w>
|
272 |
+
r an
|
273 |
+
ic al</w>
|
274 |
+
Ġfir st</w>
|
275 |
+
Ġd e
|
276 |
+
Ġin t
|
277 |
+
Ġ ro
|
278 |
+
s o</w>
|
279 |
+
ĠâĢ ĵ</w>
|
280 |
+
Ġno t</w>
|
281 |
+
d ing</w>
|
282 |
+
f ter</w>
|
283 |
+
ur e</w>
|
284 |
+
Ġp ar
|
285 |
+
Ġ :</w>
|
286 |
+
i an</w>
|
287 |
+
Ġt w
|
288 |
+
ou ld</w>
|
289 |
+
Ġal so</w>
|
290 |
+
Ġi ts</w>
|
291 |
+
Ġw or
|
292 |
+
u m</w>
|
293 |
+
Ġo r</w>
|
294 |
+
os t</w>
|
295 |
+
0 0</w>
|
296 |
+
ou r
|
297 |
+
ar d</w>
|
298 |
+
Ġre s
|
299 |
+
m p
|
300 |
+
u e</w>
|
301 |
+
Ġa b
|
302 |
+
is h</w>
|
303 |
+
Ġcon t
|
304 |
+
Ġa d
|
305 |
+
ow n</w>
|
306 |
+
al l</w>
|
307 |
+
ou g
|
308 |
+
Ġh er</w>
|
309 |
+
as t</w>
|
310 |
+
Ġ en
|
311 |
+
om e</w>
|
312 |
+
al l
|
313 |
+
d ed</w>
|
314 |
+
o w</w>
|
315 |
+
Ġha ve</w>
|
316 |
+
Ġ us
|
317 |
+
ea r</w>
|
318 |
+
ac k</w>
|
319 |
+
d uc
|
320 |
+
i al</w>
|
321 |
+
s s
|
322 |
+
en ts</w>
|
323 |
+
a in</w>
|
324 |
+
t ing</w>
|
325 |
+
Ġon e</w>
|
326 |
+
es s
|
327 |
+
Ġh as</w>
|
328 |
+
igh t</w>
|
329 |
+
a v
|
330 |
+
Ġe v
|
331 |
+
ou t</w>
|
332 |
+
a y
|
333 |
+
en ce</w>
|
334 |
+
Ġbe en</w>
|
335 |
+
e w
|
336 |
+
Ġtw o</w>
|
337 |
+
Ġc l
|
338 |
+
d er</w>
|
339 |
+
im e</w>
|
340 |
+
k s</w>
|
341 |
+
es s</w>
|
342 |
+
is h
|
343 |
+
. @</w>
|
344 |
+
Ġ@ .@</w>
|
345 |
+
Ġp la
|
346 |
+
Ġp l
|
347 |
+
Ġo r
|
348 |
+
u p</w>
|
349 |
+
m ent</w>
|
350 |
+
ur ing</w>
|
351 |
+
ol l
|
352 |
+
ĠI n
|
353 |
+
Ġth is</w>
|
354 |
+
Ġb ec
|
355 |
+
Ġcom m
|
356 |
+
Ġd is
|
357 |
+
at er</w>
|
358 |
+
ag e</w>
|
359 |
+
Ġa pp
|
360 |
+
ou s</w>
|
361 |
+
e y</w>
|
362 |
+
i l</w>
|
363 |
+
p er
|
364 |
+
ĠA l
|
365 |
+
ion al</w>
|
366 |
+
l ud
|
367 |
+
el y</w>
|
368 |
+
t t
|
369 |
+
il e</w>
|
370 |
+
i z
|
371 |
+
Ġ j
|
372 |
+
Ġwh o</w>
|
373 |
+
Ġa g
|
374 |
+
i b
|
375 |
+
Ġthe y</w>
|
376 |
+
f or
|
377 |
+
Ġo v
|
378 |
+
at h
|
379 |
+
e g
|
380 |
+
Ġs c
|
381 |
+
i p
|
382 |
+
Ġ20 1
|
383 |
+
Ġ 3
|
384 |
+
Ġp er
|
385 |
+
or y</w>
|
386 |
+
Ġd es
|
387 |
+
id e</w>
|
388 |
+
Ġs er
|
389 |
+
s e</w>
|
390 |
+
ĠH e</w>
|
391 |
+
la nd</w>
|
392 |
+
at ions</w>
|
393 |
+
r ic
|
394 |
+
i t</w>
|
395 |
+
re s</w>
|
396 |
+
er ed</w>
|
397 |
+
Ġp re
|
398 |
+
ĠS h
|
399 |
+
an ce</w>
|
400 |
+
or t</w>
|
401 |
+
an t</w>
|
402 |
+
, @</w>
|
403 |
+
Ġ@ ,@</w>
|
404 |
+
el l</w>
|
405 |
+
Ġ Y
|
406 |
+
n ed</w>
|
407 |
+
el l
|
408 |
+
it e</w>
|
409 |
+
Ġinc lud
|
410 |
+
Ġre p
|
411 |
+
Ġa fter</w>
|
412 |
+
Ġs uc
|
413 |
+
re e</w>
|
414 |
+
an y</w>
|
415 |
+
i m</w>
|
416 |
+
or t
|
417 |
+
Ġ1 8
|
418 |
+
Ġs u
|
419 |
+
ad e</w>
|
420 |
+
ou r</w>
|
421 |
+
ĠU n
|
422 |
+
ĠI t</w>
|
423 |
+
i k
|
424 |
+
ĠM ar
|
425 |
+
em ber</w>
|
426 |
+
Ġ 1</w>
|
427 |
+
e en</w>
|
428 |
+
a nd</w>
|
429 |
+
Ġs ec
|
430 |
+
ic e</w>
|
431 |
+
Ġt ime</w>
|
432 |
+
ĠA n
|
433 |
+
Ġint o</w>
|
434 |
+
Ġf in
|
435 |
+
Ġo ther</w>
|
436 |
+
Ġa tt
|
437 |
+
il l</w>
|
438 |
+
re n
|
439 |
+
ac h
|
440 |
+
as s
|
441 |
+
er al</w>
|
442 |
+
es e</w>
|
443 |
+
s h
|
444 |
+
al s</w>
|
445 |
+
it ion</w>
|
446 |
+
oug h</w>
|
447 |
+
l es</w>
|
448 |
+
am p
|
449 |
+
Ġw ould</w>
|
450 |
+
Ġm ore</w>
|
451 |
+
ro ug
|
452 |
+
ri b
|
453 |
+
er y</w>
|
454 |
+
ac e</w>
|
455 |
+
Ġ A</w>
|
456 |
+
Ġpla y
|
457 |
+
it ed</w>
|
458 |
+
k ed</w>
|
459 |
+
is t</w>
|
460 |
+
i ed</w>
|
461 |
+
Ġ 2</w>
|
462 |
+
as ed</w>
|
463 |
+
ing s</w>
|
464 |
+
an g
|
465 |
+
a m</w>
|
466 |
+
i p</w>
|
467 |
+
Ġb o
|
468 |
+
ab le</w>
|
469 |
+
t y</w>
|
470 |
+
Ġch ar
|
471 |
+
Ġc ent
|
472 |
+
et w
|
473 |
+
at es</w>
|
474 |
+
ro p
|
475 |
+
Ġ I</w>
|
476 |
+
u nd</w>
|
477 |
+
ĠA m
|
478 |
+
c es</w>
|
479 |
+
o in
|
480 |
+
Ġin ter
|
481 |
+
u p
|
482 |
+
c t
|
483 |
+
on e</w>
|
484 |
+
Ġt ra
|
485 |
+
an t
|
486 |
+
ec t
|
487 |
+
Ġal l</w>
|
488 |
+
e f
|
489 |
+
Ġcon s
|
490 |
+
ub l
|
491 |
+
n ing</w>
|
492 |
+
an s</w>
|
493 |
+
Ġf e
|
494 |
+
us t</w>
|
495 |
+
Ġ 0
|
496 |
+
Ġre m
|
497 |
+
as e</w>
|
498 |
+
on g
|
499 |
+
Ġwh en</w>
|
500 |
+
e b
|
501 |
+
ĠW h
|
502 |
+
Ġe ar
|
503 |
+
ev er</w>
|
504 |
+
Ġov er</w>
|
505 |
+
Ġk n
|
506 |
+
a us
|
507 |
+
Ġp os
|
508 |
+
a d</w>
|
509 |
+
er m
|
510 |
+
Ġsh e</w>
|
511 |
+
Ġ ra
|
512 |
+
Ġd uring</w>
|
513 |
+
as on</w>
|
514 |
+
v i
|
515 |
+
Ġex p
|
516 |
+
Ġl ea
|
517 |
+
Ġ el
|
518 |
+
Ġ 4
|
519 |
+
Ġon ly</w>
|
520 |
+
o nd</w>
|
521 |
+
Ġd ec
|
522 |
+
Ġac c
|
523 |
+
Ġo ff
|
524 |
+
is s
|
525 |
+
Ġf l
|
526 |
+
ĠE n
|
527 |
+
o t</w>
|
528 |
+
en s
|
529 |
+
os e</w>
|
530 |
+
ak e</w>
|
531 |
+
o m</w>
|
532 |
+
Ġs ev
|
533 |
+
ac h</w>
|
534 |
+
etw een</w>
|
535 |
+
er n
|
536 |
+
Ġ 3</w>
|
537 |
+
Ġp r
|
538 |
+
Ġg ro
|
539 |
+
r uc
|
540 |
+
Ġd i
|
541 |
+
Ġ19 9
|
542 |
+
ĠA r
|
543 |
+
Ġg ame</w>
|
544 |
+
Ġh im</w>
|
545 |
+
oo k</w>
|
546 |
+
Ġ up</w>
|
547 |
+
Ġab out</w>
|
548 |
+
Ġre l
|
549 |
+
for m
|
550 |
+
Ġth ree</w>
|
551 |
+
at t
|
552 |
+
ĠC om
|
553 |
+
Ġs a
|
554 |
+
ear s</w>
|
555 |
+
Ġ 5
|
556 |
+
r y</w>
|
557 |
+
Ġi mp
|
558 |
+
Ġm ost</w>
|
559 |
+
f er
|
560 |
+
Ġp res
|
561 |
+
Ġf il
|
562 |
+
Ġb etween</w>
|
563 |
+
Ġbe g
|
564 |
+
p h
|
565 |
+
or s</w>
|
566 |
+
Ġth an</w>
|
567 |
+
Ġrec or
|
568 |
+
o b
|
569 |
+
er ic
|
570 |
+
at ing</w>
|
571 |
+
Ġth roug
|
572 |
+
k ing</w>
|
573 |
+
Ġo ut</w>
|
574 |
+
Ġn um
|
575 |
+
oo d</w>
|
576 |
+
oll ow
|
577 |
+
ac t
|
578 |
+
u il
|
579 |
+
Ġc re
|
580 |
+
ol og
|
581 |
+
at ional</w>
|
582 |
+
Ġpro duc
|
583 |
+
Ġwh ile</w>
|
584 |
+
Ġl ater</w>
|
585 |
+
Ġw rit
|
586 |
+
e x
|
587 |
+
Ġst ar
|
588 |
+
Ġsp ec
|
589 |
+
e e
|
590 |
+
ish ed</w>
|
591 |
+
Ġre g
|
592 |
+
is ion</w>
|
593 |
+
ou th</w>
|
594 |
+
Ġre le
|
595 |
+
Ġa ss
|
596 |
+
Ġse ason</w>
|
597 |
+
Ġm ade</w>
|
598 |
+
il y</w>
|
599 |
+
r u
|
600 |
+
o y
|
601 |
+
t ur
|
602 |
+
t e</w>
|
603 |
+
Ġ qu
|
604 |
+
Ġm ov
|
605 |
+
ur y</w>
|
606 |
+
ĠAm eric
|
607 |
+
em ent</w>
|
608 |
+
c c
|
609 |
+
ou nd</w>
|
610 |
+
Ġl ar
|
611 |
+
Ġfor m
|
612 |
+
ec t</w>
|
613 |
+
Ġde f
|
614 |
+
Ġm us
|
615 |
+
ĠP ar
|
616 |
+
Ġm e
|
617 |
+
Ġs ub
|
618 |
+
w ay</w>
|
619 |
+
o p</w>
|
620 |
+
o h
|
621 |
+
el d</w>
|
622 |
+
i e</w>
|
623 |
+
em p
|
624 |
+
am es</w>
|
625 |
+
er n</w>
|
626 |
+
Ġn or
|
627 |
+
iv ed</w>
|
628 |
+
ev el
|
629 |
+
Ġsuc h</w>
|
630 |
+
ar ds</w>
|
631 |
+
Ġin d
|
632 |
+
ik e</w>
|
633 |
+
Ġg en
|
634 |
+
er t
|
635 |
+
Ġy ear</w>
|
636 |
+
Ġus ed</w>
|
637 |
+
Ġn ew</w>
|
638 |
+
Ġ 5</w>
|
639 |
+
Ġal b
|
640 |
+
s p
|
641 |
+
y p
|
642 |
+
Ġwit h
|
643 |
+
Ġwh ere</w>
|
644 |
+
ic s</w>
|
645 |
+
ĠTh is</w>
|
646 |
+
Ġthe m</w>
|
647 |
+
w n</w>
|
neuronxcc-2.15.128.0+56dc5a86/MODULE_0ea3601cf80a3b68f5bb/tokenizer/special_tokens_map.json
ADDED
@@ -0,0 +1,24 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
{
|
2 |
+
"bos_token": {
|
3 |
+
"content": "<|startoftext|>",
|
4 |
+
"lstrip": false,
|
5 |
+
"normalized": true,
|
6 |
+
"rstrip": false,
|
7 |
+
"single_word": false
|
8 |
+
},
|
9 |
+
"eos_token": {
|
10 |
+
"content": "<|endoftext|>",
|
11 |
+
"lstrip": false,
|
12 |
+
"normalized": true,
|
13 |
+
"rstrip": false,
|
14 |
+
"single_word": false
|
15 |
+
},
|
16 |
+
"pad_token": "<|endoftext|>",
|
17 |
+
"unk_token": {
|
18 |
+
"content": "<|endoftext|>",
|
19 |
+
"lstrip": false,
|
20 |
+
"normalized": true,
|
21 |
+
"rstrip": false,
|
22 |
+
"single_word": false
|
23 |
+
}
|
24 |
+
}
|