Synchronizing local compiler cache.
Browse filesThis view is limited to 50 files because it contains too many changes.
See raw diff
- .gitattributes +5 -0
- neuronxcc-2.15.128.0+56dc5a86/0_REGISTRY/0.0.25.dev0/inference/bloom/hf-internal-testing/tiny-random-BloomForCausalLM/12469cfbaa2be7979be9.json +1 -0
- neuronxcc-2.15.128.0+56dc5a86/0_REGISTRY/0.0.25.dev0/inference/bloom/hf-internal-testing/tiny-random-BloomForCausalLM/3047f14e5cd104f6da88.json +1 -0
- neuronxcc-2.15.128.0+56dc5a86/0_REGISTRY/0.0.25.dev0/inference/bloom/hf-internal-testing/tiny-random-BloomForCausalLM/cbf5e07921ee14bd6483.json +1 -0
- neuronxcc-2.15.128.0+56dc5a86/0_REGISTRY/0.0.25.dev0/inference/gpt2/gpt2/36aa91009c1c430c8dcf.json +1 -0
- neuronxcc-2.15.128.0+56dc5a86/0_REGISTRY/0.0.25.dev0/inference/gpt2/gpt2/98f807ee88912bf4b9d8.json +1 -0
- neuronxcc-2.15.128.0+56dc5a86/0_REGISTRY/0.0.25.dev0/inference/gpt2/hf-internal-testing/tiny-random-gpt2/0b5b5348467d20a868da.json +1 -0
- neuronxcc-2.15.128.0+56dc5a86/0_REGISTRY/0.0.25.dev0/inference/gpt2/hf-internal-testing/tiny-random-gpt2/3a30f319d7cae1f3076c.json +1 -0
- neuronxcc-2.15.128.0+56dc5a86/0_REGISTRY/0.0.25.dev0/inference/gpt2/hf-internal-testing/tiny-random-gpt2/74bc6efd0eb46f234bef.json +1 -0
- neuronxcc-2.15.128.0+56dc5a86/0_REGISTRY/0.0.25.dev0/inference/gpt2/hf-internal-testing/tiny-random-gpt2/8a6582d10bf46ea7004e.json +1 -0
- neuronxcc-2.15.128.0+56dc5a86/0_REGISTRY/0.0.25.dev0/inference/llama/NousResearch/Hermes-2-Theta-Llama-3-8B/eda26b3891458ca144a4.json +1 -0
- neuronxcc-2.15.128.0+56dc5a86/0_REGISTRY/0.0.25.dev0/inference/llama/dacorvo/tiny-random-llama/08720e1cfeb1befa20c2.json +1 -0
- neuronxcc-2.15.128.0+56dc5a86/0_REGISTRY/0.0.25.dev0/inference/llama/dacorvo/tiny-random-llama/5ba8a7533020030cff1d.json +1 -0
- neuronxcc-2.15.128.0+56dc5a86/0_REGISTRY/0.0.25.dev0/inference/llama/dacorvo/tiny-random-llama/e7337f4af7481c9d827e.json +1 -0
- neuronxcc-2.15.128.0+56dc5a86/0_REGISTRY/0.0.25.dev0/inference/mistral/dacorvo/tiny-random-MistralForCausalLM/3bdcb3d6d5fb4b0b7097.json +1 -0
- neuronxcc-2.15.128.0+56dc5a86/0_REGISTRY/0.0.25.dev0/inference/mistral/dacorvo/tiny-random-MistralForCausalLM/3d676d12f1566545d6b0.json +1 -0
- neuronxcc-2.15.128.0+56dc5a86/0_REGISTRY/0.0.25.dev0/inference/mistral/dacorvo/tiny-random-MistralForCausalLM/ad4d8b389573be75fbee.json +1 -0
- neuronxcc-2.15.128.0+56dc5a86/0_REGISTRY/0.0.25.dev0/inference/mistral/optimum/mistral-1.1b-testing/081b187113a5f417a9e0.json +1 -0
- neuronxcc-2.15.128.0+56dc5a86/0_REGISTRY/0.0.25.dev0/inference/mixtral/dacorvo/Mixtral-tiny/3a688b7dcd45d9a80b14.json +1 -0
- neuronxcc-2.15.128.0+56dc5a86/0_REGISTRY/0.0.25.dev0/inference/mixtral/dacorvo/Mixtral-tiny/89980755d7c8bc1b31b0.json +1 -0
- neuronxcc-2.15.128.0+56dc5a86/0_REGISTRY/0.0.25.dev0/inference/mixtral/dacorvo/Mixtral-tiny/9f85ef35436cc6a20682.json +1 -0
- neuronxcc-2.15.128.0+56dc5a86/0_REGISTRY/0.0.25.dev0/inference/mixtral/dacorvo/Mixtral-tiny/f1477b1f14a6669df1a7.json +1 -0
- neuronxcc-2.15.128.0+56dc5a86/0_REGISTRY/0.0.25.dev0/inference/opt/hf-internal-testing/tiny-random-OPTForCausalLM/63ef91a527f73ba3b87e.json +1 -0
- neuronxcc-2.15.128.0+56dc5a86/0_REGISTRY/0.0.25.dev0/inference/opt/hf-internal-testing/tiny-random-OPTForCausalLM/aa9fef08c5f79c9217fb.json +1 -0
- neuronxcc-2.15.128.0+56dc5a86/0_REGISTRY/0.0.25.dev0/inference/opt/hf-internal-testing/tiny-random-OPTForCausalLM/fe9dc58f7d028523a0a6.json +1 -0
- neuronxcc-2.15.128.0+56dc5a86/MODULE_00a895c500f614d5e448+39f12043/compile_flags.json +1 -0
- neuronxcc-2.15.128.0+56dc5a86/MODULE_00a895c500f614d5e448+39f12043/model.done +0 -0
- neuronxcc-2.15.128.0+56dc5a86/MODULE_00a895c500f614d5e448+39f12043/model.hlo_module.pb +3 -0
- neuronxcc-2.15.128.0+56dc5a86/MODULE_00a895c500f614d5e448+39f12043/model.neff +0 -0
- neuronxcc-2.15.128.0+56dc5a86/MODULE_0270f34f54c498cedeec+39f12043/compile_flags.json +1 -0
- neuronxcc-2.15.128.0+56dc5a86/MODULE_0270f34f54c498cedeec+39f12043/model.done +0 -0
- neuronxcc-2.15.128.0+56dc5a86/MODULE_0270f34f54c498cedeec+39f12043/model.hlo_module.pb +3 -0
- neuronxcc-2.15.128.0+56dc5a86/MODULE_0270f34f54c498cedeec+39f12043/model.neff +0 -0
- neuronxcc-2.15.128.0+56dc5a86/MODULE_05d01a658506892e6bcf+39f12043/compile_flags.json +1 -0
- neuronxcc-2.15.128.0+56dc5a86/MODULE_05d01a658506892e6bcf+39f12043/model.done +0 -0
- neuronxcc-2.15.128.0+56dc5a86/MODULE_05d01a658506892e6bcf+39f12043/model.hlo_module.pb +3 -0
- neuronxcc-2.15.128.0+56dc5a86/MODULE_05d01a658506892e6bcf+39f12043/model.neff +0 -0
- neuronxcc-2.15.128.0+56dc5a86/MODULE_0d7bde45dbf2d5c06852+39f12043/compile_flags.json +1 -0
- neuronxcc-2.15.128.0+56dc5a86/MODULE_0d7bde45dbf2d5c06852+39f12043/model.done +0 -0
- neuronxcc-2.15.128.0+56dc5a86/MODULE_0d7bde45dbf2d5c06852+39f12043/model.hlo_module.pb +3 -0
- neuronxcc-2.15.128.0+56dc5a86/MODULE_0d7bde45dbf2d5c06852+39f12043/model.neff +0 -0
- neuronxcc-2.15.128.0+56dc5a86/MODULE_162afdb41ad92db9e04c+39f12043/compile_flags.json +1 -0
- neuronxcc-2.15.128.0+56dc5a86/MODULE_162afdb41ad92db9e04c+39f12043/model.done +0 -0
- neuronxcc-2.15.128.0+56dc5a86/MODULE_162afdb41ad92db9e04c+39f12043/model.hlo_module.pb +3 -0
- neuronxcc-2.15.128.0+56dc5a86/MODULE_162afdb41ad92db9e04c+39f12043/model.neff +0 -0
- neuronxcc-2.15.128.0+56dc5a86/MODULE_176378d23aabee0ca25a+39f12043/compile_flags.json +1 -0
- neuronxcc-2.15.128.0+56dc5a86/MODULE_176378d23aabee0ca25a+39f12043/model.done +0 -0
- neuronxcc-2.15.128.0+56dc5a86/MODULE_176378d23aabee0ca25a+39f12043/model.hlo_module.pb +3 -0
- neuronxcc-2.15.128.0+56dc5a86/MODULE_176378d23aabee0ca25a+39f12043/model.neff +0 -0
- neuronxcc-2.15.128.0+56dc5a86/MODULE_1bfeeb4391752eee11c4+aa032128/compile_flags.json +1 -0
.gitattributes
CHANGED
@@ -917,3 +917,8 @@ neuronxcc-2.14.227.0+2d4f85be/MODULE_fcd85d066cb013a59267+2c2d707e/model.neff fi
|
|
917 |
neuronxcc-2.14.227.0+2d4f85be/MODULE_ff563de6b8eaa289eed4+2c2d707e/model.neff filter=lfs diff=lfs merge=lfs -text
|
918 |
neuronxcc-2.14.227.0+2d4f85be/MODULE_ffa764237818d906983b+2c2d707e/model.neff filter=lfs diff=lfs merge=lfs -text
|
919 |
neuronxcc-2.14.227.0+2d4f85be/MODULE_ffd9059473e785125324+2c2d707e/model.neff filter=lfs diff=lfs merge=lfs -text
|
|
|
|
|
|
|
|
|
|
|
|
917 |
neuronxcc-2.14.227.0+2d4f85be/MODULE_ff563de6b8eaa289eed4+2c2d707e/model.neff filter=lfs diff=lfs merge=lfs -text
|
918 |
neuronxcc-2.14.227.0+2d4f85be/MODULE_ffa764237818d906983b+2c2d707e/model.neff filter=lfs diff=lfs merge=lfs -text
|
919 |
neuronxcc-2.14.227.0+2d4f85be/MODULE_ffd9059473e785125324+2c2d707e/model.neff filter=lfs diff=lfs merge=lfs -text
|
920 |
+
neuronxcc-2.15.128.0+56dc5a86/MODULE_3a7777a73798af802ac1+39f12043/model.neff filter=lfs diff=lfs merge=lfs -text
|
921 |
+
neuronxcc-2.15.128.0+56dc5a86/MODULE_3da92ce16fff3bc522f8+39f12043/model.neff filter=lfs diff=lfs merge=lfs -text
|
922 |
+
neuronxcc-2.15.128.0+56dc5a86/MODULE_3e0ab9335d16a5ec49ef+39f12043/model.neff filter=lfs diff=lfs merge=lfs -text
|
923 |
+
neuronxcc-2.15.128.0+56dc5a86/MODULE_d1618c04ba6ca3abe828+39f12043/model.neff filter=lfs diff=lfs merge=lfs -text
|
924 |
+
neuronxcc-2.15.128.0+56dc5a86/MODULE_d36186ef4c022d90fe8c+39f12043/model.neff filter=lfs diff=lfs merge=lfs -text
|
neuronxcc-2.15.128.0+56dc5a86/0_REGISTRY/0.0.25.dev0/inference/bloom/hf-internal-testing/tiny-random-BloomForCausalLM/12469cfbaa2be7979be9.json
ADDED
@@ -0,0 +1 @@
|
|
|
|
|
1 |
+
{"apply_residual_connection_post_layernorm": false, "architectures": ["BloomForCausalLM"], "attention_dropout": 0.1, "bos_token_id": 1, "dtype": "float32", "eos_token_id": 2, "gradient_checkpointing": false, "hidden_dropout": 0.1, "hidden_size": 32, "id2label": {"0": "LABEL_0", "1": "LABEL_1", "2": "LABEL_2"}, "initializer_range": 0.02, "is_decoder": true, "label2id": {"LABEL_0": 0, "LABEL_1": 1, "LABEL_2": 2}, "layer_norm_epsilon": 1e-05, "model_type": "bloom", "n_head": 4, "n_layer": 5, "n_positions": 512, "neuron": {"auto_cast_type": "fp16", "batch_size": 1, "checkpoint_id": "hf-internal-testing/tiny-random-BloomForCausalLM", "checkpoint_revision": "92b07e9b7b4f986fa7c54e2ac3b9201b4ba5212e", "compiler_type": "neuronx-cc", "compiler_version": "2.15.128.0+56dc5a86", "num_cores": 2, "sequence_length": 100, "task": "text-generation"}, "pad_token_id": 3, "pretraining_tp": 1, "seq_length": 7, "slow_but_exact": true, "torch_dtype": "float32", "type_vocab_size": 16, "use_cache": true, "vocab_size": 1024}
|
neuronxcc-2.15.128.0+56dc5a86/0_REGISTRY/0.0.25.dev0/inference/bloom/hf-internal-testing/tiny-random-BloomForCausalLM/3047f14e5cd104f6da88.json
ADDED
@@ -0,0 +1 @@
|
|
|
|
|
1 |
+
{"apply_residual_connection_post_layernorm": false, "architectures": ["BloomForCausalLM"], "attention_dropout": 0.1, "bos_token_id": 1, "dtype": "float32", "eos_token_id": 2, "gradient_checkpointing": false, "hidden_dropout": 0.1, "hidden_size": 32, "id2label": {"0": "LABEL_0", "1": "LABEL_1", "2": "LABEL_2"}, "initializer_range": 0.02, "is_decoder": true, "label2id": {"LABEL_0": 0, "LABEL_1": 1, "LABEL_2": 2}, "layer_norm_epsilon": 1e-05, "model_type": "bloom", "n_head": 4, "n_layer": 5, "n_positions": 512, "neuron": {"auto_cast_type": "fp32", "batch_size": 1, "checkpoint_id": "hf-internal-testing/tiny-random-BloomForCausalLM", "checkpoint_revision": "92b07e9b7b4f986fa7c54e2ac3b9201b4ba5212e", "compiler_type": "neuronx-cc", "compiler_version": "2.15.128.0+56dc5a86", "num_cores": 2, "sequence_length": 100, "task": "text-generation"}, "pad_token_id": 3, "pretraining_tp": 1, "seq_length": 7, "slow_but_exact": true, "torch_dtype": "float32", "type_vocab_size": 16, "use_cache": true, "vocab_size": 1024}
|
neuronxcc-2.15.128.0+56dc5a86/0_REGISTRY/0.0.25.dev0/inference/bloom/hf-internal-testing/tiny-random-BloomForCausalLM/cbf5e07921ee14bd6483.json
ADDED
@@ -0,0 +1 @@
|
|
|
|
|
1 |
+
{"apply_residual_connection_post_layernorm": false, "architectures": ["BloomForCausalLM"], "attention_dropout": 0.1, "bos_token_id": 1, "dtype": "float32", "eos_token_id": 2, "gradient_checkpointing": false, "hidden_dropout": 0.1, "hidden_size": 32, "id2label": {"0": "LABEL_0", "1": "LABEL_1", "2": "LABEL_2"}, "initializer_range": 0.02, "is_decoder": true, "label2id": {"LABEL_0": 0, "LABEL_1": 1, "LABEL_2": 2}, "layer_norm_epsilon": 1e-05, "model_type": "bloom", "n_head": 4, "n_layer": 5, "n_positions": 512, "neuron": {"auto_cast_type": "fp16", "batch_size": 2, "checkpoint_id": "hf-internal-testing/tiny-random-BloomForCausalLM", "checkpoint_revision": "92b07e9b7b4f986fa7c54e2ac3b9201b4ba5212e", "compiler_type": "neuronx-cc", "compiler_version": "2.15.128.0+56dc5a86", "num_cores": 2, "sequence_length": 100, "task": "text-generation"}, "pad_token_id": 3, "pretraining_tp": 1, "seq_length": 7, "slow_but_exact": true, "torch_dtype": "float32", "type_vocab_size": 16, "use_cache": true, "vocab_size": 1024}
|
neuronxcc-2.15.128.0+56dc5a86/0_REGISTRY/0.0.25.dev0/inference/gpt2/gpt2/36aa91009c1c430c8dcf.json
ADDED
@@ -0,0 +1 @@
|
|
|
|
|
1 |
+
{"activation_function": "gelu_new", "architectures": ["GPT2LMHeadModel"], "attn_pdrop": 0.1, "bos_token_id": 50256, "embd_pdrop": 0.1, "eos_token_id": 50256, "initializer_range": 0.02, "layer_norm_epsilon": 1e-05, "model_type": "gpt2", "n_ctx": 1024, "n_embd": 768, "n_head": 12, "n_inner": null, "n_layer": 12, "n_positions": 1024, "neuron": {"auto_cast_type": "fp32", "batch_size": 1, "checkpoint_id": "gpt2", "checkpoint_revision": "607a30d783dfa663caf39e06633721c8d4cfcd7e", "compiler_type": "neuronx-cc", "compiler_version": "2.15.128.0+56dc5a86", "num_cores": 2, "sequence_length": 128, "task": "text-generation"}, "reorder_and_upcast_attn": false, "resid_pdrop": 0.1, "scale_attn_by_inverse_layer_idx": false, "scale_attn_weights": true, "summary_activation": null, "summary_first_dropout": 0.1, "summary_proj_to_labels": true, "summary_type": "cls_index", "summary_use_proj": true, "task_specific_params": {"text-generation": {"do_sample": true, "max_length": 50}}, "use_cache": true, "vocab_size": 50257}
|
neuronxcc-2.15.128.0+56dc5a86/0_REGISTRY/0.0.25.dev0/inference/gpt2/gpt2/98f807ee88912bf4b9d8.json
ADDED
@@ -0,0 +1 @@
|
|
|
|
|
1 |
+
{"activation_function": "gelu_new", "architectures": ["GPT2LMHeadModel"], "attn_pdrop": 0.1, "bos_token_id": 50256, "embd_pdrop": 0.1, "eos_token_id": 50256, "initializer_range": 0.02, "layer_norm_epsilon": 1e-05, "model_type": "gpt2", "n_ctx": 1024, "n_embd": 768, "n_head": 12, "n_inner": null, "n_layer": 12, "n_positions": 1024, "neuron": {"auto_cast_type": "fp16", "batch_size": 4, "checkpoint_id": "gpt2", "checkpoint_revision": "607a30d783dfa663caf39e06633721c8d4cfcd7e", "compiler_type": "neuronx-cc", "compiler_version": "2.15.128.0+56dc5a86", "num_cores": 2, "sequence_length": 1024, "task": "text-generation"}, "reorder_and_upcast_attn": false, "resid_pdrop": 0.1, "scale_attn_by_inverse_layer_idx": false, "scale_attn_weights": true, "summary_activation": null, "summary_first_dropout": 0.1, "summary_proj_to_labels": true, "summary_type": "cls_index", "summary_use_proj": true, "task_specific_params": {"text-generation": {"do_sample": true, "max_length": 50}}, "use_cache": true, "vocab_size": 50257}
|
neuronxcc-2.15.128.0+56dc5a86/0_REGISTRY/0.0.25.dev0/inference/gpt2/hf-internal-testing/tiny-random-gpt2/0b5b5348467d20a868da.json
ADDED
@@ -0,0 +1 @@
|
|
|
|
|
1 |
+
{"activation_function": "gelu_new", "attention_probs_dropout_prob": 0.1, "attn_pdrop": 0.1, "bos_token_id": 98, "embd_pdrop": 0.1, "eos_token_id": 98, "gradient_checkpointing": false, "hidden_act": "gelu", "hidden_dropout_prob": 0.1, "initializer_range": 0.02, "intermediate_size": 37, "layer_norm_epsilon": 1e-05, "model_type": "gpt2", "n_ctx": 512, "n_embd": 32, "n_head": 4, "n_inner": null, "n_layer": 5, "n_positions": 512, "neuron": {"auto_cast_type": "fp16", "batch_size": 2, "checkpoint_id": "hf-internal-testing/tiny-random-gpt2", "checkpoint_revision": "71034c5d8bde858ff824298bdedc65515b97d2b9", "compiler_type": "neuronx-cc", "compiler_version": "2.15.128.0+56dc5a86", "num_cores": 2, "sequence_length": 100, "task": "text-generation"}, "pad_token_id": 98, "reorder_and_upcast_attn": false, "resid_pdrop": 0.1, "scale_attn_by_inverse_layer_idx": false, "scale_attn_weights": true, "summary_activation": null, "summary_first_dropout": 0.1, "summary_proj_to_labels": true, "summary_type": "cls_index", "summary_use_proj": true, "type_vocab_size": 16, "use_cache": true, "vocab_size": 1000}
|
neuronxcc-2.15.128.0+56dc5a86/0_REGISTRY/0.0.25.dev0/inference/gpt2/hf-internal-testing/tiny-random-gpt2/3a30f319d7cae1f3076c.json
ADDED
@@ -0,0 +1 @@
|
|
|
|
|
1 |
+
{"activation_function": "gelu_new", "attention_probs_dropout_prob": 0.1, "attn_pdrop": 0.1, "bos_token_id": 98, "embd_pdrop": 0.1, "eos_token_id": 98, "gradient_checkpointing": false, "hidden_act": "gelu", "hidden_dropout_prob": 0.1, "initializer_range": 0.02, "intermediate_size": 37, "layer_norm_epsilon": 1e-05, "model_type": "gpt2", "n_ctx": 512, "n_embd": 32, "n_head": 4, "n_inner": null, "n_layer": 5, "n_positions": 512, "neuron": {"auto_cast_type": "fp32", "batch_size": 1, "checkpoint_id": "hf-internal-testing/tiny-random-gpt2", "checkpoint_revision": "71034c5d8bde858ff824298bdedc65515b97d2b9", "compiler_type": "neuronx-cc", "compiler_version": "2.15.128.0+56dc5a86", "num_cores": 2, "sequence_length": 100, "task": "text-generation"}, "pad_token_id": 98, "reorder_and_upcast_attn": false, "resid_pdrop": 0.1, "scale_attn_by_inverse_layer_idx": false, "scale_attn_weights": true, "summary_activation": null, "summary_first_dropout": 0.1, "summary_proj_to_labels": true, "summary_type": "cls_index", "summary_use_proj": true, "type_vocab_size": 16, "use_cache": true, "vocab_size": 1000}
|
neuronxcc-2.15.128.0+56dc5a86/0_REGISTRY/0.0.25.dev0/inference/gpt2/hf-internal-testing/tiny-random-gpt2/74bc6efd0eb46f234bef.json
ADDED
@@ -0,0 +1 @@
|
|
|
|
|
1 |
+
{"activation_function": "gelu_new", "attention_probs_dropout_prob": 0.1, "attn_pdrop": 0.1, "bos_token_id": 98, "embd_pdrop": 0.1, "eos_token_id": 98, "gradient_checkpointing": false, "hidden_act": "gelu", "hidden_dropout_prob": 0.1, "initializer_range": 0.02, "intermediate_size": 37, "layer_norm_epsilon": 1e-05, "model_type": "gpt2", "n_ctx": 512, "n_embd": 32, "n_head": 4, "n_inner": null, "n_layer": 5, "n_positions": 512, "neuron": {"auto_cast_type": "fp32", "batch_size": 1, "checkpoint_id": "hf-internal-testing/tiny-random-gpt2", "checkpoint_revision": "71034c5d8bde858ff824298bdedc65515b97d2b9", "compiler_type": "neuronx-cc", "compiler_version": "2.15.128.0+56dc5a86", "num_cores": 2, "sequence_length": 512, "task": "text-generation"}, "pad_token_id": 98, "reorder_and_upcast_attn": false, "resid_pdrop": 0.1, "scale_attn_by_inverse_layer_idx": false, "scale_attn_weights": true, "summary_activation": null, "summary_first_dropout": 0.1, "summary_proj_to_labels": true, "summary_type": "cls_index", "summary_use_proj": true, "type_vocab_size": 16, "use_cache": true, "vocab_size": 1000}
|
neuronxcc-2.15.128.0+56dc5a86/0_REGISTRY/0.0.25.dev0/inference/gpt2/hf-internal-testing/tiny-random-gpt2/8a6582d10bf46ea7004e.json
ADDED
@@ -0,0 +1 @@
|
|
|
|
|
1 |
+
{"activation_function": "gelu_new", "attention_probs_dropout_prob": 0.1, "attn_pdrop": 0.1, "bos_token_id": 98, "embd_pdrop": 0.1, "eos_token_id": 98, "gradient_checkpointing": false, "hidden_act": "gelu", "hidden_dropout_prob": 0.1, "initializer_range": 0.02, "intermediate_size": 37, "layer_norm_epsilon": 1e-05, "model_type": "gpt2", "n_ctx": 512, "n_embd": 32, "n_head": 4, "n_inner": null, "n_layer": 5, "n_positions": 512, "neuron": {"auto_cast_type": "fp16", "batch_size": 1, "checkpoint_id": "hf-internal-testing/tiny-random-gpt2", "checkpoint_revision": "71034c5d8bde858ff824298bdedc65515b97d2b9", "compiler_type": "neuronx-cc", "compiler_version": "2.15.128.0+56dc5a86", "num_cores": 2, "sequence_length": 100, "task": "text-generation"}, "pad_token_id": 98, "reorder_and_upcast_attn": false, "resid_pdrop": 0.1, "scale_attn_by_inverse_layer_idx": false, "scale_attn_weights": true, "summary_activation": null, "summary_first_dropout": 0.1, "summary_proj_to_labels": true, "summary_type": "cls_index", "summary_use_proj": true, "type_vocab_size": 16, "use_cache": true, "vocab_size": 1000}
|
neuronxcc-2.15.128.0+56dc5a86/0_REGISTRY/0.0.25.dev0/inference/llama/NousResearch/Hermes-2-Theta-Llama-3-8B/eda26b3891458ca144a4.json
ADDED
@@ -0,0 +1 @@
|
|
|
|
|
1 |
+
{"architectures": ["LlamaForCausalLM"], "attention_bias": false, "attention_dropout": 0.0, "bos_token_id": 128000, "eos_token_id": 128003, "hidden_act": "silu", "hidden_size": 4096, "initializer_range": 0.02, "intermediate_size": 14336, "max_position_embeddings": 8192, "mlp_bias": false, "model_type": "llama", "neuron": {"auto_cast_type": "fp16", "batch_size": 4, "checkpoint_id": "NousResearch/Hermes-2-Theta-Llama-3-8B", "checkpoint_revision": "d62e0c7237c7b851e8d9ae9277f9f107d174542c", "compiler_type": "neuronx-cc", "compiler_version": "2.15.128.0+56dc5a86", "num_cores": 2, "sequence_length": 4096, "task": "text-generation"}, "num_attention_heads": 32, "num_hidden_layers": 32, "num_key_value_heads": 8, "pretraining_tp": 1, "rms_norm_eps": 1e-05, "rope_scaling": null, "rope_theta": 500000.0, "tie_word_embeddings": false, "torch_dtype": "bfloat16", "use_cache": true, "vocab_size": 128256}
|
neuronxcc-2.15.128.0+56dc5a86/0_REGISTRY/0.0.25.dev0/inference/llama/dacorvo/tiny-random-llama/08720e1cfeb1befa20c2.json
ADDED
@@ -0,0 +1 @@
|
|
|
|
|
1 |
+
{"architectures": ["LlamaForCausalLM"], "attention_bias": false, "attention_dropout": 0.0, "bos_token_id": 1, "eos_token_id": 2, "hidden_act": "silu", "hidden_size": 128, "initializer_range": 0.02, "intermediate_size": 256, "max_position_embeddings": 512, "mlp_bias": false, "model_type": "llama", "neuron": {"auto_cast_type": "fp16", "batch_size": 2, "checkpoint_id": "dacorvo/tiny-random-llama", "checkpoint_revision": "7fdafd2fe6a2d31c6abb72ae60db606d8bb23196", "compiler_type": "neuronx-cc", "compiler_version": "2.15.128.0+56dc5a86", "num_cores": 2, "sequence_length": 100, "task": "text-generation"}, "num_attention_heads": 1, "num_hidden_layers": 1, "num_key_value_heads": 1, "pretraining_tp": 1, "rms_norm_eps": 1e-06, "rope_scaling": null, "rope_theta": 10000.0, "tie_word_embeddings": false, "torch_dtype": "float32", "use_cache": true, "vocab_size": 32000}
|
neuronxcc-2.15.128.0+56dc5a86/0_REGISTRY/0.0.25.dev0/inference/llama/dacorvo/tiny-random-llama/5ba8a7533020030cff1d.json
ADDED
@@ -0,0 +1 @@
|
|
|
|
|
1 |
+
{"architectures": ["LlamaForCausalLM"], "attention_bias": false, "attention_dropout": 0.0, "bos_token_id": 1, "eos_token_id": 2, "hidden_act": "silu", "hidden_size": 128, "initializer_range": 0.02, "intermediate_size": 256, "max_position_embeddings": 512, "mlp_bias": false, "model_type": "llama", "neuron": {"auto_cast_type": "fp16", "batch_size": 1, "checkpoint_id": "dacorvo/tiny-random-llama", "checkpoint_revision": "7fdafd2fe6a2d31c6abb72ae60db606d8bb23196", "compiler_type": "neuronx-cc", "compiler_version": "2.15.128.0+56dc5a86", "num_cores": 2, "sequence_length": 100, "task": "text-generation"}, "num_attention_heads": 1, "num_hidden_layers": 1, "num_key_value_heads": 1, "pretraining_tp": 1, "rms_norm_eps": 1e-06, "rope_scaling": null, "rope_theta": 10000.0, "tie_word_embeddings": false, "torch_dtype": "float32", "use_cache": true, "vocab_size": 32000}
|
neuronxcc-2.15.128.0+56dc5a86/0_REGISTRY/0.0.25.dev0/inference/llama/dacorvo/tiny-random-llama/e7337f4af7481c9d827e.json
ADDED
@@ -0,0 +1 @@
|
|
|
|
|
1 |
+
{"architectures": ["LlamaForCausalLM"], "attention_bias": false, "attention_dropout": 0.0, "bos_token_id": 1, "eos_token_id": 2, "hidden_act": "silu", "hidden_size": 128, "initializer_range": 0.02, "intermediate_size": 256, "max_position_embeddings": 512, "mlp_bias": false, "model_type": "llama", "neuron": {"auto_cast_type": "fp32", "batch_size": 1, "checkpoint_id": "dacorvo/tiny-random-llama", "checkpoint_revision": "7fdafd2fe6a2d31c6abb72ae60db606d8bb23196", "compiler_type": "neuronx-cc", "compiler_version": "2.15.128.0+56dc5a86", "num_cores": 2, "sequence_length": 100, "task": "text-generation"}, "num_attention_heads": 1, "num_hidden_layers": 1, "num_key_value_heads": 1, "pretraining_tp": 1, "rms_norm_eps": 1e-06, "rope_scaling": null, "rope_theta": 10000.0, "tie_word_embeddings": false, "torch_dtype": "float32", "use_cache": true, "vocab_size": 32000}
|
neuronxcc-2.15.128.0+56dc5a86/0_REGISTRY/0.0.25.dev0/inference/mistral/dacorvo/tiny-random-MistralForCausalLM/3bdcb3d6d5fb4b0b7097.json
ADDED
@@ -0,0 +1 @@
|
|
|
|
|
1 |
+
{"architectures": ["MistralForCausalLM"], "attention_dropout": 0.0, "attention_probs_dropout_prob": 0.1, "bos_token_id": 1, "eos_token_id": 2, "head_dim": 8, "hidden_act": "gelu", "hidden_dropout_prob": 0.1, "hidden_size": 32, "initializer_range": 0.02, "intermediate_size": 37, "is_decoder": true, "max_position_embeddings": 512, "model_type": "mistral", "neuron": {"auto_cast_type": "fp32", "batch_size": 1, "checkpoint_id": "dacorvo/tiny-random-MistralForCausalLM", "checkpoint_revision": "81d453e3c8985649e9ee3d4c9378461029d1c73a", "compiler_type": "neuronx-cc", "compiler_version": "2.15.128.0+56dc5a86", "num_cores": 2, "sequence_length": 100, "task": "text-generation"}, "num_attention_heads": 4, "num_hidden_layers": 2, "num_key_value_heads": 2, "pad_token_id": 0, "rms_norm_eps": 1e-06, "rope_theta": 10000.0, "sliding_window": 4096, "tie_word_embeddings": false, "torch_dtype": "float32", "type_vocab_size": 16, "use_cache": true, "vocab_size": 32000}
|
neuronxcc-2.15.128.0+56dc5a86/0_REGISTRY/0.0.25.dev0/inference/mistral/dacorvo/tiny-random-MistralForCausalLM/3d676d12f1566545d6b0.json
ADDED
@@ -0,0 +1 @@
|
|
|
|
|
1 |
+
{"architectures": ["MistralForCausalLM"], "attention_dropout": 0.0, "attention_probs_dropout_prob": 0.1, "bos_token_id": 1, "eos_token_id": 2, "head_dim": 8, "hidden_act": "gelu", "hidden_dropout_prob": 0.1, "hidden_size": 32, "initializer_range": 0.02, "intermediate_size": 37, "is_decoder": true, "max_position_embeddings": 512, "model_type": "mistral", "neuron": {"auto_cast_type": "fp16", "batch_size": 2, "checkpoint_id": "dacorvo/tiny-random-MistralForCausalLM", "checkpoint_revision": "81d453e3c8985649e9ee3d4c9378461029d1c73a", "compiler_type": "neuronx-cc", "compiler_version": "2.15.128.0+56dc5a86", "num_cores": 2, "sequence_length": 100, "task": "text-generation"}, "num_attention_heads": 4, "num_hidden_layers": 2, "num_key_value_heads": 2, "pad_token_id": 0, "rms_norm_eps": 1e-06, "rope_theta": 10000.0, "sliding_window": 4096, "tie_word_embeddings": false, "torch_dtype": "float32", "type_vocab_size": 16, "use_cache": true, "vocab_size": 32000}
|
neuronxcc-2.15.128.0+56dc5a86/0_REGISTRY/0.0.25.dev0/inference/mistral/dacorvo/tiny-random-MistralForCausalLM/ad4d8b389573be75fbee.json
ADDED
@@ -0,0 +1 @@
|
|
|
|
|
1 |
+
{"architectures": ["MistralForCausalLM"], "attention_dropout": 0.0, "attention_probs_dropout_prob": 0.1, "bos_token_id": 1, "eos_token_id": 2, "head_dim": 8, "hidden_act": "gelu", "hidden_dropout_prob": 0.1, "hidden_size": 32, "initializer_range": 0.02, "intermediate_size": 37, "is_decoder": true, "max_position_embeddings": 512, "model_type": "mistral", "neuron": {"auto_cast_type": "fp16", "batch_size": 1, "checkpoint_id": "dacorvo/tiny-random-MistralForCausalLM", "checkpoint_revision": "81d453e3c8985649e9ee3d4c9378461029d1c73a", "compiler_type": "neuronx-cc", "compiler_version": "2.15.128.0+56dc5a86", "num_cores": 2, "sequence_length": 100, "task": "text-generation"}, "num_attention_heads": 4, "num_hidden_layers": 2, "num_key_value_heads": 2, "pad_token_id": 0, "rms_norm_eps": 1e-06, "rope_theta": 10000.0, "sliding_window": 4096, "tie_word_embeddings": false, "torch_dtype": "float32", "type_vocab_size": 16, "use_cache": true, "vocab_size": 32000}
|
neuronxcc-2.15.128.0+56dc5a86/0_REGISTRY/0.0.25.dev0/inference/mistral/optimum/mistral-1.1b-testing/081b187113a5f417a9e0.json
ADDED
@@ -0,0 +1 @@
|
|
|
|
|
1 |
+
{"architectures": ["MistralForCausalLM"], "attention_dropout": 0.0, "bos_token_id": 1, "eos_token_id": 2, "head_dim": 64, "hidden_act": "silu", "hidden_size": 2048, "initializer_range": 0.02, "intermediate_size": 5632, "max_position_embeddings": 32768, "model_type": "mistral", "neuron": {"auto_cast_type": "bf16", "batch_size": 4, "checkpoint_id": "optimum/mistral-1.1b-testing", "checkpoint_revision": "ce03bc8d47dbd2c173ff65f3a8de1325ba724195", "compiler_type": "neuronx-cc", "compiler_version": "2.15.128.0+56dc5a86", "num_cores": 2, "sequence_length": 4096, "task": "text-generation"}, "num_attention_heads": 32, "num_hidden_layers": 22, "num_key_value_heads": 4, "rms_norm_eps": 1e-05, "rope_theta": 1000000.0, "sliding_window": null, "tie_word_embeddings": false, "torch_dtype": "bfloat16", "use_cache": true, "vocab_size": 32000}
|
neuronxcc-2.15.128.0+56dc5a86/0_REGISTRY/0.0.25.dev0/inference/mixtral/dacorvo/Mixtral-tiny/3a688b7dcd45d9a80b14.json
ADDED
@@ -0,0 +1 @@
|
|
|
|
|
1 |
+
{"architectures": ["MixtralForCausalLM"], "attention_dropout": 0.0, "bos_token_id": 1, "eos_token_id": 2, "hidden_act": "silu", "hidden_size": 1024, "initializer_range": 0.02, "intermediate_size": 3584, "max_position_embeddings": 1024, "model_type": "mixtral", "neuron": {"auto_cast_type": "fp32", "batch_size": 1, "checkpoint_id": "dacorvo/Mixtral-tiny", "checkpoint_revision": "c557ba205ddff6ea911f4719e0d543d6c08356b6", "compiler_type": "neuronx-cc", "compiler_version": "2.15.128.0+56dc5a86", "num_cores": 2, "sequence_length": 100, "task": "text-generation"}, "num_attention_heads": 32, "num_experts_per_tok": 2, "num_hidden_layers": 2, "num_key_value_heads": 8, "num_local_experts": 8, "output_router_logits": false, "rms_norm_eps": 1e-05, "rope_theta": 10000.0, "router_aux_loss_coef": 0.001, "router_jitter_noise": 0.0, "sliding_window": 4096, "tie_word_embeddings": false, "torch_dtype": "float16", "use_cache": true, "vocab_size": 32000}
|
neuronxcc-2.15.128.0+56dc5a86/0_REGISTRY/0.0.25.dev0/inference/mixtral/dacorvo/Mixtral-tiny/89980755d7c8bc1b31b0.json
ADDED
@@ -0,0 +1 @@
|
|
|
|
|
1 |
+
{"architectures": ["MixtralForCausalLM"], "attention_dropout": 0.0, "bos_token_id": 1, "eos_token_id": 2, "hidden_act": "silu", "hidden_size": 1024, "initializer_range": 0.02, "intermediate_size": 3584, "max_position_embeddings": 1024, "model_type": "mixtral", "neuron": {"auto_cast_type": "fp16", "batch_size": 4, "checkpoint_id": "dacorvo/Mixtral-tiny", "checkpoint_revision": "c557ba205ddff6ea911f4719e0d543d6c08356b6", "compiler_type": "neuronx-cc", "compiler_version": "2.15.128.0+56dc5a86", "num_cores": 2, "sequence_length": 1024, "task": "text-generation"}, "num_attention_heads": 32, "num_experts_per_tok": 2, "num_hidden_layers": 2, "num_key_value_heads": 8, "num_local_experts": 8, "output_router_logits": false, "rms_norm_eps": 1e-05, "rope_theta": 10000.0, "router_aux_loss_coef": 0.001, "router_jitter_noise": 0.0, "sliding_window": 4096, "tie_word_embeddings": false, "torch_dtype": "float16", "use_cache": true, "vocab_size": 32000}
|
neuronxcc-2.15.128.0+56dc5a86/0_REGISTRY/0.0.25.dev0/inference/mixtral/dacorvo/Mixtral-tiny/9f85ef35436cc6a20682.json
ADDED
@@ -0,0 +1 @@
|
|
|
|
|
1 |
+
{"architectures": ["MixtralForCausalLM"], "attention_dropout": 0.0, "bos_token_id": 1, "eos_token_id": 2, "hidden_act": "silu", "hidden_size": 1024, "initializer_range": 0.02, "intermediate_size": 3584, "max_position_embeddings": 1024, "model_type": "mixtral", "neuron": {"auto_cast_type": "fp16", "batch_size": 1, "checkpoint_id": "dacorvo/Mixtral-tiny", "checkpoint_revision": "c557ba205ddff6ea911f4719e0d543d6c08356b6", "compiler_type": "neuronx-cc", "compiler_version": "2.15.128.0+56dc5a86", "num_cores": 2, "sequence_length": 100, "task": "text-generation"}, "num_attention_heads": 32, "num_experts_per_tok": 2, "num_hidden_layers": 2, "num_key_value_heads": 8, "num_local_experts": 8, "output_router_logits": false, "rms_norm_eps": 1e-05, "rope_theta": 10000.0, "router_aux_loss_coef": 0.001, "router_jitter_noise": 0.0, "sliding_window": 4096, "tie_word_embeddings": false, "torch_dtype": "float16", "use_cache": true, "vocab_size": 32000}
|
neuronxcc-2.15.128.0+56dc5a86/0_REGISTRY/0.0.25.dev0/inference/mixtral/dacorvo/Mixtral-tiny/f1477b1f14a6669df1a7.json
ADDED
@@ -0,0 +1 @@
|
|
|
|
|
1 |
+
{"architectures": ["MixtralForCausalLM"], "attention_dropout": 0.0, "bos_token_id": 1, "eos_token_id": 2, "hidden_act": "silu", "hidden_size": 1024, "initializer_range": 0.02, "intermediate_size": 3584, "max_position_embeddings": 1024, "model_type": "mixtral", "neuron": {"auto_cast_type": "fp16", "batch_size": 2, "checkpoint_id": "dacorvo/Mixtral-tiny", "checkpoint_revision": "c557ba205ddff6ea911f4719e0d543d6c08356b6", "compiler_type": "neuronx-cc", "compiler_version": "2.15.128.0+56dc5a86", "num_cores": 2, "sequence_length": 100, "task": "text-generation"}, "num_attention_heads": 32, "num_experts_per_tok": 2, "num_hidden_layers": 2, "num_key_value_heads": 8, "num_local_experts": 8, "output_router_logits": false, "rms_norm_eps": 1e-05, "rope_theta": 10000.0, "router_aux_loss_coef": 0.001, "router_jitter_noise": 0.0, "sliding_window": 4096, "tie_word_embeddings": false, "torch_dtype": "float16", "use_cache": true, "vocab_size": 32000}
|
neuronxcc-2.15.128.0+56dc5a86/0_REGISTRY/0.0.25.dev0/inference/opt/hf-internal-testing/tiny-random-OPTForCausalLM/63ef91a527f73ba3b87e.json
ADDED
@@ -0,0 +1 @@
|
|
|
|
|
1 |
+
{"_remove_final_layer_norm": false, "activation_function": "relu", "architectures": ["OPTForCausalLM"], "attention_dropout": 0.1, "bos_token_id": 2, "do_layer_norm_before": true, "dropout": 0.1, "embed_dim": 16, "enable_bias": true, "eos_token_id": 2, "ffn_dim": 4, "hidden_size": 16, "init_std": 0.02, "is_decoder": true, "layer_norm_elementwise_affine": true, "layerdrop": 0.0, "max_position_embeddings": 100, "model_type": "opt", "neuron": {"auto_cast_type": "fp16", "batch_size": 1, "checkpoint_id": "hf-internal-testing/tiny-random-OPTForCausalLM", "checkpoint_revision": "190d1f4fc0011d2eaeaa05282e0fbd2445e4b11f", "compiler_type": "neuronx-cc", "compiler_version": "2.15.128.0+56dc5a86", "num_cores": 2, "sequence_length": 100, "task": "text-generation"}, "num_attention_heads": 4, "num_hidden_layers": 5, "pad_token_id": 1, "torch_dtype": "float32", "use_cache": true, "vocab_size": 50265, "word_embed_proj_dim": 16}
|
neuronxcc-2.15.128.0+56dc5a86/0_REGISTRY/0.0.25.dev0/inference/opt/hf-internal-testing/tiny-random-OPTForCausalLM/aa9fef08c5f79c9217fb.json
ADDED
@@ -0,0 +1 @@
|
|
|
|
|
1 |
+
{"_remove_final_layer_norm": false, "activation_function": "relu", "architectures": ["OPTForCausalLM"], "attention_dropout": 0.1, "bos_token_id": 2, "do_layer_norm_before": true, "dropout": 0.1, "embed_dim": 16, "enable_bias": true, "eos_token_id": 2, "ffn_dim": 4, "hidden_size": 16, "init_std": 0.02, "is_decoder": true, "layer_norm_elementwise_affine": true, "layerdrop": 0.0, "max_position_embeddings": 100, "model_type": "opt", "neuron": {"auto_cast_type": "fp32", "batch_size": 1, "checkpoint_id": "hf-internal-testing/tiny-random-OPTForCausalLM", "checkpoint_revision": "190d1f4fc0011d2eaeaa05282e0fbd2445e4b11f", "compiler_type": "neuronx-cc", "compiler_version": "2.15.128.0+56dc5a86", "num_cores": 2, "sequence_length": 100, "task": "text-generation"}, "num_attention_heads": 4, "num_hidden_layers": 5, "pad_token_id": 1, "torch_dtype": "float32", "use_cache": true, "vocab_size": 50265, "word_embed_proj_dim": 16}
|
neuronxcc-2.15.128.0+56dc5a86/0_REGISTRY/0.0.25.dev0/inference/opt/hf-internal-testing/tiny-random-OPTForCausalLM/fe9dc58f7d028523a0a6.json
ADDED
@@ -0,0 +1 @@
|
|
|
|
|
1 |
+
{"_remove_final_layer_norm": false, "activation_function": "relu", "architectures": ["OPTForCausalLM"], "attention_dropout": 0.1, "bos_token_id": 2, "do_layer_norm_before": true, "dropout": 0.1, "embed_dim": 16, "enable_bias": true, "eos_token_id": 2, "ffn_dim": 4, "hidden_size": 16, "init_std": 0.02, "is_decoder": true, "layer_norm_elementwise_affine": true, "layerdrop": 0.0, "max_position_embeddings": 100, "model_type": "opt", "neuron": {"auto_cast_type": "fp16", "batch_size": 2, "checkpoint_id": "hf-internal-testing/tiny-random-OPTForCausalLM", "checkpoint_revision": "190d1f4fc0011d2eaeaa05282e0fbd2445e4b11f", "compiler_type": "neuronx-cc", "compiler_version": "2.15.128.0+56dc5a86", "num_cores": 2, "sequence_length": 100, "task": "text-generation"}, "num_attention_heads": 4, "num_hidden_layers": 5, "pad_token_id": 1, "torch_dtype": "float32", "use_cache": true, "vocab_size": 50265, "word_embed_proj_dim": 16}
|
neuronxcc-2.15.128.0+56dc5a86/MODULE_00a895c500f614d5e448+39f12043/compile_flags.json
ADDED
@@ -0,0 +1 @@
|
|
|
|
|
1 |
+
["--model-type=transformer", "--auto-cast=none", "--execute-repetition=1"]
|
neuronxcc-2.15.128.0+56dc5a86/MODULE_00a895c500f614d5e448+39f12043/model.done
ADDED
File without changes
|
neuronxcc-2.15.128.0+56dc5a86/MODULE_00a895c500f614d5e448+39f12043/model.hlo_module.pb
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:2fd039f830e38474b35ffa7875e575823b49051252f489fa859a10027c0b0176
|
3 |
+
size 56998
|
neuronxcc-2.15.128.0+56dc5a86/MODULE_00a895c500f614d5e448+39f12043/model.neff
ADDED
Binary file (124 kB). View file
|
|
neuronxcc-2.15.128.0+56dc5a86/MODULE_0270f34f54c498cedeec+39f12043/compile_flags.json
ADDED
@@ -0,0 +1 @@
|
|
|
|
|
1 |
+
["--model-type=transformer", "--auto-cast=none", "--execute-repetition=1"]
|
neuronxcc-2.15.128.0+56dc5a86/MODULE_0270f34f54c498cedeec+39f12043/model.done
ADDED
File without changes
|
neuronxcc-2.15.128.0+56dc5a86/MODULE_0270f34f54c498cedeec+39f12043/model.hlo_module.pb
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:55232d1f8a7bf3216037d725f1d9f971826f44b3f0bfe700dba381587050e2ea
|
3 |
+
size 46555
|
neuronxcc-2.15.128.0+56dc5a86/MODULE_0270f34f54c498cedeec+39f12043/model.neff
ADDED
Binary file (400 kB). View file
|
|
neuronxcc-2.15.128.0+56dc5a86/MODULE_05d01a658506892e6bcf+39f12043/compile_flags.json
ADDED
@@ -0,0 +1 @@
|
|
|
|
|
1 |
+
["--model-type=transformer", "--auto-cast=none", "--execute-repetition=1"]
|
neuronxcc-2.15.128.0+56dc5a86/MODULE_05d01a658506892e6bcf+39f12043/model.done
ADDED
File without changes
|
neuronxcc-2.15.128.0+56dc5a86/MODULE_05d01a658506892e6bcf+39f12043/model.hlo_module.pb
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:19bad7c7a1aca39e7acb9d1644e06f34534612f3f9d8da135a7ac71b567941c4
|
3 |
+
size 142880
|
neuronxcc-2.15.128.0+56dc5a86/MODULE_05d01a658506892e6bcf+39f12043/model.neff
ADDED
Binary file (646 kB). View file
|
|
neuronxcc-2.15.128.0+56dc5a86/MODULE_0d7bde45dbf2d5c06852+39f12043/compile_flags.json
ADDED
@@ -0,0 +1 @@
|
|
|
|
|
1 |
+
["--model-type=transformer", "--auto-cast=none", "--execute-repetition=1"]
|
neuronxcc-2.15.128.0+56dc5a86/MODULE_0d7bde45dbf2d5c06852+39f12043/model.done
ADDED
File without changes
|
neuronxcc-2.15.128.0+56dc5a86/MODULE_0d7bde45dbf2d5c06852+39f12043/model.hlo_module.pb
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:e8a4ae1405062b3adda459c6aa4a11045fe5add4801765e60a77d687a3c26e78
|
3 |
+
size 46216
|
neuronxcc-2.15.128.0+56dc5a86/MODULE_0d7bde45dbf2d5c06852+39f12043/model.neff
ADDED
Binary file (441 kB). View file
|
|
neuronxcc-2.15.128.0+56dc5a86/MODULE_162afdb41ad92db9e04c+39f12043/compile_flags.json
ADDED
@@ -0,0 +1 @@
|
|
|
|
|
1 |
+
["--model-type=transformer", "--auto-cast=none", "--execute-repetition=1"]
|
neuronxcc-2.15.128.0+56dc5a86/MODULE_162afdb41ad92db9e04c+39f12043/model.done
ADDED
File without changes
|
neuronxcc-2.15.128.0+56dc5a86/MODULE_162afdb41ad92db9e04c+39f12043/model.hlo_module.pb
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:8fcaf6b7fc177ea5bc252882dc99c281dd23ccf44230a9b3df3fff23fc2ea424
|
3 |
+
size 22310
|
neuronxcc-2.15.128.0+56dc5a86/MODULE_162afdb41ad92db9e04c+39f12043/model.neff
ADDED
Binary file (144 kB). View file
|
|
neuronxcc-2.15.128.0+56dc5a86/MODULE_176378d23aabee0ca25a+39f12043/compile_flags.json
ADDED
@@ -0,0 +1 @@
|
|
|
|
|
1 |
+
["--model-type=transformer", "--auto-cast=none", "--execute-repetition=1"]
|
neuronxcc-2.15.128.0+56dc5a86/MODULE_176378d23aabee0ca25a+39f12043/model.done
ADDED
File without changes
|
neuronxcc-2.15.128.0+56dc5a86/MODULE_176378d23aabee0ca25a+39f12043/model.hlo_module.pb
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:33d1eefd32ab3d717d23c885b1837f80d73d1b8d40119924095e9d7b358c41b2
|
3 |
+
size 61713
|
neuronxcc-2.15.128.0+56dc5a86/MODULE_176378d23aabee0ca25a+39f12043/model.neff
ADDED
Binary file (155 kB). View file
|
|
neuronxcc-2.15.128.0+56dc5a86/MODULE_1bfeeb4391752eee11c4+aa032128/compile_flags.json
ADDED
@@ -0,0 +1 @@
|
|
|
|
|
1 |
+
["--model-type=transformer", "--auto-cast=none", "--execute-repetition=5"]
|