| |
| |
|
|
| base_model: dineth554/legion-coder-8m |
| library_name: transformers |
| license: mit |
| pipeline_tag: text-generation |
| language: |
| - en |
| - code |
| tags: |
| - transformers |
| - pytorch |
| - safetensors |
| - text-generation |
| - code-generation |
| - python |
| - javascript |
| - coding |
| - programming |
| - sagemaker |
| - amazon-sagemaker |
| - cpu |
| - compact |
| - efficient |
| - nvdya-kit |
| - death-legion |
|
|
| datasets: |
| - the-stack-v2 |
|
|
| metrics: |
| - perplexity |
| - accuracy |
|
|
| model-index: |
| - name: Legion Coder 8M |
| results: [] |
|
|
| inference: |
| parameters: |
| temperature: 0.8 |
| top_p: 0.95 |
| top_k: 50 |
| max_new_tokens: 200 |
|
|
| sagemaker: |
| sdk_version: "2.200.0" |
| instance_type: "ml.m5.large" |
| instance_count: 1 |
| container_image: "huggingface-pytorch-inference:2.0.0-transformers4.28.1-cpu-py310-ubuntu20.04-v1.0" |
|
|
| |
| model_details: |
| name: Legion Coder 8M |
| version: 1.0.0 |
| description: A compact yet powerful 44M parameter transformer model optimized for coding tasks |
| developer: DEATH LEGION |
| powered_by: nvdya-kit |
| architecture: GPT-style Transformer |
| parameters: 44,341,632 |
| model_size: 170MB |
| hidden_size: 576 |
| num_layers: 13 |
| num_heads: 16 |
| context_length: 1024 |
| vocabulary_size: 16000 |
| format: Safetensors |
| precision: float32 |
|
|
| |
| training_details: |
| optimizer: AdamW |
| learning_rate: 5e-4 |
| lr_schedule: cosine_decay |
| batch_size: 4 |
| gradient_accumulation: true |
| training_steps: 10000 |
| precision: float32 |
|
|
| |
| intended_use: |
| primary_use_cases: |
| - Code completion and generation |
| - Function generation from descriptions |
| - Debugging assistance |
| - Code explanation and documentation |
| - Programming concept explanations |
| - Code scaffolding and prototyping |
| target_users: |
| - Software developers |
| - Students learning to code |
| - Data scientists |
| - DevOps engineers |
| - Technical writers |
|
|
| |
| limitations: |
| - Limited to 1,024 token context window |
| - Trained primarily on Python code |
| - May generate code that requires review before production use |
| - Not suitable for non-coding tasks |
|
|
| |
| ethical_considerations: |
| - Generated code should be reviewed before deployment |
| - May reproduce patterns from training data |
| - Not a replacement for human code review |
| - Users are responsible for compliance with licenses of generated code |
|
|
| |
| citation: | |
| @misc{legioncoder2026, |
| title={Legion Coder 8M: A Compact Transformer for Code Generation}, |
| author={DEATH LEGION}, |
| year={2026}, |
| howpublished={\url{https://huggingface.co/dineth554/legion-coder-8m}} |
| } |
|
|
| |
| contact: |
| developer: DEATH LEGION |
| powered_by: nvdya-kit |
| repository: https://huggingface.co/dineth554/legion-coder-8m |
|
|
| |
| branding: |
| tagline: MADE WITH BY DEATH LEGION |
| powered_by: nvdya-kit |
| copyright: 2026 DEATH LEGION. All rights reserved. |
|
|