YAML Metadata Warning: empty or missing yaml metadata in repo card (https://huggingface.co/docs/hub/model-cards#model-card-metadata)

Llama-3-3B CodeSearchNet Fine-tuned

This repository hosts a Llama 3 (3B) model fine-tuned on the CodeSearchNet dataset, which contains code in six programming languages.

πŸ“ Model Details

  • Base Model: Llama 3 (3B)
  • Fine-tuning Dataset: CodeSearchNet
  • Languages Covered: Python, Java, JavaScript, PHP, Ruby, Go
  • Training Method: Supervised fine-tuning (SFT) with a contrastive loss objective for code search tasks
  • Tokenization: Llama 3 tokenizer with additional tokens for code-specific keywords
  • Frameworks Used: Hugging Face transformers, PyTorch, PEFT (for LoRA-based tuning)

πŸ“š Dataset

The model is trained on the CodeSearchNet dataset, which contains:

  • Function-level code snippets
  • Paired natural language descriptions
  • Multiple programming languages for multi-language search support

Dataset Sources

πŸš€ Training Setup

  • Hardware: NVIDIA A100 GPUs
  • Batch Size: 16
  • Learning Rate: 2e-5 with cosine annealing
  • Max Sequence Length: 512
  • Fine-tuning Duration: 3 epochs

πŸ” Intended Use

  • Code Search: Retrieve relevant code snippets given a natural language query
  • Code Completion: Provide context-aware code suggestions
  • Code-to-Text Generation: Explain code functionality in natural language
  • Multi-language Code Retrieval: Search across different programming languages
Downloads last month
6
Safetensors
Model size
3.21B params
Tensor type
F32
Β·
Inference Providers NEW
This model is not currently available via any of the supported third-party Inference Providers, and HF Inference API was unable to determine this model's library.