Text Generation
PEFT
English
llama
llama2
python
code
willyninja30's picture
Update README.md
cf5c9fc
metadata
library_name: peft
license: apache-2.0
language:
  - en
tags:
  - llama
  - llama2
  - python
  - code
pipeline_tag: text-generation
datasets:
  - iamtarun/python_code_instructions_18k_alpaca

Aria Code Light is based on llama 2 chat. Training procedure

Aria Code light is a finetuned llama 2 chat HF on python dataset with over 18.000 tokens of coding prompts and answers. Our goal was to create a model which can run on a single GPU with more language skills and better coding performance than the initial LLama 2 especially in Python. ..

GPU used for training : NVIDIA A100.

....

Timing: Less than 24 hours.

.....

Method : Lora + PEFT

Update following LLAMA CODE 7B release.

As Meta just released LLAMA CODE 7B, and even a LLAMA CODE PYTHON, which is trained on a larger python dataset than Aria Code light, we still believe Aria Code light has a more user friendly approach by adding coding skills to a chat model. It has been noticed by many community users that specialized models in Coding often loose "non-coding" and natural language performance. That being said,we encourage you to try both and use the model which fit your needs better,everything done for the open source community is always useful. Congratulations to Meta Team for achieving this new milestone in "Coding LLMS" area.

Contact

Support : contact@faradaylab.fr

The following bitsandbytes quantization config was used during training:

  • quant_method: bitsandbytes
  • load_in_8bit: True
  • load_in_4bit: False
  • llm_int8_threshold: 6.0
  • llm_int8_skip_modules: None
  • llm_int8_enable_fp32_cpu_offload: False
  • llm_int8_has_fp16_weight: False
  • bnb_4bit_quant_type: fp4
  • bnb_4bit_use_double_quant: False
  • bnb_4bit_compute_dtype: float32

Framework versions

  • PEFT 0.6.0.dev0