metadata
license: apache-2.0
datasets:
- hkust-nlp/deita-10k-v0
language:
- en
base_model: meta-llama/Llama-2-13b-hf
Model Card for Deita Llama2 13B V1.0 SFT
Deita is an open-sourced project designed to facilitate Automatic Data Selection for instruction tuning in Large Language Models (LLMs). Deita Llama2 13B V1.0 SFT is a fine-tuned version of Llama 2 that was trained on 10k automatically selected lightweight, high-quality alignment SFT data: Deita 10K V0.
Model description
- Model type: Model fine tuned on automatically selected lightweight, high-quality alignment SFT data.
- Language(s) (NLP): Primarily English
- Finetuned from model: meta-llama/Llama-2-13b-hf
Model Sources
- Repository: https://github.com/hkust-nlp/deita
- Model Family: Other models and the dataset are found in the Deita collection.
Performance
Input Format
The model is trained using the vicuna_v1.1 template
A chat between a curious user and an artificial intelligence assistant. The assistant gives helpful, detailed, and polite answers to the user's questions. USER: Hello! ASSISTANT: Hi!</s>USER: How are you? ASSISTANT:
Training hyperparameters
The following hyperparameters were used during fine tuning:
- learning_rate: 2e-05
- total_train_batch_size: 128
- optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
- lr_scheduler_type: linear
- lr_scheduler_warmup_ratio: 0.1
- num_epochs: 3.0