Edit model card
YAML Metadata Warning: empty or missing yaml metadata in repo card (https://huggingface.co/docs/hub/model-cards#model-card-metadata)

Quantization made by Richard Erkhov.

Github

Discord

Request more models

TinyLlama-1.1B-step-2T-lr-5-5ep-oasst1-top1-instruct-V1 - GGUF

Name Quant method Size
TinyLlama-1.1B-step-2T-lr-5-5ep-oasst1-top1-instruct-V1.Q2_K.gguf Q2_K 0.4GB
TinyLlama-1.1B-step-2T-lr-5-5ep-oasst1-top1-instruct-V1.IQ3_XS.gguf IQ3_XS 0.44GB
TinyLlama-1.1B-step-2T-lr-5-5ep-oasst1-top1-instruct-V1.IQ3_S.gguf IQ3_S 0.47GB
TinyLlama-1.1B-step-2T-lr-5-5ep-oasst1-top1-instruct-V1.Q3_K_S.gguf Q3_K_S 0.47GB
TinyLlama-1.1B-step-2T-lr-5-5ep-oasst1-top1-instruct-V1.IQ3_M.gguf IQ3_M 0.48GB
TinyLlama-1.1B-step-2T-lr-5-5ep-oasst1-top1-instruct-V1.Q3_K.gguf Q3_K 0.51GB
TinyLlama-1.1B-step-2T-lr-5-5ep-oasst1-top1-instruct-V1.Q3_K_M.gguf Q3_K_M 0.51GB
TinyLlama-1.1B-step-2T-lr-5-5ep-oasst1-top1-instruct-V1.Q3_K_L.gguf Q3_K_L 0.55GB
TinyLlama-1.1B-step-2T-lr-5-5ep-oasst1-top1-instruct-V1.IQ4_XS.gguf IQ4_XS 0.57GB
TinyLlama-1.1B-step-2T-lr-5-5ep-oasst1-top1-instruct-V1.Q4_0.gguf Q4_0 0.59GB
TinyLlama-1.1B-step-2T-lr-5-5ep-oasst1-top1-instruct-V1.IQ4_NL.gguf IQ4_NL 0.6GB
TinyLlama-1.1B-step-2T-lr-5-5ep-oasst1-top1-instruct-V1.Q4_K_S.gguf Q4_K_S 0.6GB
TinyLlama-1.1B-step-2T-lr-5-5ep-oasst1-top1-instruct-V1.Q4_K.gguf Q4_K 0.62GB
TinyLlama-1.1B-step-2T-lr-5-5ep-oasst1-top1-instruct-V1.Q4_K_M.gguf Q4_K_M 0.62GB
TinyLlama-1.1B-step-2T-lr-5-5ep-oasst1-top1-instruct-V1.Q4_1.gguf Q4_1 0.65GB
TinyLlama-1.1B-step-2T-lr-5-5ep-oasst1-top1-instruct-V1.Q5_0.gguf Q5_0 0.71GB
TinyLlama-1.1B-step-2T-lr-5-5ep-oasst1-top1-instruct-V1.Q5_K_S.gguf Q5_K_S 0.71GB
TinyLlama-1.1B-step-2T-lr-5-5ep-oasst1-top1-instruct-V1.Q5_K.gguf Q5_K 0.73GB
TinyLlama-1.1B-step-2T-lr-5-5ep-oasst1-top1-instruct-V1.Q5_K_M.gguf Q5_K_M 0.73GB
TinyLlama-1.1B-step-2T-lr-5-5ep-oasst1-top1-instruct-V1.Q5_1.gguf Q5_1 0.77GB
TinyLlama-1.1B-step-2T-lr-5-5ep-oasst1-top1-instruct-V1.Q6_K.gguf Q6_K 0.84GB
TinyLlama-1.1B-step-2T-lr-5-5ep-oasst1-top1-instruct-V1.Q8_0.gguf Q8_0 1.09GB

Original model description:

language: - en license: apache-2.0 datasets: - OpenAssistant/oasst_top1_2023-08-25 pipeline_tag: text-generation base_model: TinyLlama/TinyLlama-1.1B-intermediate-step-955k-token-2T model-index: - name: TinyLlama-1.1B-step-2T-lr-5-5ep-oasst1-top1-instruct-V1 results: - task: type: text-generation name: Text Generation dataset: name: AI2 Reasoning Challenge (25-Shot) type: ai2_arc config: ARC-Challenge split: test args: num_few_shot: 25 metrics: - type: acc_norm value: 31.06 name: normalized accuracy source: url: https://huggingface.co/spaces/HuggingFaceH4/open_llm_leaderboard?query=habanoz/TinyLlama-1.1B-step-2T-lr-5-5ep-oasst1-top1-instruct-V1 name: Open LLM Leaderboard - task: type: text-generation name: Text Generation dataset: name: HellaSwag (10-Shot) type: hellaswag split: validation args: num_few_shot: 10 metrics: - type: acc_norm value: 55.02 name: normalized accuracy source: url: https://huggingface.co/spaces/HuggingFaceH4/open_llm_leaderboard?query=habanoz/TinyLlama-1.1B-step-2T-lr-5-5ep-oasst1-top1-instruct-V1 name: Open LLM Leaderboard - task: type: text-generation name: Text Generation dataset: name: MMLU (5-Shot) type: cais/mmlu config: all split: test args: num_few_shot: 5 metrics: - type: acc value: 26.41 name: accuracy source: url: https://huggingface.co/spaces/HuggingFaceH4/open_llm_leaderboard?query=habanoz/TinyLlama-1.1B-step-2T-lr-5-5ep-oasst1-top1-instruct-V1 name: Open LLM Leaderboard - task: type: text-generation name: Text Generation dataset: name: TruthfulQA (0-shot) type: truthful_qa config: multiple_choice split: validation args: num_few_shot: 0 metrics: - type: mc2 value: 35.08 source: url: https://huggingface.co/spaces/HuggingFaceH4/open_llm_leaderboard?query=habanoz/TinyLlama-1.1B-step-2T-lr-5-5ep-oasst1-top1-instruct-V1 name: Open LLM Leaderboard - task: type: text-generation name: Text Generation dataset: name: Winogrande (5-shot) type: winogrande config: winogrande_xl split: validation args: num_few_shot: 5 metrics: - type: acc value: 58.01 name: accuracy source: url: https://huggingface.co/spaces/HuggingFaceH4/open_llm_leaderboard?query=habanoz/TinyLlama-1.1B-step-2T-lr-5-5ep-oasst1-top1-instruct-V1 name: Open LLM Leaderboard - task: type: text-generation name: Text Generation dataset: name: GSM8k (5-shot) type: gsm8k config: main split: test args: num_few_shot: 5 metrics: - type: acc value: 1.59 name: accuracy source: url: https://huggingface.co/spaces/HuggingFaceH4/open_llm_leaderboard?query=habanoz/TinyLlama-1.1B-step-2T-lr-5-5ep-oasst1-top1-instruct-V1 name: Open LLM Leaderboard

TinyLlama/TinyLlama-1.1B-intermediate-step-955k-token-2T finetuned using OpenAssistant/oasst_top1_2023-08-25 dataset.

Trained for 5 epochs using Qlora. Adapter is merged.

SFT code: https://github.com/habanoz/qlora.git

Command used:

accelerate launch $BASE_DIR/qlora/train.py \
  --model_name_or_path $BASE_MODEL \
  --working_dir $BASE_DIR/$OUTPUT_NAME-checkpoints \
  --output_dir $BASE_DIR/$OUTPUT_NAME-peft \
  --merged_output_dir $BASE_DIR/$OUTPUT_NAME \
  --final_output_dir $BASE_DIR/$OUTPUT_NAME-final \
  --num_train_epochs 5 \
  --logging_steps 1 \
  --save_strategy steps \
  --save_steps 75 \
  --save_total_limit 2 \
  --data_seed 11422 \
  --evaluation_strategy steps \
  --per_device_eval_batch_size 4 \
  --eval_dataset_size 0.01 \
  --eval_steps 75 \
  --max_new_tokens 1024 \
  --dataloader_num_workers 3 \
  --logging_strategy steps \
  --do_train \
  --do_eval \
  --lora_r 64 \
  --lora_alpha 16 \
  --lora_modules all \
  --bits 4 \
  --double_quant \
  --quant_type nf4 \
  --lr_scheduler_type constant \
  --dataset oasst1-top1 \
  --dataset_format oasst1 \
  --model_max_len 1024 \
  --per_device_train_batch_size 4 \
  --gradient_accumulation_steps 4 \
  --learning_rate 1e-5 \
  --adam_beta2 0.999 \
  --max_grad_norm 0.3 \
  --lora_dropout 0.0 \
  --weight_decay 0.0 \
  --seed 11422 \
  --gradient_checkpointing \
  --use_flash_attention_2 \
  --ddp_find_unused_parameters False

Open LLM Leaderboard Evaluation Results

Detailed results can be found here

Metric Value
Avg. 34.53
AI2 Reasoning Challenge (25-Shot) 31.06
HellaSwag (10-Shot) 55.02
MMLU (5-Shot) 26.41
TruthfulQA (0-shot) 35.08
Winogrande (5-shot) 58.01
GSM8k (5-shot) 1.59
Downloads last month
195
GGUF
Model size
1.1B params
Architecture
llama

2-bit

3-bit

4-bit

5-bit

6-bit

8-bit

Inference API
Unable to determine this model's library. Check the docs .