YanaS's picture
Upload quantized model versions of Photolens/llama-2-7b-langchain-chat
e95191e
metadata
language:
  - en
  - es
  - ru
  - de
  - pl
  - th
  - vi
  - sv
  - bn
  - da
  - he
  - it
  - fa
  - sk
  - id
  - nb
  - el
  - nl
  - hu
  - eu
  - zh
  - eo
  - ja
  - ca
  - cs
  - bg
  - fi
  - pt
  - tr
  - ro
  - ar
  - uk
  - gl
  - fr
  - ko
task_categories:
  - conversational
license: llama2
datasets:
  - Photolens/oasst1-langchain-llama-2-formatted

Model Overview

Model license: Llama-2
This model is trained based on NousResearch/Llama-2-7b-chat-hf model that is QLoRA finetuned on Photolens/oasst1-langchain-llama-2-formatted dataset.

Prompt Template: Llama-2

<s>[INST] Prompter Message [/INST] Assistant Message </s>

Intended Use

Dataset that is used to finetune base model is optimized for langchain applications.
So this model is intended for a langchain LLM.

Training Details

This model took 1:14:16 to train in QLoRA on a single A100 40gb GPU.

  • epochs: 1
  • train batch size: 8
  • eval batch size: 8
  • gradient accumulation steps: 1
  • maximum gradient normal: 0.3
  • learning rate: 2e-4
  • weight decay: 0.001
  • optimizer: paged_adamw_32bit
  • learning rate schedule: cosine
  • warmup ratio (linear): 0.03

Models in this series