Saxo's picture
Update README.md
b23f0b8 verified
|
raw
history blame
1.37 kB
metadata
library_name: transformers
license: apache-2.0
basemodel: beomi/llama-2-koen-13b
datasets:
  - Saxo/total_ko_train_set_small_basic
  - beomi/KoAlpaca-v1.1a
  - kyujinpy/KOR-OpenOrca-Platypus-v2
  - nlpai-lab/databricks-dolly-15k-ko
language:
  - ko
  - en
pipeline_tag: text-generation

Model Card for Model ID

AI 와 빅데이터 분석 전문 기업인 Linkbricks의 데이터사이언티스트인 지윤성 박사(Saxo)가 beomi/llama-2-koen-13b 베이스모델을 GCP상의 A100-40G 4개를 통해 SFT 훈련을 한(2048 Tokens) 인스트럭션 모델. Accelerate, Deepspeed Zero-3 라이브러리를 사용했으며 Flash Attention 은 Disable 로 설정

Dr. Yunsung Ji (Saxo), a data scientist at Linkbricks, a company specializing in AI and big data analytics, trained the beomi/llama-2-koen-13b base model on 4 A100-40Gs on GCC for 4 hours of instructional training (2048 Tokens). Accelerate, Deepspeed Zero-3 libraries were used.

www.linkbricks.com, www.linkbricks.vc

Configuration including BitsandBytes


learning_rate = 2e-4 num_epochs = 5 batch_size = 4 block_size = 2048 trainer = "sft" warmup_ratio = 0.1 weight_decay = 0.01 gradient_accumulation = 4 mixed_precision = "fp16" peft = True quantization = "int4" lora_r = 64 lora_alpha = 16 lora_dropout = 0.1 model_max_length = 2048

Dataset Format

Alpaca Format Prompt Text