Saxo's picture
Update README.md
b23f0b8 verified
|
raw
history blame
1.37 kB
---
library_name: transformers
license: apache-2.0
basemodel: beomi/llama-2-koen-13b
datasets:
- Saxo/total_ko_train_set_small_basic
- beomi/KoAlpaca-v1.1a
- kyujinpy/KOR-OpenOrca-Platypus-v2
- nlpai-lab/databricks-dolly-15k-ko
language:
- ko
- en
pipeline_tag: text-generation
---
# Model Card for Model ID
AI 와 빅데이터 분석 전문 기업인 Linkbricks의 데이터사이언티스트인 지윤성 박사(Saxo)가 beomi/llama-2-koen-13b 베이스모델을 GCP상의 A100-40G 4개를 통해 SFT 훈련을 한(2048 Tokens) 인스트럭션 모델.
Accelerate, Deepspeed Zero-3 라이브러리를 사용했으며 Flash Attention 은 Disable 로 설정
Dr. Yunsung Ji (Saxo), a data scientist at Linkbricks, a company specializing in AI and big data analytics, trained the beomi/llama-2-koen-13b base model on 4 A100-40Gs on GCC for 4 hours of instructional training (2048 Tokens).
Accelerate, Deepspeed Zero-3 libraries were used.
www.linkbricks.com, www.linkbricks.vc
## Configuration including BitsandBytes
---
learning_rate = 2e-4
num_epochs = 5
batch_size = 4
block_size = 2048
trainer = "sft"
warmup_ratio = 0.1
weight_decay = 0.01
gradient_accumulation = 4
mixed_precision = "fp16"
peft = True
quantization = "int4"
lora_r = 64
lora_alpha = 16
lora_dropout = 0.1
model_max_length = 2048
---
## Dataset Format
Alpaca Format Prompt Text