Edit model card
YAML Metadata Warning: empty or missing yaml metadata in repo card (https://huggingface.co/docs/hub/model-cards#model-card-metadata)

base model = beomi-Llama-3-Open-Ko-8B-Instruct-preview base model = hansoldeco-beomi-Llama-3-Open-Ko-8B-Instruct-preview (Trained via Axolotl)

dora_train config (from fsdp_qlora repo)

export CUDA_VISIBLE_DEVICES=0,1
python train.py \
--train_type bnb_dora \
--model_name sosoai/hansoldeco-beomi-Llama-3-Open-Ko-8B-Instruct-preview \
--dataset orca_math \
--dataset_samples 193789 \
--batch_size 4 \
--context_length 8192 \
--gradient_accumulation_steps 2 \
--sharding_strategy full_shard \
--use_gradient_checkpointing true \
--reentrant_checkpointing true \
--use_cpu_offload false \
--use_activation_cpu_offload false \
--log_to wandb \
--project_name "sosoai-fsdp-quantized-ft-exps" \
--save_model true \
--output_dir models/llama-8b-orca-math-10k-bnb-QDoRA

Dataset = hansoldeco domain own dataset (Non open) Dataset = kuotient/orca-math-word-problems-193k-korean

Downloads last month
15
Safetensors
Model size
8.03B params
Tensor type
BF16
·