|
dataset: |
|
align_stage_components: |
|
- download/llava-laion-cc-sbu-558k/chat.json |
|
- download/llava-laion-cc-sbu-558k |
|
dataset_id: llava-lrv |
|
dataset_root_dir: /data/projects/12003782/training_data |
|
finetune_stage_components: |
|
- download/llava-v1.5-instruct/llava_v1_5_lrv_mix1008k.json |
|
- download/llava-v1.5-instruct |
|
type: llava-lrv |
|
model: |
|
align_epochs: 1 |
|
align_global_batch_size: 64 |
|
align_learning_rate: 0.001 |
|
align_lr_scheduler_type: linear-warmup+cosine-decay |
|
align_max_grad_norm: 1.0 |
|
align_max_steps: null |
|
align_per_device_batch_size: 8 |
|
align_train_strategy: fsdp-shard-grad-op |
|
align_warmup_ratio: 0.03 |
|
align_weight_decay: 0.0 |
|
arch_specifier: fused-gelu-mlp |
|
enable_gradient_checkpointing: true |
|
enable_mixed_precision_training: true |
|
finetune_epochs: 1 |
|
finetune_global_batch_size: 32 |
|
finetune_learning_rate: 2.0e-05 |
|
finetune_lr_scheduler_type: linear-warmup+cosine-decay |
|
finetune_max_grad_norm: 1.0 |
|
finetune_max_steps: null |
|
finetune_per_device_batch_size: 4 |
|
finetune_train_strategy: fsdp-full-shard |
|
finetune_warmup_ratio: 0.03 |
|
finetune_weight_decay: 0.1 |
|
image_resize_strategy: resize-naive |
|
llm_backbone_id: phi3_base |
|
llm_max_length: 2048 |
|
model_id: dino-siglip-phi3-lora-model |
|
reduce_in_full_precision: false |
|
type: dino-siglip-phi3-lora-model |
|
vision_backbone_id: dinosiglip-vit-so-384px |
|
pretrained_checkpoint: null |
|
run_id: llava-lrv+dino-siglip-phi3-lora-model+stage-align+x7 |
|
run_root_dir: /data/projects/12003782/model_weights/runs |
|
seed: 7 |
|
stage: align |
|
trackers: |
|
- jsonl |
|
- wandb |
|
wandb_entity: null |
|
wandb_project: nscc-prismatic-phi3 |
|
|