File size: 1,387 Bytes
3ec648b |
1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 16 17 18 19 20 21 22 23 24 25 26 27 28 29 30 31 32 33 34 35 36 37 38 39 40 41 42 43 44 45 46 47 48 49 50 |
model:
_component_: torchtune.models.llama3_2_vision.llama3_2_vision_11b
decoder_trainable: false
encoder_trainable: true
fusion_trainable: true
image_size: 560
tokenizer:
_component_: torchtune.models.llama3_2_vision.llama3_2_vision_transform
path: /tmp/Llama-3.2-11B-Vision-Instruct/original/tokenizer.model
image_size: 560
checkpointer:
_component_: torchtune.training.FullModelMetaCheckpointer
checkpoint_dir: /tmp/Llama-3.2-11B-Vision-Instruct/original/
checkpoint_files:
- consolidated.pth
recipe_checkpoint: null
output_dir: /tmp/Llama-3.2-11B-Vision-Instruct/
model_type: LLAMA3_VISION
resume_from_checkpoint: false
dataset:
_component_: data.chart_dataset
source: jrc/data-viz-qa
split: train
seed: 42
shuffle: true
collate_fn: torchtune.data.padded_collate_tiled_images_and_mask
epochs: 2
max_steps_per_epoch: null
batch_size: 8
gradient_accumulation_steps: 4
optimizer:
_component_: torch.optim.AdamW
lr: 2.0e-05
fused: true
loss:
_component_: torchtune.modules.loss.CEWithChunkedOutputLoss
clip_grad_norm: 1.0
compile: false
device: cuda
enable_activation_checkpointing: true
custom_sharded_layers: []
dtype: bf16
output_dir: /tmp/full-llama3.2-vision--finetune
metric_logger:
_component_: torchtune.training.metric_logging.WandBLogger
project: plot-huh
name: dataviz-qa-full
log_every_n_steps: 1
log_peak_memory_stats: true
|