--- license: wtfpl --- batch_size: 500 micro_batch_size: 24 num_epochs: 1 learning_rate: 0.0003 cutoff_len: 512 val_set_size: 0 lora_r: 8 lora_alpha: 16 lora_dropout: 0.05 lora_target_modules: ['q_proj', 'v_proj'] train_on_inputs: True add_eos_token: True group_by_length: False resume_from_checkpoint: False prompt template: alpaca trainable params: 4194304 || all params: 6742609920 || trainable%: 0.06220594176090199