|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
model_hub: "HaNguyen/test_llama2" |
|
llama2_folder: recipes/MultiWOZ/response_generation/llama2/results/train_with_llama2/1995/save/llama2_checkpoint/ |
|
|
|
|
|
|
|
max_history: 2 |
|
|
|
|
|
freeze_model: True |
|
num_beams: 8 |
|
max_new_tokens: 50 |
|
top_k: 45 |
|
top_p: 0.9 |
|
|
|
|
|
model: !new:custom.LLAMA2_expanded |
|
source: !ref <model_hub> |
|
freeze: !ref <freeze_model> |
|
save_path: !ref <llama2_folder> |
|
max_new_tokens: !ref <max_new_tokens> |
|
num_beams: !ref <num_beams> |
|
top_k: !ref <top_k> |
|
top_p: !ref <top_p> |
|
with_peft: True |
|
|
|
|
|
|
|
padding_mask: !name:speechbrain.lobes.models.transformer.Transformer.get_key_padding_mask |
|
|
|
pretrainer: !new:speechbrain.utils.parameter_transfer.Pretrainer |
|
loadables: |
|
model: !ref <model> |
|
|
|
modules: |
|
model: !ref <model> |
|
|
|
|