datasets: | |
- name: Scottie201/seq2seq_data | |
split: train | |
- name: Scottie201/seq2seq_data | |
split: validation | |
decision: | |
continue_training: false | |
reason: Model performance has not improved enough to warrant continued training. | |
suggestions: | |
- Consider adjusting the learning rate. | |
- Review training data for any inconsistencies. | |
model: | |
architecture: seq2seq | |
base_model: facebook/bart-large | |
batch_size: 4 | |
epochs: 3 | |
learning_rate: 3e-5 | |
output_dir: /workspaces/wakin_bake/trained_seq2seq | |
resources: | |
device_map: auto | |
torch_dtype: bfloat16 | |
use_gpu: true | |
tokenization: | |
max_length: 512 | |
padding: max_length | |
source_column: input_text | |
target_column: target_text | |
truncation: true | |
training_args: | |
eval_strategy: epoch | |
generation_max_length: 150 | |
generation_num_beams: 5 | |
hub_model_id: Scottie201/seq2seq_finetuned | |
logging_dir: ./logs | |
logging_steps: 1 | |
push_to_hub: true | |
save_steps: 350 | |
save_total_limit: 1 | |
training_continue: false | |