File size: 968 Bytes
105f4ae
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
datasets:
- name: Scottie201/seq2seq_data
  split: train
- name: Scottie201/seq2seq_data
  split: validation
decision:
  continue_training: false
  reason: Model performance has not improved enough to warrant continued training.
  suggestions:
  - Consider adjusting the learning rate.
  - Review training data for any inconsistencies.
model:
  architecture: seq2seq
  base_model: facebook/bart-large
  batch_size: 4
  epochs: 3
  learning_rate: 3e-5
  output_dir: /workspaces/wakin_bake/trained_seq2seq
resources:
  device_map: auto
  torch_dtype: bfloat16
  use_gpu: true
tokenization:
  max_length: 512
  padding: max_length
  source_column: input_text
  target_column: target_text
  truncation: true
training_args:
  eval_strategy: epoch
  generation_max_length: 150
  generation_num_beams: 5
  hub_model_id: Scottie201/seq2seq_finetuned
  logging_dir: ./logs
  logging_steps: 1
  push_to_hub: true
  save_steps: 350
  save_total_limit: 1
training_continue: false