transmogrifier commited on
Commit
667be72
1 Parent(s): 277bc96

Upload cfg.yaml

Browse files
Files changed (1) hide show
  1. cfg.yaml +18 -18
cfg.yaml CHANGED
@@ -1,14 +1,14 @@
1
  architecture:
2
- backbone_dtype: int8
3
  force_embedding_gradients: true
4
  gradient_checkpointing: true
5
  intermediate_dropout: 0.0
6
  pretrained: true
7
- pretrained_weights: /media/akshay/datasets/largeModels/llms/h2o/h2o-llmstudio/output/user/economic-ferret.1/checkpoint.pth
8
  augmentation:
9
- random_parent_probability: 0.5
10
- skip_parent_probability: 0.0
11
- token_mask_probability: 0.0
12
  dataset:
13
  add_eos_token_to_answer: true
14
  add_eos_token_to_prompt: true
@@ -22,7 +22,7 @@ dataset:
22
  - Validation
23
  limit_chained_samples: false
24
  mask_prompt_labels: true
25
- parent_id_column: None
26
  personalize: false
27
  prompt_column:
28
  - instruction
@@ -30,7 +30,7 @@ dataset:
30
  text_answer_separator: <|answer|>
31
  text_prompt_start: <|prompt|>
32
  text_system_start: <|system|>
33
- train_dataframe: /media/akshay/datasets/largeModels/llms/h2o/h2o-llmstudio/data/user/PR-singleQA-July13/singleQA.csv
34
  validation_dataframe: None
35
  validation_size: 0.01
36
  validation_strategy: automatic
@@ -45,13 +45,13 @@ environment:
45
  seed: -1
46
  trust_remote_code: true
47
  use_fsdp: false
48
- experiment_name: economic-ferret.1.1
49
- llm_backbone: tiiuae/falcon-7b
50
  logging:
51
  logger: None
52
  neptune_project: ''
53
  number_of_texts: 10
54
- output_directory: /media/akshay/datasets/largeModels/llms/h2o/h2o-llmstudio/output/user/economic-ferret.1.1/
55
  prediction:
56
  batch_size_inference: 0
57
  do_sample: false
@@ -69,22 +69,22 @@ prediction:
69
  problem_type: text_causal_language_modeling
70
  tokenizer:
71
  add_prefix_space: false
72
- add_prompt_answer_tokens: false
73
- max_length: 1760
74
- max_length_answer: 512
75
- max_length_prompt: 1024
76
  padding_quantile: 1.0
77
  use_fast: true
78
  training:
79
  adaptive_kl_control: true
80
  advantages_gamma: 0.99
81
  advantages_lambda: 0.95
82
- batch_size: 2
83
  differential_learning_rate: 1.0e-05
84
  differential_learning_rate_layers: []
85
  drop_last_batch: true
86
- epochs: 3
87
- evaluate_before_training: true
88
  evaluation_epochs: 1.0
89
  grad_accumulation: 4
90
  gradient_clip: 0.9
@@ -96,7 +96,7 @@ training:
96
  lora_alpha: 16
97
  lora_dropout: 0.05
98
  lora_r: 8
99
- lora_target_modules: query_key_value, dense, dense_h_to_4h, dense_4h_to_h
100
  loss_function: TokenAveragedCrossEntropy
101
  offload_reward_model: false
102
  optimizer: AdamW
 
1
  architecture:
2
+ backbone_dtype: int4
3
  force_embedding_gradients: true
4
  gradient_checkpointing: true
5
  intermediate_dropout: 0.0
6
  pretrained: true
7
+ pretrained_weights: /media/akshay/datasets/largeModels/llms/h2o/h2o-llmstudio/output/user/raspberry-crocodile.1.2.1.1.1/checkpoint.pth
8
  augmentation:
9
+ random_parent_probability: 0.0
10
+ skip_parent_probability: 0.05
11
+ token_mask_probability: 0.3
12
  dataset:
13
  add_eos_token_to_answer: true
14
  add_eos_token_to_prompt: true
 
22
  - Validation
23
  limit_chained_samples: false
24
  mask_prompt_labels: true
25
+ parent_id_column: parent_id
26
  personalize: false
27
  prompt_column:
28
  - instruction
 
30
  text_answer_separator: <|answer|>
31
  text_prompt_start: <|prompt|>
32
  text_system_start: <|system|>
33
+ train_dataframe: /media/akshay/datasets/largeModels/llms/h2o/h2o-llmstudio/data/user/PR-Conversations-July21/conversations.csv
34
  validation_dataframe: None
35
  validation_size: 0.01
36
  validation_strategy: automatic
 
45
  seed: -1
46
  trust_remote_code: true
47
  use_fsdp: false
48
+ experiment_name: raspberry-crocodile-conv
49
+ llm_backbone: tiiuae/falcon-7b-instruct
50
  logging:
51
  logger: None
52
  neptune_project: ''
53
  number_of_texts: 10
54
+ output_directory: /media/akshay/datasets/largeModels/llms/h2o/h2o-llmstudio/output/user/raspberry-crocodile-conv/
55
  prediction:
56
  batch_size_inference: 0
57
  do_sample: false
 
69
  problem_type: text_causal_language_modeling
70
  tokenizer:
71
  add_prefix_space: false
72
+ add_prompt_answer_tokens: true
73
+ max_length: 4096
74
+ max_length_answer: 2048
75
+ max_length_prompt: 2048
76
  padding_quantile: 1.0
77
  use_fast: true
78
  training:
79
  adaptive_kl_control: true
80
  advantages_gamma: 0.99
81
  advantages_lambda: 0.95
82
+ batch_size: 3
83
  differential_learning_rate: 1.0e-05
84
  differential_learning_rate_layers: []
85
  drop_last_batch: true
86
+ epochs: 4
87
+ evaluate_before_training: false
88
  evaluation_epochs: 1.0
89
  grad_accumulation: 4
90
  gradient_clip: 0.9
 
96
  lora_alpha: 16
97
  lora_dropout: 0.05
98
  lora_r: 8
99
+ lora_target_modules: ''
100
  loss_function: TokenAveragedCrossEntropy
101
  offload_reward_model: false
102
  optimizer: AdamW