Upload 9 files
Browse files- sft_args.json +15 -14
- tokenizer_config.json +0 -4
sft_args.json
CHANGED
@@ -7,7 +7,7 @@
|
|
7 |
"additional_trainable_parameters": [],
|
8 |
"tuner_backend": "peft",
|
9 |
"template_type": "phi3-vl",
|
10 |
-
"output_dir": "D:\\_____NEW_NN\\LLM\\MiniCPM-V\\finetune\\output\\phi3-vision-128k-instruct\\
|
11 |
"add_output_dir_suffix": true,
|
12 |
"ddp_backend": null,
|
13 |
"ddp_find_unused_parameters": null,
|
@@ -22,7 +22,7 @@
|
|
22 |
],
|
23 |
"val_dataset": [],
|
24 |
"dataset_seed": 42,
|
25 |
-
"dataset_test_ratio": 0.
|
26 |
"use_loss_scale": false,
|
27 |
"system": null,
|
28 |
"max_length": 2048,
|
@@ -45,18 +45,18 @@
|
|
45 |
"bnb_4bit_use_double_quant": true,
|
46 |
"bnb_4bit_quant_storage": null,
|
47 |
"lora_target_modules": [
|
48 |
-
"
|
49 |
"img_projection.0",
|
|
|
|
|
|
|
|
|
50 |
"qkv_proj",
|
51 |
-
"down_proj",
|
52 |
"o_proj",
|
53 |
-
"fc1",
|
54 |
"q_proj",
|
55 |
-
"
|
56 |
-
"
|
57 |
-
"gate_up_proj"
|
58 |
-
"img_projection.2",
|
59 |
-
"v_proj"
|
60 |
],
|
61 |
"lora_rank": 128,
|
62 |
"lora_alpha": 128,
|
@@ -130,11 +130,11 @@
|
|
130 |
"lr_scheduler_type": "cosine",
|
131 |
"warmup_ratio": 0.05,
|
132 |
"eval_steps": 50,
|
133 |
-
"save_steps":
|
134 |
"save_only_model": false,
|
135 |
"save_total_limit": 8,
|
136 |
"logging_steps": 5,
|
137 |
-
"dataloader_num_workers":
|
138 |
"dataloader_pin_memory": true,
|
139 |
"dataloader_drop_last": false,
|
140 |
"push_to_hub": false,
|
@@ -149,7 +149,7 @@
|
|
149 |
"use_flash_attn": null,
|
150 |
"ignore_args_error": false,
|
151 |
"check_model_is_latest": true,
|
152 |
-
"logging_dir": "D:\\_____NEW_NN\\LLM\\MiniCPM-V\\finetune\\output\\phi3-vision-128k-instruct\\
|
153 |
"report_to": [
|
154 |
"tensorboard"
|
155 |
],
|
@@ -179,6 +179,7 @@
|
|
179 |
"fsdp_num": 1,
|
180 |
"per_device_train_batch_size": null,
|
181 |
"per_device_eval_batch_size": null,
|
|
|
182 |
"self_cognition_sample": 0,
|
183 |
"train_dataset_mix_ratio": 0.0,
|
184 |
"train_dataset_mix_ds": [
|
@@ -205,5 +206,5 @@
|
|
205 |
"load_in_4bit": false,
|
206 |
"load_in_8bit": false,
|
207 |
"train_sampler_random": true,
|
208 |
-
"training_args": "Seq2SeqTrainingArguments(output_dir='D:\\\\_____NEW_NN\\\\LLM\\\\MiniCPM-V\\\\finetune\\\\output\\\\phi3-vision-128k-instruct\\\\
|
209 |
}
|
|
|
7 |
"additional_trainable_parameters": [],
|
8 |
"tuner_backend": "peft",
|
9 |
"template_type": "phi3-vl",
|
10 |
+
"output_dir": "D:\\_____NEW_NN\\LLM\\MiniCPM-V\\finetune\\output\\phi3-vision-128k-instruct\\v0-20240629-080216",
|
11 |
"add_output_dir_suffix": true,
|
12 |
"ddp_backend": null,
|
13 |
"ddp_find_unused_parameters": null,
|
|
|
22 |
],
|
23 |
"val_dataset": [],
|
24 |
"dataset_seed": 42,
|
25 |
+
"dataset_test_ratio": 0.07,
|
26 |
"use_loss_scale": false,
|
27 |
"system": null,
|
28 |
"max_length": 2048,
|
|
|
45 |
"bnb_4bit_use_double_quant": true,
|
46 |
"bnb_4bit_quant_storage": null,
|
47 |
"lora_target_modules": [
|
48 |
+
"fc1",
|
49 |
"img_projection.0",
|
50 |
+
"img_projection.2",
|
51 |
+
"k_proj",
|
52 |
+
"out_proj",
|
53 |
+
"fc2",
|
54 |
"qkv_proj",
|
|
|
55 |
"o_proj",
|
|
|
56 |
"q_proj",
|
57 |
+
"down_proj",
|
58 |
+
"v_proj",
|
59 |
+
"gate_up_proj"
|
|
|
|
|
60 |
],
|
61 |
"lora_rank": 128,
|
62 |
"lora_alpha": 128,
|
|
|
130 |
"lr_scheduler_type": "cosine",
|
131 |
"warmup_ratio": 0.05,
|
132 |
"eval_steps": 50,
|
133 |
+
"save_steps": 300,
|
134 |
"save_only_model": false,
|
135 |
"save_total_limit": 8,
|
136 |
"logging_steps": 5,
|
137 |
+
"dataloader_num_workers": 0,
|
138 |
"dataloader_pin_memory": true,
|
139 |
"dataloader_drop_last": false,
|
140 |
"push_to_hub": false,
|
|
|
149 |
"use_flash_attn": null,
|
150 |
"ignore_args_error": false,
|
151 |
"check_model_is_latest": true,
|
152 |
+
"logging_dir": "D:\\_____NEW_NN\\LLM\\MiniCPM-V\\finetune\\output\\phi3-vision-128k-instruct\\v0-20240629-080216/runs",
|
153 |
"report_to": [
|
154 |
"tensorboard"
|
155 |
],
|
|
|
179 |
"fsdp_num": 1,
|
180 |
"per_device_train_batch_size": null,
|
181 |
"per_device_eval_batch_size": null,
|
182 |
+
"eval_strategy": null,
|
183 |
"self_cognition_sample": 0,
|
184 |
"train_dataset_mix_ratio": 0.0,
|
185 |
"train_dataset_mix_ds": [
|
|
|
206 |
"load_in_4bit": false,
|
207 |
"load_in_8bit": false,
|
208 |
"train_sampler_random": true,
|
209 |
+
"training_args": "Seq2SeqTrainingArguments(output_dir='D:\\\\_____NEW_NN\\\\LLM\\\\MiniCPM-V\\\\finetune\\\\output\\\\phi3-vision-128k-instruct\\\\v0-20240629-080216', overwrite_output_dir=False, do_train=False, do_eval=True, do_predict=False, eval_strategy=<IntervalStrategy.STEPS: 'steps'>, prediction_loss_only=False, per_device_train_batch_size=2, per_device_eval_batch_size=2, per_gpu_train_batch_size=None, per_gpu_eval_batch_size=None, gradient_accumulation_steps=2, eval_accumulation_steps=None, eval_delay=0, learning_rate=0.00014, weight_decay=0.1, adam_beta1=0.9, adam_beta2=0.95, adam_epsilon=1e-08, max_grad_norm=0.5, num_train_epochs=4, max_steps=-1, lr_scheduler_type=<SchedulerType.COSINE: 'cosine'>, lr_scheduler_kwargs={}, warmup_ratio=0.05, warmup_steps=0, log_level='passive', log_level_replica='warning', log_on_each_node=True, logging_dir='D:\\\\_____NEW_NN\\\\LLM\\\\MiniCPM-V\\\\finetune\\\\output\\\\phi3-vision-128k-instruct\\\\v0-20240629-080216/runs', logging_strategy=<IntervalStrategy.STEPS: 'steps'>, logging_first_step=True, logging_steps=5, logging_nan_inf_filter=True, save_strategy=<IntervalStrategy.STEPS: 'steps'>, save_steps=300, save_total_limit=8, save_safetensors=True, save_on_each_node=True, save_only_model=False, restore_callback_states_from_checkpoint=False, no_cuda=False, use_cpu=False, use_mps_device=False, seed=42, data_seed=None, jit_mode_eval=False, use_ipex=False, bf16=True, fp16=False, fp16_opt_level='O1', half_precision_backend='auto', bf16_full_eval=False, fp16_full_eval=False, tf32=None, local_rank=0, ddp_backend=None, tpu_num_cores=None, tpu_metrics_debug=False, debug=[], dataloader_drop_last=False, eval_steps=50, dataloader_num_workers=0, dataloader_prefetch_factor=None, past_index=-1, run_name='D:\\\\_____NEW_NN\\\\LLM\\\\MiniCPM-V\\\\finetune\\\\output\\\\phi3-vision-128k-instruct\\\\v0-20240629-080216', disable_tqdm=False, remove_unused_columns=False, label_names=None, load_best_model_at_end=False, metric_for_best_model='loss', greater_is_better=False, ignore_data_skip=False, fsdp=[], fsdp_min_num_params=0, fsdp_config={'min_num_params': 0, 'xla': False, 'xla_fsdp_v2': False, 'xla_fsdp_grad_ckpt': False}, fsdp_transformer_layer_cls_to_wrap=None, accelerator_config=AcceleratorConfig(split_batches=False, dispatch_batches=None, even_batches=True, use_seedable_sampler=True, non_blocking=False, gradient_accumulation_kwargs=None), deepspeed=None, label_smoothing_factor=0.0, optim=<OptimizerNames.ADAMW_TORCH: 'adamw_torch'>, optim_args=None, adafactor=False, group_by_length=False, length_column_name='length', report_to=['tensorboard'], ddp_find_unused_parameters=None, ddp_bucket_cap_mb=None, ddp_broadcast_buffers=None, dataloader_pin_memory=True, dataloader_persistent_workers=False, skip_memory_metrics=True, use_legacy_prediction_loop=False, push_to_hub=False, resume_from_checkpoint=None, hub_model_id=None, hub_strategy=<HubStrategy.EVERY_SAVE: 'every_save'>, hub_token=None, hub_private_repo=False, hub_always_push=False, gradient_checkpointing=True, gradient_checkpointing_kwargs=None, include_inputs_for_metrics=False, eval_do_concat_batches=True, fp16_backend='auto', evaluation_strategy=None, push_to_hub_model_id=None, push_to_hub_organization=None, push_to_hub_token=None, mp_parameters='', auto_find_batch_size=False, full_determinism=False, torchdynamo=None, ray_scope='last', ddp_timeout=1800, torch_compile=False, torch_compile_backend=None, torch_compile_mode=None, dispatch_batches=None, split_batches=None, include_tokens_per_second=False, include_num_input_tokens_seen=False, neftune_noise_alpha=None, optim_target_modules=None, batch_eval_metrics=False, sortish_sampler=True, predict_with_generate=False, generation_max_length=None, generation_num_beams=None, generation_config=GenerationConfig {\n \"do_sample\": true,\n \"eos_token_id\": 32000,\n \"max_new_tokens\": 2048,\n \"pad_token_id\": 32000,\n \"temperature\": 0.3,\n \"top_k\": 20,\n \"top_p\": 0.7\n}\n, train_sampler_random=True, push_hub_strategy='push_best', acc_strategy='token', additional_saved_files=[], metric_warmup_step=0, train_dataset_sample=1021)"
|
210 |
}
|
tokenizer_config.json
CHANGED
@@ -393,9 +393,6 @@
|
|
393 |
"<|user|>",
|
394 |
"<|end|>"
|
395 |
],
|
396 |
-
"auto_map": {
|
397 |
-
"AutoProcessor": "processing_phi3_v.Phi3VProcessor"
|
398 |
-
},
|
399 |
"bos_token": "<s>",
|
400 |
"chat_template": "{% for message in messages %}{{'<|' + message['role'] + '|>' + '\n' + message['content'] + '<|end|>\n' }}{% endfor %}{% if add_generation_prompt and messages[-1]['role'] != 'assistant' %}{{- '<|assistant|>\n' -}}{% endif %}",
|
401 |
"clean_up_tokenization_spaces": false,
|
@@ -404,7 +401,6 @@
|
|
404 |
"model_max_length": 131072,
|
405 |
"pad_token": "<|endoftext|>",
|
406 |
"padding_side": "right",
|
407 |
-
"processor_class": "Phi3VProcessor",
|
408 |
"sp_model_kwargs": {},
|
409 |
"tokenizer_class": "LlamaTokenizer",
|
410 |
"unk_token": "<unk>",
|
|
|
393 |
"<|user|>",
|
394 |
"<|end|>"
|
395 |
],
|
|
|
|
|
|
|
396 |
"bos_token": "<s>",
|
397 |
"chat_template": "{% for message in messages %}{{'<|' + message['role'] + '|>' + '\n' + message['content'] + '<|end|>\n' }}{% endfor %}{% if add_generation_prompt and messages[-1]['role'] != 'assistant' %}{{- '<|assistant|>\n' -}}{% endif %}",
|
398 |
"clean_up_tokenization_spaces": false,
|
|
|
401 |
"model_max_length": 131072,
|
402 |
"pad_token": "<|endoftext|>",
|
403 |
"padding_side": "right",
|
|
|
404 |
"sp_model_kwargs": {},
|
405 |
"tokenizer_class": "LlamaTokenizer",
|
406 |
"unk_token": "<unk>",
|