svjack's picture
Upload folder using huggingface_hub
3b24357 verified
raw
history blame contribute delete
No virus
25.6 kB
05/20/2024 15:00:49 - INFO - transformers.tokenization_utils_base - loading file vocab.json from cache at /home/featurize/.cache/huggingface/hub/models--Qwen--Qwen1.5-7B-Chat/snapshots/5f4f5e69ac7f1d508f8369e977de208b4803444b/vocab.json
05/20/2024 15:00:49 - INFO - transformers.tokenization_utils_base - loading file merges.txt from cache at /home/featurize/.cache/huggingface/hub/models--Qwen--Qwen1.5-7B-Chat/snapshots/5f4f5e69ac7f1d508f8369e977de208b4803444b/merges.txt
05/20/2024 15:00:49 - INFO - transformers.tokenization_utils_base - loading file tokenizer.json from cache at /home/featurize/.cache/huggingface/hub/models--Qwen--Qwen1.5-7B-Chat/snapshots/5f4f5e69ac7f1d508f8369e977de208b4803444b/tokenizer.json
05/20/2024 15:00:49 - INFO - transformers.tokenization_utils_base - loading file added_tokens.json from cache at None
05/20/2024 15:00:49 - INFO - transformers.tokenization_utils_base - loading file special_tokens_map.json from cache at None
05/20/2024 15:00:49 - INFO - transformers.tokenization_utils_base - loading file tokenizer_config.json from cache at /home/featurize/.cache/huggingface/hub/models--Qwen--Qwen1.5-7B-Chat/snapshots/5f4f5e69ac7f1d508f8369e977de208b4803444b/tokenizer_config.json
05/20/2024 15:00:49 - WARNING - transformers.tokenization_utils_base - Special tokens have been added in the vocabulary, make sure the associated word embeddings are fine-tuned or trained.
05/20/2024 15:00:49 - INFO - llamafactory.data.template - Replace eos token: <|im_end|>
05/20/2024 15:00:49 - INFO - llamafactory.data.loader - Loading dataset svjack/ShareGPT-Genshin-Impact-Human-Gpt-half...
05/20/2024 15:00:59 - INFO - transformers.configuration_utils - loading configuration file config.json from cache at /home/featurize/.cache/huggingface/hub/models--Qwen--Qwen1.5-7B-Chat/snapshots/5f4f5e69ac7f1d508f8369e977de208b4803444b/config.json
05/20/2024 15:00:59 - INFO - transformers.configuration_utils - Model config Qwen2Config {
"_name_or_path": "Qwen/Qwen1.5-7B-Chat",
"architectures": [
"Qwen2ForCausalLM"
],
"attention_dropout": 0.0,
"bos_token_id": 151643,
"eos_token_id": 151645,
"hidden_act": "silu",
"hidden_size": 4096,
"initializer_range": 0.02,
"intermediate_size": 11008,
"max_position_embeddings": 32768,
"max_window_layers": 28,
"model_type": "qwen2",
"num_attention_heads": 32,
"num_hidden_layers": 32,
"num_key_value_heads": 32,
"rms_norm_eps": 1e-06,
"rope_theta": 1000000.0,
"sliding_window": 32768,
"tie_word_embeddings": false,
"torch_dtype": "bfloat16",
"transformers_version": "4.41.0",
"use_cache": true,
"use_sliding_window": false,
"vocab_size": 151936
}
05/20/2024 15:00:59 - INFO - llamafactory.model.utils.quantization - Quantizing model to 4 bit.
05/20/2024 15:00:59 - INFO - transformers.modeling_utils - loading weights file model.safetensors from cache at /home/featurize/.cache/huggingface/hub/models--Qwen--Qwen1.5-7B-Chat/snapshots/5f4f5e69ac7f1d508f8369e977de208b4803444b/model.safetensors.index.json
05/20/2024 15:00:59 - INFO - transformers.modeling_utils - Instantiating Qwen2ForCausalLM model under default dtype torch.float16.
05/20/2024 15:00:59 - INFO - transformers.generation.configuration_utils - Generate config GenerationConfig {
"bos_token_id": 151643,
"eos_token_id": 151645
}
05/20/2024 15:01:11 - INFO - transformers.modeling_utils - All model checkpoint weights were used when initializing Qwen2ForCausalLM.
05/20/2024 15:01:11 - INFO - transformers.modeling_utils - All the weights of Qwen2ForCausalLM were initialized from the model checkpoint at Qwen/Qwen1.5-7B-Chat.
If your task is similar to the task the model of the checkpoint was trained on, you can already use Qwen2ForCausalLM for predictions without further training.
05/20/2024 15:01:11 - INFO - transformers.generation.configuration_utils - loading configuration file generation_config.json from cache at /home/featurize/.cache/huggingface/hub/models--Qwen--Qwen1.5-7B-Chat/snapshots/5f4f5e69ac7f1d508f8369e977de208b4803444b/generation_config.json
05/20/2024 15:01:11 - INFO - transformers.generation.configuration_utils - Generate config GenerationConfig {
"bos_token_id": 151643,
"do_sample": true,
"eos_token_id": [
151645,
151643
],
"pad_token_id": 151643,
"repetition_penalty": 1.05,
"temperature": 0.7,
"top_k": 20,
"top_p": 0.8
}
05/20/2024 15:01:11 - INFO - llamafactory.model.utils.checkpointing - Gradient checkpointing enabled.
05/20/2024 15:01:11 - INFO - llamafactory.model.utils.attention - Using torch SDPA for faster training and inference.
05/20/2024 15:01:11 - INFO - llamafactory.model.adapter - Upcasting trainable params to float32.
05/20/2024 15:01:11 - INFO - llamafactory.model.adapter - Fine-tuning method: LoRA
05/20/2024 15:01:12 - INFO - llamafactory.model.loader - trainable params: 4194304 || all params: 7725518848 || trainable%: 0.0543
05/20/2024 15:01:12 - WARNING - accelerate.utils.other - Detected kernel version 5.4.0, which is below the recommended minimum of 5.5.0; this can cause the process to hang. It is recommended to upgrade the kernel to the minimum version or higher.
05/20/2024 15:01:12 - INFO - transformers.trainer - Using auto half precision backend
05/20/2024 15:01:12 - INFO - transformers.trainer - ***** Running training *****
05/20/2024 15:01:12 - INFO - transformers.trainer - Num examples = 1,258
05/20/2024 15:01:12 - INFO - transformers.trainer - Num Epochs = 3
05/20/2024 15:01:12 - INFO - transformers.trainer - Instantaneous batch size per device = 1
05/20/2024 15:01:12 - INFO - transformers.trainer - Total train batch size (w. parallel, distributed & accumulation) = 8
05/20/2024 15:01:12 - INFO - transformers.trainer - Gradient Accumulation steps = 8
05/20/2024 15:01:12 - INFO - transformers.trainer - Total optimization steps = 471
05/20/2024 15:01:12 - INFO - transformers.trainer - Number of trainable parameters = 4,194,304
05/20/2024 15:02:34 - INFO - llamafactory.extras.callbacks - {'loss': 2.9252, 'learning_rate': 4.9986e-05, 'epoch': 0.03}
05/20/2024 15:03:56 - INFO - llamafactory.extras.callbacks - {'loss': 2.4825, 'learning_rate': 4.9944e-05, 'epoch': 0.06}
05/20/2024 15:05:17 - INFO - llamafactory.extras.callbacks - {'loss': 2.2139, 'learning_rate': 4.9875e-05, 'epoch': 0.10}
05/20/2024 15:06:39 - INFO - llamafactory.extras.callbacks - {'loss': 2.0351, 'learning_rate': 4.9778e-05, 'epoch': 0.13}
05/20/2024 15:07:59 - INFO - llamafactory.extras.callbacks - {'loss': 2.1238, 'learning_rate': 4.9653e-05, 'epoch': 0.16}
05/20/2024 15:09:20 - INFO - llamafactory.extras.callbacks - {'loss': 1.9683, 'learning_rate': 4.9501e-05, 'epoch': 0.19}
05/20/2024 15:10:41 - INFO - llamafactory.extras.callbacks - {'loss': 1.9120, 'learning_rate': 4.9322e-05, 'epoch': 0.22}
05/20/2024 15:12:03 - INFO - llamafactory.extras.callbacks - {'loss': 1.9160, 'learning_rate': 4.9115e-05, 'epoch': 0.25}
05/20/2024 15:13:24 - INFO - llamafactory.extras.callbacks - {'loss': 1.8891, 'learning_rate': 4.8882e-05, 'epoch': 0.29}
05/20/2024 15:14:46 - INFO - llamafactory.extras.callbacks - {'loss': 1.9039, 'learning_rate': 4.8623e-05, 'epoch': 0.32}
05/20/2024 15:16:05 - INFO - llamafactory.extras.callbacks - {'loss': 1.8915, 'learning_rate': 4.8337e-05, 'epoch': 0.35}
05/20/2024 15:17:27 - INFO - llamafactory.extras.callbacks - {'loss': 1.8870, 'learning_rate': 4.8025e-05, 'epoch': 0.38}
05/20/2024 15:18:48 - INFO - llamafactory.extras.callbacks - {'loss': 1.9291, 'learning_rate': 4.7687e-05, 'epoch': 0.41}
05/20/2024 15:20:07 - INFO - llamafactory.extras.callbacks - {'loss': 1.8363, 'learning_rate': 4.7324e-05, 'epoch': 0.45}
05/20/2024 15:21:26 - INFO - llamafactory.extras.callbacks - {'loss': 1.8624, 'learning_rate': 4.6937e-05, 'epoch': 0.48}
05/20/2024 15:22:48 - INFO - llamafactory.extras.callbacks - {'loss': 1.8554, 'learning_rate': 4.6524e-05, 'epoch': 0.51}
05/20/2024 15:24:10 - INFO - llamafactory.extras.callbacks - {'loss': 1.8476, 'learning_rate': 4.6089e-05, 'epoch': 0.54}
05/20/2024 15:25:31 - INFO - llamafactory.extras.callbacks - {'loss': 1.7998, 'learning_rate': 4.5629e-05, 'epoch': 0.57}
05/20/2024 15:26:51 - INFO - llamafactory.extras.callbacks - {'loss': 1.9258, 'learning_rate': 4.5147e-05, 'epoch': 0.60}
05/20/2024 15:28:11 - INFO - llamafactory.extras.callbacks - {'loss': 1.7941, 'learning_rate': 4.4642e-05, 'epoch': 0.64}
05/20/2024 15:28:11 - INFO - transformers.trainer - Saving model checkpoint to saves/Qwen1.5-7B-Chat/lora/train_2024-05-20-14-55-51/checkpoint-100
05/20/2024 15:28:12 - INFO - transformers.configuration_utils - loading configuration file config.json from cache at /home/featurize/.cache/huggingface/hub/models--Qwen--Qwen1.5-7B-Chat/snapshots/5f4f5e69ac7f1d508f8369e977de208b4803444b/config.json
05/20/2024 15:28:12 - INFO - transformers.configuration_utils - Model config Qwen2Config {
"architectures": [
"Qwen2ForCausalLM"
],
"attention_dropout": 0.0,
"bos_token_id": 151643,
"eos_token_id": 151645,
"hidden_act": "silu",
"hidden_size": 4096,
"initializer_range": 0.02,
"intermediate_size": 11008,
"max_position_embeddings": 32768,
"max_window_layers": 28,
"model_type": "qwen2",
"num_attention_heads": 32,
"num_hidden_layers": 32,
"num_key_value_heads": 32,
"rms_norm_eps": 1e-06,
"rope_theta": 1000000.0,
"sliding_window": 32768,
"tie_word_embeddings": false,
"torch_dtype": "bfloat16",
"transformers_version": "4.41.0",
"use_cache": true,
"use_sliding_window": false,
"vocab_size": 151936
}
05/20/2024 15:28:12 - INFO - transformers.tokenization_utils_base - tokenizer config file saved in saves/Qwen1.5-7B-Chat/lora/train_2024-05-20-14-55-51/checkpoint-100/tokenizer_config.json
05/20/2024 15:28:12 - INFO - transformers.tokenization_utils_base - Special tokens file saved in saves/Qwen1.5-7B-Chat/lora/train_2024-05-20-14-55-51/checkpoint-100/special_tokens_map.json
05/20/2024 15:29:33 - INFO - llamafactory.extras.callbacks - {'loss': 1.8889, 'learning_rate': 4.4115e-05, 'epoch': 0.67}
05/20/2024 15:30:53 - INFO - llamafactory.extras.callbacks - {'loss': 1.8287, 'learning_rate': 4.3567e-05, 'epoch': 0.70}
05/20/2024 15:32:13 - INFO - llamafactory.extras.callbacks - {'loss': 1.7811, 'learning_rate': 4.2999e-05, 'epoch': 0.73}
05/20/2024 15:33:31 - INFO - llamafactory.extras.callbacks - {'loss': 1.7925, 'learning_rate': 4.2410e-05, 'epoch': 0.76}
05/20/2024 15:34:52 - INFO - llamafactory.extras.callbacks - {'loss': 1.8646, 'learning_rate': 4.1802e-05, 'epoch': 0.79}
05/20/2024 15:36:12 - INFO - llamafactory.extras.callbacks - {'loss': 1.8213, 'learning_rate': 4.1176e-05, 'epoch': 0.83}
05/20/2024 15:37:35 - INFO - llamafactory.extras.callbacks - {'loss': 1.8301, 'learning_rate': 4.0531e-05, 'epoch': 0.86}
05/20/2024 15:38:54 - INFO - llamafactory.extras.callbacks - {'loss': 1.8081, 'learning_rate': 3.9869e-05, 'epoch': 0.89}
05/20/2024 15:40:13 - INFO - llamafactory.extras.callbacks - {'loss': 1.8472, 'learning_rate': 3.9191e-05, 'epoch': 0.92}
05/20/2024 15:41:34 - INFO - llamafactory.extras.callbacks - {'loss': 1.8186, 'learning_rate': 3.8497e-05, 'epoch': 0.95}
05/20/2024 15:42:53 - INFO - llamafactory.extras.callbacks - {'loss': 1.8542, 'learning_rate': 3.7788e-05, 'epoch': 0.99}
05/20/2024 15:44:13 - INFO - llamafactory.extras.callbacks - {'loss': 1.6997, 'learning_rate': 3.7064e-05, 'epoch': 1.02}
05/20/2024 15:45:34 - INFO - llamafactory.extras.callbacks - {'loss': 1.8274, 'learning_rate': 3.6327e-05, 'epoch': 1.05}
05/20/2024 15:46:55 - INFO - llamafactory.extras.callbacks - {'loss': 1.7660, 'learning_rate': 3.5578e-05, 'epoch': 1.08}
05/20/2024 15:48:15 - INFO - llamafactory.extras.callbacks - {'loss': 1.7501, 'learning_rate': 3.4817e-05, 'epoch': 1.11}
05/20/2024 15:49:37 - INFO - llamafactory.extras.callbacks - {'loss': 1.7754, 'learning_rate': 3.4045e-05, 'epoch': 1.14}
05/20/2024 15:50:54 - INFO - llamafactory.extras.callbacks - {'loss': 1.8031, 'learning_rate': 3.3263e-05, 'epoch': 1.18}
05/20/2024 15:52:14 - INFO - llamafactory.extras.callbacks - {'loss': 1.7226, 'learning_rate': 3.2471e-05, 'epoch': 1.21}
05/20/2024 15:53:34 - INFO - llamafactory.extras.callbacks - {'loss': 1.7752, 'learning_rate': 3.1672e-05, 'epoch': 1.24}
05/20/2024 15:54:55 - INFO - llamafactory.extras.callbacks - {'loss': 1.7353, 'learning_rate': 3.0865e-05, 'epoch': 1.27}
05/20/2024 15:54:55 - INFO - transformers.trainer - Saving model checkpoint to saves/Qwen1.5-7B-Chat/lora/train_2024-05-20-14-55-51/checkpoint-200
05/20/2024 15:54:56 - INFO - transformers.configuration_utils - loading configuration file config.json from cache at /home/featurize/.cache/huggingface/hub/models--Qwen--Qwen1.5-7B-Chat/snapshots/5f4f5e69ac7f1d508f8369e977de208b4803444b/config.json
05/20/2024 15:54:56 - INFO - transformers.configuration_utils - Model config Qwen2Config {
"architectures": [
"Qwen2ForCausalLM"
],
"attention_dropout": 0.0,
"bos_token_id": 151643,
"eos_token_id": 151645,
"hidden_act": "silu",
"hidden_size": 4096,
"initializer_range": 0.02,
"intermediate_size": 11008,
"max_position_embeddings": 32768,
"max_window_layers": 28,
"model_type": "qwen2",
"num_attention_heads": 32,
"num_hidden_layers": 32,
"num_key_value_heads": 32,
"rms_norm_eps": 1e-06,
"rope_theta": 1000000.0,
"sliding_window": 32768,
"tie_word_embeddings": false,
"torch_dtype": "bfloat16",
"transformers_version": "4.41.0",
"use_cache": true,
"use_sliding_window": false,
"vocab_size": 151936
}
05/20/2024 15:54:56 - INFO - transformers.tokenization_utils_base - tokenizer config file saved in saves/Qwen1.5-7B-Chat/lora/train_2024-05-20-14-55-51/checkpoint-200/tokenizer_config.json
05/20/2024 15:54:56 - INFO - transformers.tokenization_utils_base - Special tokens file saved in saves/Qwen1.5-7B-Chat/lora/train_2024-05-20-14-55-51/checkpoint-200/special_tokens_map.json
05/20/2024 15:56:19 - INFO - llamafactory.extras.callbacks - {'loss': 1.7834, 'learning_rate': 3.0051e-05, 'epoch': 1.30}
05/20/2024 15:57:40 - INFO - llamafactory.extras.callbacks - {'loss': 1.7668, 'learning_rate': 2.9232e-05, 'epoch': 1.34}
05/20/2024 15:59:02 - INFO - llamafactory.extras.callbacks - {'loss': 1.7912, 'learning_rate': 2.8408e-05, 'epoch': 1.37}
05/20/2024 16:00:24 - INFO - llamafactory.extras.callbacks - {'loss': 1.7674, 'learning_rate': 2.7580e-05, 'epoch': 1.40}
05/20/2024 16:01:48 - INFO - llamafactory.extras.callbacks - {'loss': 1.7827, 'learning_rate': 2.6749e-05, 'epoch': 1.43}
05/20/2024 16:03:10 - INFO - llamafactory.extras.callbacks - {'loss': 1.8176, 'learning_rate': 2.5917e-05, 'epoch': 1.46}
05/20/2024 16:04:32 - INFO - llamafactory.extras.callbacks - {'loss': 1.7613, 'learning_rate': 2.5083e-05, 'epoch': 1.49}
05/20/2024 16:05:52 - INFO - llamafactory.extras.callbacks - {'loss': 1.7506, 'learning_rate': 2.4250e-05, 'epoch': 1.53}
05/20/2024 16:07:13 - INFO - llamafactory.extras.callbacks - {'loss': 1.7380, 'learning_rate': 2.3417e-05, 'epoch': 1.56}
05/20/2024 16:08:35 - INFO - llamafactory.extras.callbacks - {'loss': 1.8030, 'learning_rate': 2.2586e-05, 'epoch': 1.59}
05/20/2024 16:09:56 - INFO - llamafactory.extras.callbacks - {'loss': 1.7351, 'learning_rate': 2.1758e-05, 'epoch': 1.62}
05/20/2024 16:11:19 - INFO - llamafactory.extras.callbacks - {'loss': 1.7596, 'learning_rate': 2.0933e-05, 'epoch': 1.65}
05/20/2024 16:12:40 - INFO - llamafactory.extras.callbacks - {'loss': 1.7875, 'learning_rate': 2.0113e-05, 'epoch': 1.69}
05/20/2024 16:14:02 - INFO - llamafactory.extras.callbacks - {'loss': 1.7871, 'learning_rate': 1.9298e-05, 'epoch': 1.72}
05/20/2024 16:15:24 - INFO - llamafactory.extras.callbacks - {'loss': 1.7658, 'learning_rate': 1.8489e-05, 'epoch': 1.75}
05/20/2024 16:16:46 - INFO - llamafactory.extras.callbacks - {'loss': 1.7510, 'learning_rate': 1.7688e-05, 'epoch': 1.78}
05/20/2024 16:18:05 - INFO - llamafactory.extras.callbacks - {'loss': 1.7664, 'learning_rate': 1.6895e-05, 'epoch': 1.81}
05/20/2024 16:19:26 - INFO - llamafactory.extras.callbacks - {'loss': 1.8125, 'learning_rate': 1.6111e-05, 'epoch': 1.84}
05/20/2024 16:20:47 - INFO - llamafactory.extras.callbacks - {'loss': 1.7941, 'learning_rate': 1.5337e-05, 'epoch': 1.88}
05/20/2024 16:22:09 - INFO - llamafactory.extras.callbacks - {'loss': 1.7435, 'learning_rate': 1.4573e-05, 'epoch': 1.91}
05/20/2024 16:22:09 - INFO - transformers.trainer - Saving model checkpoint to saves/Qwen1.5-7B-Chat/lora/train_2024-05-20-14-55-51/checkpoint-300
05/20/2024 16:22:10 - INFO - transformers.configuration_utils - loading configuration file config.json from cache at /home/featurize/.cache/huggingface/hub/models--Qwen--Qwen1.5-7B-Chat/snapshots/5f4f5e69ac7f1d508f8369e977de208b4803444b/config.json
05/20/2024 16:22:10 - INFO - transformers.configuration_utils - Model config Qwen2Config {
"architectures": [
"Qwen2ForCausalLM"
],
"attention_dropout": 0.0,
"bos_token_id": 151643,
"eos_token_id": 151645,
"hidden_act": "silu",
"hidden_size": 4096,
"initializer_range": 0.02,
"intermediate_size": 11008,
"max_position_embeddings": 32768,
"max_window_layers": 28,
"model_type": "qwen2",
"num_attention_heads": 32,
"num_hidden_layers": 32,
"num_key_value_heads": 32,
"rms_norm_eps": 1e-06,
"rope_theta": 1000000.0,
"sliding_window": 32768,
"tie_word_embeddings": false,
"torch_dtype": "bfloat16",
"transformers_version": "4.41.0",
"use_cache": true,
"use_sliding_window": false,
"vocab_size": 151936
}
05/20/2024 16:22:10 - INFO - transformers.tokenization_utils_base - tokenizer config file saved in saves/Qwen1.5-7B-Chat/lora/train_2024-05-20-14-55-51/checkpoint-300/tokenizer_config.json
05/20/2024 16:22:10 - INFO - transformers.tokenization_utils_base - Special tokens file saved in saves/Qwen1.5-7B-Chat/lora/train_2024-05-20-14-55-51/checkpoint-300/special_tokens_map.json
05/20/2024 16:23:32 - INFO - llamafactory.extras.callbacks - {'loss': 1.7548, 'learning_rate': 1.3821e-05, 'epoch': 1.94}
05/20/2024 16:24:54 - INFO - llamafactory.extras.callbacks - {'loss': 1.8122, 'learning_rate': 1.3082e-05, 'epoch': 1.97}
05/20/2024 16:26:15 - INFO - llamafactory.extras.callbacks - {'loss': 1.7795, 'learning_rate': 1.2356e-05, 'epoch': 2.00}
05/20/2024 16:27:36 - INFO - llamafactory.extras.callbacks - {'loss': 1.7171, 'learning_rate': 1.1644e-05, 'epoch': 2.03}
05/20/2024 16:28:57 - INFO - llamafactory.extras.callbacks - {'loss': 1.6727, 'learning_rate': 1.0947e-05, 'epoch': 2.07}
05/20/2024 16:30:18 - INFO - llamafactory.extras.callbacks - {'loss': 1.8067, 'learning_rate': 1.0265e-05, 'epoch': 2.10}
05/20/2024 16:31:40 - INFO - llamafactory.extras.callbacks - {'loss': 1.7300, 'learning_rate': 9.5997e-06, 'epoch': 2.13}
05/20/2024 16:32:59 - INFO - llamafactory.extras.callbacks - {'loss': 1.7939, 'learning_rate': 8.9516e-06, 'epoch': 2.16}
05/20/2024 16:34:21 - INFO - llamafactory.extras.callbacks - {'loss': 1.7272, 'learning_rate': 8.3214e-06, 'epoch': 2.19}
05/20/2024 16:35:42 - INFO - llamafactory.extras.callbacks - {'loss': 1.7387, 'learning_rate': 7.7097e-06, 'epoch': 2.23}
05/20/2024 16:37:03 - INFO - llamafactory.extras.callbacks - {'loss': 1.7606, 'learning_rate': 7.1172e-06, 'epoch': 2.26}
05/20/2024 16:38:24 - INFO - llamafactory.extras.callbacks - {'loss': 1.6940, 'learning_rate': 6.5446e-06, 'epoch': 2.29}
05/20/2024 16:39:47 - INFO - llamafactory.extras.callbacks - {'loss': 1.7982, 'learning_rate': 5.9926e-06, 'epoch': 2.32}
05/20/2024 16:41:07 - INFO - llamafactory.extras.callbacks - {'loss': 1.7905, 'learning_rate': 5.4616e-06, 'epoch': 2.35}
05/20/2024 16:42:29 - INFO - llamafactory.extras.callbacks - {'loss': 1.7480, 'learning_rate': 4.9525e-06, 'epoch': 2.38}
05/20/2024 16:43:51 - INFO - llamafactory.extras.callbacks - {'loss': 1.7059, 'learning_rate': 4.4656e-06, 'epoch': 2.42}
05/20/2024 16:45:13 - INFO - llamafactory.extras.callbacks - {'loss': 1.7080, 'learning_rate': 4.0015e-06, 'epoch': 2.45}
05/20/2024 16:46:34 - INFO - llamafactory.extras.callbacks - {'loss': 1.7037, 'learning_rate': 3.5608e-06, 'epoch': 2.48}
05/20/2024 16:47:56 - INFO - llamafactory.extras.callbacks - {'loss': 1.7715, 'learning_rate': 3.1439e-06, 'epoch': 2.51}
05/20/2024 16:49:17 - INFO - llamafactory.extras.callbacks - {'loss': 1.7584, 'learning_rate': 2.7514e-06, 'epoch': 2.54}
05/20/2024 16:49:17 - INFO - transformers.trainer - Saving model checkpoint to saves/Qwen1.5-7B-Chat/lora/train_2024-05-20-14-55-51/checkpoint-400
05/20/2024 16:49:18 - INFO - transformers.configuration_utils - loading configuration file config.json from cache at /home/featurize/.cache/huggingface/hub/models--Qwen--Qwen1.5-7B-Chat/snapshots/5f4f5e69ac7f1d508f8369e977de208b4803444b/config.json
05/20/2024 16:49:18 - INFO - transformers.configuration_utils - Model config Qwen2Config {
"architectures": [
"Qwen2ForCausalLM"
],
"attention_dropout": 0.0,
"bos_token_id": 151643,
"eos_token_id": 151645,
"hidden_act": "silu",
"hidden_size": 4096,
"initializer_range": 0.02,
"intermediate_size": 11008,
"max_position_embeddings": 32768,
"max_window_layers": 28,
"model_type": "qwen2",
"num_attention_heads": 32,
"num_hidden_layers": 32,
"num_key_value_heads": 32,
"rms_norm_eps": 1e-06,
"rope_theta": 1000000.0,
"sliding_window": 32768,
"tie_word_embeddings": false,
"torch_dtype": "bfloat16",
"transformers_version": "4.41.0",
"use_cache": true,
"use_sliding_window": false,
"vocab_size": 151936
}
05/20/2024 16:49:18 - INFO - transformers.tokenization_utils_base - tokenizer config file saved in saves/Qwen1.5-7B-Chat/lora/train_2024-05-20-14-55-51/checkpoint-400/tokenizer_config.json
05/20/2024 16:49:18 - INFO - transformers.tokenization_utils_base - Special tokens file saved in saves/Qwen1.5-7B-Chat/lora/train_2024-05-20-14-55-51/checkpoint-400/special_tokens_map.json
05/20/2024 16:50:40 - INFO - llamafactory.extras.callbacks - {'loss': 1.7246, 'learning_rate': 2.3836e-06, 'epoch': 2.58}
05/20/2024 16:52:03 - INFO - llamafactory.extras.callbacks - {'loss': 1.6902, 'learning_rate': 2.0409e-06, 'epoch': 2.61}
05/20/2024 16:53:25 - INFO - llamafactory.extras.callbacks - {'loss': 1.7111, 'learning_rate': 1.7238e-06, 'epoch': 2.64}
05/20/2024 16:54:47 - INFO - llamafactory.extras.callbacks - {'loss': 1.6956, 'learning_rate': 1.4326e-06, 'epoch': 2.67}
05/20/2024 16:56:09 - INFO - llamafactory.extras.callbacks - {'loss': 1.7892, 'learning_rate': 1.1675e-06, 'epoch': 2.70}
05/20/2024 16:57:31 - INFO - llamafactory.extras.callbacks - {'loss': 1.6886, 'learning_rate': 9.2902e-07, 'epoch': 2.73}
05/20/2024 16:58:52 - INFO - llamafactory.extras.callbacks - {'loss': 1.7672, 'learning_rate': 7.1727e-07, 'epoch': 2.77}
05/20/2024 17:00:14 - INFO - llamafactory.extras.callbacks - {'loss': 1.7667, 'learning_rate': 5.3253e-07, 'epoch': 2.80}
05/20/2024 17:01:36 - INFO - llamafactory.extras.callbacks - {'loss': 1.7337, 'learning_rate': 3.7500e-07, 'epoch': 2.83}
05/20/2024 17:02:58 - INFO - llamafactory.extras.callbacks - {'loss': 1.6704, 'learning_rate': 2.4485e-07, 'epoch': 2.86}
05/20/2024 17:04:20 - INFO - llamafactory.extras.callbacks - {'loss': 1.6902, 'learning_rate': 1.4223e-07, 'epoch': 2.89}
05/20/2024 17:05:41 - INFO - llamafactory.extras.callbacks - {'loss': 1.7985, 'learning_rate': 6.7260e-08, 'epoch': 2.93}
05/20/2024 17:07:02 - INFO - llamafactory.extras.callbacks - {'loss': 1.7464, 'learning_rate': 2.0018e-08, 'epoch': 2.96}
05/20/2024 17:08:24 - INFO - llamafactory.extras.callbacks - {'loss': 1.7430, 'learning_rate': 5.5612e-10, 'epoch': 2.99}
05/20/2024 17:08:40 - INFO - transformers.trainer -
Training completed. Do not forget to share your model on huggingface.co/models =)
05/20/2024 17:08:40 - INFO - transformers.trainer - Saving model checkpoint to saves/Qwen1.5-7B-Chat/lora/train_2024-05-20-14-55-51
05/20/2024 17:08:42 - INFO - transformers.configuration_utils - loading configuration file config.json from cache at /home/featurize/.cache/huggingface/hub/models--Qwen--Qwen1.5-7B-Chat/snapshots/5f4f5e69ac7f1d508f8369e977de208b4803444b/config.json
05/20/2024 17:08:42 - INFO - transformers.configuration_utils - Model config Qwen2Config {
"architectures": [
"Qwen2ForCausalLM"
],
"attention_dropout": 0.0,
"bos_token_id": 151643,
"eos_token_id": 151645,
"hidden_act": "silu",
"hidden_size": 4096,
"initializer_range": 0.02,
"intermediate_size": 11008,
"max_position_embeddings": 32768,
"max_window_layers": 28,
"model_type": "qwen2",
"num_attention_heads": 32,
"num_hidden_layers": 32,
"num_key_value_heads": 32,
"rms_norm_eps": 1e-06,
"rope_theta": 1000000.0,
"sliding_window": 32768,
"tie_word_embeddings": false,
"torch_dtype": "bfloat16",
"transformers_version": "4.41.0",
"use_cache": true,
"use_sliding_window": false,
"vocab_size": 151936
}
05/20/2024 17:08:42 - INFO - transformers.tokenization_utils_base - tokenizer config file saved in saves/Qwen1.5-7B-Chat/lora/train_2024-05-20-14-55-51/tokenizer_config.json
05/20/2024 17:08:42 - INFO - transformers.tokenization_utils_base - Special tokens file saved in saves/Qwen1.5-7B-Chat/lora/train_2024-05-20-14-55-51/special_tokens_map.json
05/20/2024 17:08:42 - WARNING - llamafactory.extras.ploting - No metric eval_loss to plot.
05/20/2024 17:08:42 - INFO - transformers.modelcard - Dropping the following result as it does not have all the necessary fields:
{'task': {'name': 'Causal Language Modeling', 'type': 'text-generation'}}