06/06/2024 20:45:49 - INFO - transformers.tokenization_utils_base - loading file vocab.json from cache at /home/zeus/.cache/huggingface/hub/models--Qwen--Qwen2-7B-Instruct-GPTQ-Int8/snapshots/22d9ef924dfae7700c4b620be6600d5c7ea31f88/vocab.json 06/06/2024 20:45:49 - INFO - transformers.tokenization_utils_base - loading file merges.txt from cache at /home/zeus/.cache/huggingface/hub/models--Qwen--Qwen2-7B-Instruct-GPTQ-Int8/snapshots/22d9ef924dfae7700c4b620be6600d5c7ea31f88/merges.txt 06/06/2024 20:45:49 - INFO - transformers.tokenization_utils_base - loading file tokenizer.json from cache at /home/zeus/.cache/huggingface/hub/models--Qwen--Qwen2-7B-Instruct-GPTQ-Int8/snapshots/22d9ef924dfae7700c4b620be6600d5c7ea31f88/tokenizer.json 06/06/2024 20:45:49 - INFO - transformers.tokenization_utils_base - loading file added_tokens.json from cache at None 06/06/2024 20:45:49 - INFO - transformers.tokenization_utils_base - loading file special_tokens_map.json from cache at None 06/06/2024 20:45:49 - INFO - transformers.tokenization_utils_base - loading file tokenizer_config.json from cache at /home/zeus/.cache/huggingface/hub/models--Qwen--Qwen2-7B-Instruct-GPTQ-Int8/snapshots/22d9ef924dfae7700c4b620be6600d5c7ea31f88/tokenizer_config.json 06/06/2024 20:45:49 - WARNING - transformers.tokenization_utils_base - Special tokens have been added in the vocabulary, make sure the associated word embeddings are fine-tuned or trained. 06/06/2024 20:45:49 - INFO - llamafactory.data.template - Replace eos token: <|im_end|> 06/06/2024 20:45:49 - INFO - llamafactory.data.loader - Loading dataset train.json... 06/06/2024 20:45:49 - INFO - transformers.configuration_utils - loading configuration file config.json from cache at /home/zeus/.cache/huggingface/hub/models--Qwen--Qwen2-7B-Instruct-GPTQ-Int8/snapshots/22d9ef924dfae7700c4b620be6600d5c7ea31f88/config.json 06/06/2024 20:45:49 - INFO - transformers.configuration_utils - Model config Qwen2Config { "_name_or_path": "Qwen/Qwen2-7B-Instruct-GPTQ-Int8", "architectures": [ "Qwen2ForCausalLM" ], "attention_dropout": 0.0, "bos_token_id": 151643, "eos_token_id": 151645, "hidden_act": "silu", "hidden_size": 3584, "initializer_range": 0.02, "intermediate_size": 18944, "max_position_embeddings": 32768, "max_window_layers": 28, "model_type": "qwen2", "num_attention_heads": 28, "num_hidden_layers": 28, "num_key_value_heads": 4, "quantization_config": { "batch_size": 1, "bits": 8, "block_name_to_quantize": null, "cache_block_outputs": true, "damp_percent": 0.1, "dataset": null, "desc_act": false, "exllama_config": { "version": 1 }, "group_size": 128, "max_input_length": null, "model_seqlen": null, "module_name_preceding_first_block": null, "modules_in_block_to_quantize": null, "pad_token_id": null, "quant_method": "gptq", "sym": true, "tokenizer": null, "true_sequential": true, "use_cuda_fp16": false, "use_exllama": true }, "rms_norm_eps": 1e-06, "rope_theta": 1000000.0, "sliding_window": 131072, "tie_word_embeddings": false, "torch_dtype": "float16", "transformers_version": "4.41.2", "use_cache": true, "use_sliding_window": false, "vocab_size": 152064 } 06/06/2024 20:45:49 - WARNING - llamafactory.model.model_utils.attention - FlashAttention-2 is not installed. 06/06/2024 20:45:49 - INFO - llamafactory.model.model_utils.quantization - Loading 8-bit GPTQ-quantized model. 06/06/2024 20:45:49 - INFO - transformers.modeling_utils - loading weights file model.safetensors from cache at /home/zeus/.cache/huggingface/hub/models--Qwen--Qwen2-7B-Instruct-GPTQ-Int8/snapshots/22d9ef924dfae7700c4b620be6600d5c7ea31f88/model.safetensors.index.json 06/06/2024 20:46:15 - INFO - transformers.modeling_utils - Instantiating Qwen2ForCausalLM model under default dtype torch.float16. 06/06/2024 20:46:15 - INFO - transformers.generation.configuration_utils - Generate config GenerationConfig { "bos_token_id": 151643, "eos_token_id": 151645, "use_cache": false } 06/06/2024 20:46:41 - WARNING - transformers.modeling_utils - Some weights of the model checkpoint at Qwen/Qwen2-7B-Instruct-GPTQ-Int8 were not used when initializing Qwen2ForCausalLM: ['model.layers.0.mlp.down_proj.bias', 'model.layers.0.mlp.gate_proj.bias', 'model.layers.0.mlp.up_proj.bias', 'model.layers.0.self_attn.o_proj.bias', 'model.layers.1.mlp.down_proj.bias', 'model.layers.1.mlp.gate_proj.bias', 'model.layers.1.mlp.up_proj.bias', 'model.layers.1.self_attn.o_proj.bias', 'model.layers.10.mlp.down_proj.bias', 'model.layers.10.mlp.gate_proj.bias', 'model.layers.10.mlp.up_proj.bias', 'model.layers.10.self_attn.o_proj.bias', 'model.layers.11.mlp.down_proj.bias', 'model.layers.11.mlp.gate_proj.bias', 'model.layers.11.mlp.up_proj.bias', 'model.layers.11.self_attn.o_proj.bias', 'model.layers.12.mlp.down_proj.bias', 'model.layers.12.mlp.gate_proj.bias', 'model.layers.12.mlp.up_proj.bias', 'model.layers.12.self_attn.o_proj.bias', 'model.layers.13.mlp.down_proj.bias', 'model.layers.13.mlp.gate_proj.bias', 'model.layers.13.mlp.up_proj.bias', 'model.layers.13.self_attn.o_proj.bias', 'model.layers.14.mlp.down_proj.bias', 'model.layers.14.mlp.gate_proj.bias', 'model.layers.14.mlp.up_proj.bias', 'model.layers.14.self_attn.o_proj.bias', 'model.layers.15.mlp.down_proj.bias', 'model.layers.15.mlp.gate_proj.bias', 'model.layers.15.mlp.up_proj.bias', 'model.layers.15.self_attn.o_proj.bias', 'model.layers.16.mlp.down_proj.bias', 'model.layers.16.mlp.gate_proj.bias', 'model.layers.16.mlp.up_proj.bias', 'model.layers.16.self_attn.o_proj.bias', 'model.layers.17.mlp.down_proj.bias', 'model.layers.17.mlp.gate_proj.bias', 'model.layers.17.mlp.up_proj.bias', 'model.layers.17.self_attn.o_proj.bias', 'model.layers.18.mlp.down_proj.bias', 'model.layers.18.mlp.gate_proj.bias', 'model.layers.18.mlp.up_proj.bias', 'model.layers.18.self_attn.o_proj.bias', 'model.layers.19.mlp.down_proj.bias', 'model.layers.19.mlp.gate_proj.bias', 'model.layers.19.mlp.up_proj.bias', 'model.layers.19.self_attn.o_proj.bias', 'model.layers.2.mlp.down_proj.bias', 'model.layers.2.mlp.gate_proj.bias', 'model.layers.2.mlp.up_proj.bias', 'model.layers.2.self_attn.o_proj.bias', 'model.layers.20.mlp.down_proj.bias', 'model.layers.20.mlp.gate_proj.bias', 'model.layers.20.mlp.up_proj.bias', 'model.layers.20.self_attn.o_proj.bias', 'model.layers.21.mlp.down_proj.bias', 'model.layers.21.mlp.gate_proj.bias', 'model.layers.21.mlp.up_proj.bias', 'model.layers.21.self_attn.o_proj.bias', 'model.layers.22.mlp.down_proj.bias', 'model.layers.22.mlp.gate_proj.bias', 'model.layers.22.mlp.up_proj.bias', 'model.layers.22.self_attn.o_proj.bias', 'model.layers.23.mlp.down_proj.bias', 'model.layers.23.mlp.gate_proj.bias', 'model.layers.23.mlp.up_proj.bias', 'model.layers.23.self_attn.o_proj.bias', 'model.layers.24.mlp.down_proj.bias', 'model.layers.24.mlp.gate_proj.bias', 'model.layers.24.mlp.up_proj.bias', 'model.layers.24.self_attn.o_proj.bias', 'model.layers.25.mlp.down_proj.bias', 'model.layers.25.mlp.gate_proj.bias', 'model.layers.25.mlp.up_proj.bias', 'model.layers.25.self_attn.o_proj.bias', 'model.layers.26.mlp.down_proj.bias', 'model.layers.26.mlp.gate_proj.bias', 'model.layers.26.mlp.up_proj.bias', 'model.layers.26.self_attn.o_proj.bias', 'model.layers.27.mlp.down_proj.bias', 'model.layers.27.mlp.gate_proj.bias', 'model.layers.27.mlp.up_proj.bias', 'model.layers.27.self_attn.o_proj.bias', 'model.layers.3.mlp.down_proj.bias', 'model.layers.3.mlp.gate_proj.bias', 'model.layers.3.mlp.up_proj.bias', 'model.layers.3.self_attn.o_proj.bias', 'model.layers.4.mlp.down_proj.bias', 'model.layers.4.mlp.gate_proj.bias', 'model.layers.4.mlp.up_proj.bias', 'model.layers.4.self_attn.o_proj.bias', 'model.layers.5.mlp.down_proj.bias', 'model.layers.5.mlp.gate_proj.bias', 'model.layers.5.mlp.up_proj.bias', 'model.layers.5.self_attn.o_proj.bias', 'model.layers.6.mlp.down_proj.bias', 'model.layers.6.mlp.gate_proj.bias', 'model.layers.6.mlp.up_proj.bias', 'model.layers.6.self_attn.o_proj.bias', 'model.layers.7.mlp.down_proj.bias', 'model.layers.7.mlp.gate_proj.bias', 'model.layers.7.mlp.up_proj.bias', 'model.layers.7.self_attn.o_proj.bias', 'model.layers.8.mlp.down_proj.bias', 'model.layers.8.mlp.gate_proj.bias', 'model.layers.8.mlp.up_proj.bias', 'model.layers.8.self_attn.o_proj.bias', 'model.layers.9.mlp.down_proj.bias', 'model.layers.9.mlp.gate_proj.bias', 'model.layers.9.mlp.up_proj.bias', 'model.layers.9.self_attn.o_proj.bias'] - This IS expected if you are initializing Qwen2ForCausalLM from the checkpoint of a model trained on another task or with another architecture (e.g. initializing a BertForSequenceClassification model from a BertForPreTraining model). - This IS NOT expected if you are initializing Qwen2ForCausalLM from the checkpoint of a model that you expect to be exactly identical (initializing a BertForSequenceClassification model from a BertForSequenceClassification model). 06/06/2024 20:46:41 - INFO - transformers.modeling_utils - All the weights of Qwen2ForCausalLM were initialized from the model checkpoint at Qwen/Qwen2-7B-Instruct-GPTQ-Int8. If your task is similar to the task the model of the checkpoint was trained on, you can already use Qwen2ForCausalLM for predictions without further training. 06/06/2024 20:46:41 - INFO - transformers.generation.configuration_utils - loading configuration file generation_config.json from cache at /home/zeus/.cache/huggingface/hub/models--Qwen--Qwen2-7B-Instruct-GPTQ-Int8/snapshots/22d9ef924dfae7700c4b620be6600d5c7ea31f88/generation_config.json 06/06/2024 20:46:41 - INFO - transformers.generation.configuration_utils - Generate config GenerationConfig { "bos_token_id": 151643, "do_sample": true, "eos_token_id": [ 151645, 151643 ], "pad_token_id": 151643, "repetition_penalty": 1.05, "top_p": 0.8 } 06/06/2024 20:46:41 - INFO - llamafactory.model.model_utils.checkpointing - Gradient checkpointing enabled. 06/06/2024 20:46:41 - INFO - llamafactory.model.model_utils.attention - Using torch SDPA for faster training and inference. 06/06/2024 20:46:41 - INFO - llamafactory.model.adapter - Upcasting trainable params to float32. 06/06/2024 20:46:41 - INFO - llamafactory.model.adapter - Fine-tuning method: LoRA 06/06/2024 20:46:41 - INFO - llamafactory.model.model_utils.misc - Found linear modules: q_proj,o_proj,k_proj,gate_proj,down_proj,up_proj,v_proj 06/06/2024 20:46:42 - INFO - llamafactory.model.loader - trainable params: 20185088 || all params: 1110384128 || trainable%: 1.8178 06/06/2024 20:46:42 - INFO - transformers.trainer - Using auto half precision backend 06/06/2024 20:46:42 - INFO - transformers.trainer - ***** Running training ***** 06/06/2024 20:46:42 - INFO - transformers.trainer - Num examples = 9,600 06/06/2024 20:46:42 - INFO - transformers.trainer - Num Epochs = 3 06/06/2024 20:46:42 - INFO - transformers.trainer - Instantaneous batch size per device = 2 06/06/2024 20:46:42 - INFO - transformers.trainer - Total train batch size (w. parallel, distributed & accumulation) = 16 06/06/2024 20:46:42 - INFO - transformers.trainer - Gradient Accumulation steps = 8 06/06/2024 20:46:42 - INFO - transformers.trainer - Total optimization steps = 1,800 06/06/2024 20:46:42 - INFO - transformers.trainer - Number of trainable parameters = 20,185,088 06/06/2024 20:48:17 - INFO - llamafactory.extras.callbacks - {'loss': 0.4212, 'learning_rate': 4.9999e-05, 'epoch': 0.01, 'throughput': 317.96} 06/06/2024 20:49:49 - INFO - llamafactory.extras.callbacks - {'loss': 0.2803, 'learning_rate': 4.9996e-05, 'epoch': 0.02, 'throughput': 307.70} 06/06/2024 20:51:24 - INFO - llamafactory.extras.callbacks - {'loss': 0.1484, 'learning_rate': 4.9991e-05, 'epoch': 0.03, 'throughput': 305.82} 06/06/2024 20:52:54 - INFO - llamafactory.extras.callbacks - {'loss': 0.1151, 'learning_rate': 4.9985e-05, 'epoch': 0.03, 'throughput': 297.64} 06/06/2024 20:54:29 - INFO - llamafactory.extras.callbacks - {'loss': 0.0751, 'learning_rate': 4.9976e-05, 'epoch': 0.04, 'throughput': 297.62} 06/06/2024 20:56:07 - INFO - llamafactory.extras.callbacks - {'loss': 0.0471, 'learning_rate': 4.9966e-05, 'epoch': 0.05, 'throughput': 300.36} 06/06/2024 20:57:47 - INFO - llamafactory.extras.callbacks - {'loss': 0.0368, 'learning_rate': 4.9953e-05, 'epoch': 0.06, 'throughput': 303.39} 06/06/2024 20:59:22 - INFO - llamafactory.extras.callbacks - {'loss': 0.0311, 'learning_rate': 4.9939e-05, 'epoch': 0.07, 'throughput': 302.60} 06/06/2024 21:00:58 - INFO - llamafactory.extras.callbacks - {'loss': 0.0262, 'learning_rate': 4.9923e-05, 'epoch': 0.07, 'throughput': 302.73} 06/06/2024 21:02:36 - INFO - llamafactory.extras.callbacks - {'loss': 0.0244, 'learning_rate': 4.9905e-05, 'epoch': 0.08, 'throughput': 303.83} 06/06/2024 21:04:11 - INFO - llamafactory.extras.callbacks - {'loss': 0.0266, 'learning_rate': 4.9885e-05, 'epoch': 0.09, 'throughput': 303.09} 06/06/2024 21:05:46 - INFO - llamafactory.extras.callbacks - {'loss': 0.0283, 'learning_rate': 4.9868e-05, 'epoch': 0.10, 'throughput': 300.01} 06/06/2024 21:07:21 - INFO - llamafactory.extras.callbacks - {'loss': 0.0265, 'learning_rate': 4.9844e-05, 'epoch': 0.11, 'throughput': 299.96} 06/06/2024 21:08:54 - INFO - llamafactory.extras.callbacks - {'loss': 0.0235, 'learning_rate': 4.9819e-05, 'epoch': 0.12, 'throughput': 299.39} 06/06/2024 21:10:28 - INFO - llamafactory.extras.callbacks - {'loss': 0.0229, 'learning_rate': 4.9792e-05, 'epoch': 0.12, 'throughput': 299.15} 06/06/2024 21:12:01 - INFO - llamafactory.extras.callbacks - {'loss': 0.0238, 'learning_rate': 4.9763e-05, 'epoch': 0.13, 'throughput': 298.31} 06/06/2024 21:13:35 - INFO - llamafactory.extras.callbacks - {'loss': 0.0208, 'learning_rate': 4.9732e-05, 'epoch': 0.14, 'throughput': 298.24} 06/06/2024 21:15:11 - INFO - llamafactory.extras.callbacks - {'loss': 0.0191, 'learning_rate': 4.9699e-05, 'epoch': 0.15, 'throughput': 298.57} 06/06/2024 21:16:43 - INFO - llamafactory.extras.callbacks - {'loss': 0.0205, 'learning_rate': 4.9664e-05, 'epoch': 0.16, 'throughput': 298.01} 06/06/2024 21:18:20 - INFO - llamafactory.extras.callbacks - {'loss': 0.0193, 'learning_rate': 4.9628e-05, 'epoch': 0.17, 'throughput': 298.56} 06/06/2024 21:18:20 - INFO - transformers.trainer - Saving model checkpoint to saves/Qwen2-7B-int8-Chat/lora/train_2024-06-06-20-45-21/checkpoint-100 06/06/2024 21:18:20 - INFO - transformers.configuration_utils - loading configuration file config.json from cache at /home/zeus/.cache/huggingface/hub/models--Qwen--Qwen2-7B-Instruct-GPTQ-Int8/snapshots/22d9ef924dfae7700c4b620be6600d5c7ea31f88/config.json 06/06/2024 21:18:20 - INFO - transformers.configuration_utils - Model config Qwen2Config { "architectures": [ "Qwen2ForCausalLM" ], "attention_dropout": 0.0, "bos_token_id": 151643, "eos_token_id": 151645, "hidden_act": "silu", "hidden_size": 3584, "initializer_range": 0.02, "intermediate_size": 18944, "max_position_embeddings": 32768, "max_window_layers": 28, "model_type": "qwen2", "num_attention_heads": 28, "num_hidden_layers": 28, "num_key_value_heads": 4, "quantization_config": { "batch_size": 1, "bits": 8, "block_name_to_quantize": null, "cache_block_outputs": true, "damp_percent": 0.1, "dataset": null, "desc_act": false, "exllama_config": { "version": 1 }, "group_size": 128, "max_input_length": null, "model_seqlen": null, "module_name_preceding_first_block": null, "modules_in_block_to_quantize": null, "pad_token_id": null, "quant_method": "gptq", "sym": true, "tokenizer": null, "true_sequential": true, "use_cuda_fp16": false, "use_exllama": true }, "rms_norm_eps": 1e-06, "rope_theta": 1000000.0, "sliding_window": 131072, "tie_word_embeddings": false, "torch_dtype": "float16", "transformers_version": "4.41.2", "use_cache": true, "use_sliding_window": false, "vocab_size": 152064 } 06/06/2024 21:18:20 - INFO - transformers.tokenization_utils_base - tokenizer config file saved in saves/Qwen2-7B-int8-Chat/lora/train_2024-06-06-20-45-21/checkpoint-100/tokenizer_config.json 06/06/2024 21:18:20 - INFO - transformers.tokenization_utils_base - Special tokens file saved in saves/Qwen2-7B-int8-Chat/lora/train_2024-06-06-20-45-21/checkpoint-100/special_tokens_map.json 06/06/2024 21:19:53 - INFO - llamafactory.extras.callbacks - {'loss': 0.0240, 'learning_rate': 4.9589e-05, 'epoch': 0.17, 'throughput': 298.12} 06/06/2024 21:21:27 - INFO - llamafactory.extras.callbacks - {'loss': 0.0206, 'learning_rate': 4.9549e-05, 'epoch': 0.18, 'throughput': 297.89} 06/06/2024 21:22:59 - INFO - llamafactory.extras.callbacks - {'loss': 0.0188, 'learning_rate': 4.9507e-05, 'epoch': 0.19, 'throughput': 297.43} 06/06/2024 21:24:35 - INFO - llamafactory.extras.callbacks - {'loss': 0.0183, 'learning_rate': 4.9463e-05, 'epoch': 0.20, 'throughput': 297.64} 06/06/2024 21:26:04 - INFO - llamafactory.extras.callbacks - {'loss': 0.0189, 'learning_rate': 4.9417e-05, 'epoch': 0.21, 'throughput': 297.06} 06/06/2024 21:27:40 - INFO - llamafactory.extras.callbacks - {'loss': 0.0213, 'learning_rate': 4.9369e-05, 'epoch': 0.22, 'throughput': 297.53} 06/06/2024 21:29:12 - INFO - llamafactory.extras.callbacks - {'loss': 0.0214, 'learning_rate': 4.9319e-05, 'epoch': 0.23, 'throughput': 297.22} 06/06/2024 21:30:47 - INFO - llamafactory.extras.callbacks - {'loss': 0.0214, 'learning_rate': 4.9268e-05, 'epoch': 0.23, 'throughput': 297.22} 06/06/2024 21:32:22 - INFO - llamafactory.extras.callbacks - {'loss': 0.0203, 'learning_rate': 4.9215e-05, 'epoch': 0.24, 'throughput': 297.36} 06/06/2024 21:33:56 - INFO - llamafactory.extras.callbacks - {'loss': 0.0194, 'learning_rate': 4.9159e-05, 'epoch': 0.25, 'throughput': 297.35} 06/06/2024 21:35:27 - INFO - llamafactory.extras.callbacks - {'loss': 0.0158, 'learning_rate': 4.9102e-05, 'epoch': 0.26, 'throughput': 296.96} 06/06/2024 21:37:01 - INFO - llamafactory.extras.callbacks - {'loss': 0.0172, 'learning_rate': 4.9044e-05, 'epoch': 0.27, 'throughput': 296.94} 06/06/2024 21:38:35 - INFO - llamafactory.extras.callbacks - {'loss': 0.0193, 'learning_rate': 4.8983e-05, 'epoch': 0.28, 'throughput': 297.06} 06/06/2024 21:40:13 - INFO - llamafactory.extras.callbacks - {'loss': 0.0189, 'learning_rate': 4.8920e-05, 'epoch': 0.28, 'throughput': 297.62} 06/06/2024 21:41:49 - INFO - llamafactory.extras.callbacks - {'loss': 0.0190, 'learning_rate': 4.8856e-05, 'epoch': 0.29, 'throughput': 297.91} 06/06/2024 21:43:23 - INFO - llamafactory.extras.callbacks - {'loss': 0.0168, 'learning_rate': 4.8790e-05, 'epoch': 0.30, 'throughput': 297.99} 06/06/2024 21:45:00 - INFO - llamafactory.extras.callbacks - {'loss': 0.0194, 'learning_rate': 4.8722e-05, 'epoch': 0.31, 'throughput': 298.39} 06/06/2024 21:46:33 - INFO - llamafactory.extras.callbacks - {'loss': 0.0155, 'learning_rate': 4.8652e-05, 'epoch': 0.32, 'throughput': 298.24} 06/06/2024 21:48:05 - INFO - llamafactory.extras.callbacks - {'loss': 0.0193, 'learning_rate': 4.8581e-05, 'epoch': 0.33, 'throughput': 298.10} 06/06/2024 21:49:40 - INFO - llamafactory.extras.callbacks - {'loss': 0.0164, 'learning_rate': 4.8507e-05, 'epoch': 0.33, 'throughput': 298.12} 06/06/2024 21:49:40 - INFO - transformers.trainer - Saving model checkpoint to saves/Qwen2-7B-int8-Chat/lora/train_2024-06-06-20-45-21/checkpoint-200 06/06/2024 21:49:40 - INFO - transformers.configuration_utils - loading configuration file config.json from cache at /home/zeus/.cache/huggingface/hub/models--Qwen--Qwen2-7B-Instruct-GPTQ-Int8/snapshots/22d9ef924dfae7700c4b620be6600d5c7ea31f88/config.json 06/06/2024 21:49:40 - INFO - transformers.configuration_utils - Model config Qwen2Config { "architectures": [ "Qwen2ForCausalLM" ], "attention_dropout": 0.0, "bos_token_id": 151643, "eos_token_id": 151645, "hidden_act": "silu", "hidden_size": 3584, "initializer_range": 0.02, "intermediate_size": 18944, "max_position_embeddings": 32768, "max_window_layers": 28, "model_type": "qwen2", "num_attention_heads": 28, "num_hidden_layers": 28, "num_key_value_heads": 4, "quantization_config": { "batch_size": 1, "bits": 8, "block_name_to_quantize": null, "cache_block_outputs": true, "damp_percent": 0.1, "dataset": null, "desc_act": false, "exllama_config": { "version": 1 }, "group_size": 128, "max_input_length": null, "model_seqlen": null, "module_name_preceding_first_block": null, "modules_in_block_to_quantize": null, "pad_token_id": null, "quant_method": "gptq", "sym": true, "tokenizer": null, "true_sequential": true, "use_cuda_fp16": false, "use_exllama": true }, "rms_norm_eps": 1e-06, "rope_theta": 1000000.0, "sliding_window": 131072, "tie_word_embeddings": false, "torch_dtype": "float16", "transformers_version": "4.41.2", "use_cache": true, "use_sliding_window": false, "vocab_size": 152064 } 06/06/2024 21:49:40 - INFO - transformers.tokenization_utils_base - tokenizer config file saved in saves/Qwen2-7B-int8-Chat/lora/train_2024-06-06-20-45-21/checkpoint-200/tokenizer_config.json 06/06/2024 21:49:40 - INFO - transformers.tokenization_utils_base - Special tokens file saved in saves/Qwen2-7B-int8-Chat/lora/train_2024-06-06-20-45-21/checkpoint-200/special_tokens_map.json 06/06/2024 21:51:18 - INFO - llamafactory.extras.callbacks - {'loss': 0.0140, 'learning_rate': 4.8432e-05, 'epoch': 0.34, 'throughput': 298.66} 06/06/2024 21:52:51 - INFO - llamafactory.extras.callbacks - {'loss': 0.0154, 'learning_rate': 4.8355e-05, 'epoch': 0.35, 'throughput': 298.57} 06/06/2024 21:54:25 - INFO - llamafactory.extras.callbacks - {'loss': 0.0147, 'learning_rate': 4.8276e-05, 'epoch': 0.36, 'throughput': 298.63} 06/06/2024 21:55:59 - INFO - llamafactory.extras.callbacks - {'loss': 0.0185, 'learning_rate': 4.8196e-05, 'epoch': 0.37, 'throughput': 298.72} 06/06/2024 21:57:34 - INFO - llamafactory.extras.callbacks - {'loss': 0.0160, 'learning_rate': 4.8114e-05, 'epoch': 0.38, 'throughput': 298.74} 06/06/2024 21:59:07 - INFO - llamafactory.extras.callbacks - {'loss': 0.0168, 'learning_rate': 4.8030e-05, 'epoch': 0.38, 'throughput': 298.65} 06/06/2024 22:00:38 - INFO - llamafactory.extras.callbacks - {'loss': 0.0154, 'learning_rate': 4.7944e-05, 'epoch': 0.39, 'throughput': 298.39} 06/06/2024 22:02:21 - INFO - llamafactory.extras.callbacks - {'loss': 0.0149, 'learning_rate': 4.7856e-05, 'epoch': 0.40, 'throughput': 299.32} 06/06/2024 22:03:05 - INFO - transformers.trainer - Training completed. Do not forget to share your model on huggingface.co/models =) 06/06/2024 22:03:05 - INFO - transformers.trainer - Saving model checkpoint to saves/Qwen2-7B-int8-Chat/lora/train_2024-06-06-20-45-21 06/06/2024 22:03:06 - INFO - transformers.configuration_utils - loading configuration file config.json from cache at /home/zeus/.cache/huggingface/hub/models--Qwen--Qwen2-7B-Instruct-GPTQ-Int8/snapshots/22d9ef924dfae7700c4b620be6600d5c7ea31f88/config.json 06/06/2024 22:03:06 - INFO - transformers.configuration_utils - Model config Qwen2Config { "architectures": [ "Qwen2ForCausalLM" ], "attention_dropout": 0.0, "bos_token_id": 151643, "eos_token_id": 151645, "hidden_act": "silu", "hidden_size": 3584, "initializer_range": 0.02, "intermediate_size": 18944, "max_position_embeddings": 32768, "max_window_layers": 28, "model_type": "qwen2", "num_attention_heads": 28, "num_hidden_layers": 28, "num_key_value_heads": 4, "quantization_config": { "batch_size": 1, "bits": 8, "block_name_to_quantize": null, "cache_block_outputs": true, "damp_percent": 0.1, "dataset": null, "desc_act": false, "exllama_config": { "version": 1 }, "group_size": 128, "max_input_length": null, "model_seqlen": null, "module_name_preceding_first_block": null, "modules_in_block_to_quantize": null, "pad_token_id": null, "quant_method": "gptq", "sym": true, "tokenizer": null, "true_sequential": true, "use_cuda_fp16": false, "use_exllama": true }, "rms_norm_eps": 1e-06, "rope_theta": 1000000.0, "sliding_window": 131072, "tie_word_embeddings": false, "torch_dtype": "float16", "transformers_version": "4.41.2", "use_cache": true, "use_sliding_window": false, "vocab_size": 152064 } 06/06/2024 22:03:06 - INFO - transformers.tokenization_utils_base - tokenizer config file saved in saves/Qwen2-7B-int8-Chat/lora/train_2024-06-06-20-45-21/tokenizer_config.json 06/06/2024 22:03:06 - INFO - transformers.tokenization_utils_base - Special tokens file saved in saves/Qwen2-7B-int8-Chat/lora/train_2024-06-06-20-45-21/special_tokens_map.json 06/06/2024 22:03:06 - WARNING - llamafactory.extras.ploting - No metric eval_loss to plot. 06/06/2024 22:03:06 - INFO - transformers.modelcard - Dropping the following result as it does not have all the necessary fields: {'task': {'name': 'Causal Language Modeling', 'type': 'text-generation'}}