{ "model_name_or_path": "Qwen/Qwen1.5-0.5B", "max_length": 75, "model_kwargs": {}, "pooling_strategy": "avg", "lora_config_kwargs": { "task_type": "CAUSAL_LM", "r": 32, "lora_alpha": 32, "lora_dropout": 0.1, "bias": "none", "target_modules": [ "k_proj", "v_proj", "o_proj", "up_proj", "down_proj", "q_proj", "gate_proj" ] }, "is_llm": 1, "apply_billm": 1, "billm_model_class": "Qwen2ForCausalLM", "apply_lora": 1, "tokenizer_padding_side": null }