--- license: mit --- language: - en pipeline_tag: text-generation tags: --- # Qwen2-1.5B-Sign ## Introduction Qwen2-Sign is a text to sige model base on Qwen2. ## Finetune Details - Finetune dataset: [alpaca-zh-text2sign](https://huggingface.co/datasets/thundax/alpaca-zh-text2sign) - Finetune parameter | Parameter | Value | |------------------------------|--------| | learning_rate | 5e-05 | | train_batch_size | 4 | | eval_batch_size | 4 | | gradient_accumulation_steps | 8 | | total_train_batch_size | 32 | | lr_scheduler_type | cosine | | lr_scheduler_warmup_steps | 100 | | num_epochs | 4 | ## Quickstart ```python from transformers import AutoModelForCausalLM, AutoTokenizer device = "cuda" # the device to load the model onto model = AutoModelForCausalLM.from_pretrained( "thundax/Qwen2-1.5B-Sign", torch_dtype="auto", device_map="auto" ) tokenizer = AutoTokenizer.from_pretrained("thundax/Qwen2-1.5B-Sign") text = "你好,世界!" text = f'Translate sentence into labels\n{text}\n' model_inputs = tokenizer([text], return_tensors="pt").to(device) generated_ids = model.generate( model_inputs.input_ids, max_new_tokens=512 ) generated_ids = [ output_ids[len(input_ids):] for input_ids, output_ids in zip(model_inputs.input_ids, generated_ids) ] response = tokenizer.batch_decode(generated_ids, skip_special_tokens=True)[0] ``` ## Citation If you find our work helpful, feel free to give us a cite. ``` @software{qwen2-sign, author = {thundax}, title = {qwen2-sign: A Tool for Text to Sign}, year = {2024}, url = {https://github.com/thundax-lyp}, } ```