--- language: - en license: apache-2.0 tags: - text-generation-inference - transformers - unsloth - llama - trl base_model: unsloth/llama-3-8b-Instruct-bnb-4bit --- # Model: BAI_LLM_FinArg - **Developed by:** varadsrivastava - **License:** apache-2.0 - **Base Model :** unsloth/llama-3-8b-Instruct-bnb-4bit # For Proper Inference, please use: !pip install "unsloth[colab-new] @ git+https://GitHub.com/unslothai/unsloth.git ### Loading the fine-tuned model and the tokenizer for inference from unsloth import FastLanguageModel model, tokenizer = FastLanguageModel.from_pretrained( model_name = "varadsrivastava/BAI_LLM_FinArg", max_seq_length = 20, dtype = torch.bfloat16, load_in_4bit = True ) ### Using FastLanguageModel for fast inference FastLanguageModel.for_inference(model) # Prompt template: """<|begin_of_text|><|start_header_id|>system<|end_header_id|> {instruction}<|eot_id|><|start_header_id|>user<|end_header_id|> Sentence: {row['text']}<|eot_id|><|start_header_id|>assistant<|end_header_id|> Class: {row['label']}<|eot_id|>""" NOTE: This model was trained 2x faster using Unsloth and Huggingface's TRL library.