license: apache-2.0 | |
language: | |
- en | |
pipeline_tag: text-generation | |
Install Libs: !pip install -q -U transformers datasets accelerate peft trl bitsandbytes | |
My model trained on "garage-bAInd/Open-Platypus" data. | |
I have taken 1000 samples to fine tune LLaMA-2-7b. | |
Prompt which i used for prepare dataset: | |
def chat_template(example): | |
example["instruction"] = f"### Instruction:\n{example['instruction']}\n\n### Response:\n" | |
return example | |
dataset= dataset.map(chat_template) |