Edit model card

Alpaca-LoRA is an open-source project that reproduces results from Stanford Alpaca using Low-Rank Adaptation (LoRA) techniques. It provides an Instruct model of similar quality to text-davinci-003. Alpaca-LoRA uses the resource-efficient low-rank adaptation (LoRA) method, also widely used in Stable Diffusion, with Meta’s LLaMA to achieve results comparable to Alpaca Alpaca formula is open source, but may not be used commercially. However, the LLaMA model used for Alpaca is not released for commercial use, and the OpenAI GPT-3.5 terms of use prohibit using the model to develop AI models that compete with OpenAI. Stanford has therefore not yet released the model, only the training data and the code to generate the data and fine-tune the model.

Github link:

Inference: click red bar Space

The labeled dataset I used to fine-tune and train the Alpaca model can be found at: https://www.kaggle.com/datasets/praveengovi/emotions-dataset-for-nlp?select=train.txt

To create Space in HuggingFace: https://huggingface.co/new-space ( Select for CPU Upgrade or above)

https://huggingface.co/spaces/RinInori/alpaca_finetune_6_sentiments

Upload app.py and requirements.txt to https://huggingface.co/spaces/RinInori/alpaca_finetune_6_sentiments/tree/main

Alpaca ref: https://github.com/tloen/alpaca-lora

Downloads last month
2
Unable to determine this model’s pipeline type. Check the docs .