Edit model card

LLaMA-7B

This repo contains a low-rank adapter for LLaMA-7b trained on Bulgarian dataset.

This model was introduced in this paper.

Model description

The training data is private Bulgarian dataset.

Intended uses & limitations

This is an instruction-based model similart to ChatGPT but in Bulgarian.

How to use

Here is how to use this model in PyTorch:

>>> git clone https://github.com/tloen/alpaca-lora.git
>>> cd alpaca-lora
>>> pip install -r requirements.txt
>>>
>>> python generate.py \
    --load_8bit \
    --base_model 'yahma/llama-7b-hf' \
    --lora_weights 'rmihaylov/alpaca-lora-bg-7b' \
    --share_gradio

This will download both a base model and an adapter from huggingface. Then it will run a gradio interface for chatting.

Example using this model: Colab. You need a Colab Pro because the model needs high ram when loading.

Interface

Downloads last month
0
Inference API (serverless) has been turned off for this model.