Edit model card

Llama-2-7b-guanaco

πŸ“ Article | πŸ’» Colab | πŸ“„ Script

This is a llama-2-7b-chat-hf model fine-tuned using QLoRA (4-bit precision) on the mlabonne/guanaco-llama2 dataset.

πŸ”§ Training

It was trained on a Google Colab notebook with a T4 GPU and high RAM.

πŸ’» Usage

# pip install transformers accelerate

from transformers import AutoTokenizer
import transformers
import torch

model = "mlabonne/llama-2-7b-miniguanaco"
prompt = "What is a large language model?"

tokenizer = AutoTokenizer.from_pretrained(model)
pipeline = transformers.pipeline(
    "text-generation",
    model=model,
    torch_dtype=torch.float16,
    device_map="auto",
)

sequences = pipeline(
    f'<s>[INST] {prompt} [/INST]',
    do_sample=True,
    top_k=10,
    num_return_sequences=1,
    eos_token_id=tokenizer.eos_token_id,
    max_length=200,
)
for seq in sequences:
    print(f"Result: {seq['generated_text']}")

Output:

A large language model is a type of artificial intelligence (AI) model that is trained to generate human-like language. The models can be trained on text from a specific genre, such as news articles, or on a large corpus of text, such as the internet. They can then be used to generate text, such as articles, stories or even entire books. These models are often used in applications such as chatbots, language translation and content generation. They have been used to write books such as: "The Last Days of New Paris" by China MiΓ©ville.

The large models are also used for many other applications such as:

  • Translation
  • Summarization
  • Sentiment Analysis
  • Text classification
  • Generative writing (creates articles, stories, and more.)
  • Conversational language generation.
Downloads last month
13

Dataset used to train mlabonne/llama-2-7b-guanaco

Space using mlabonne/llama-2-7b-guanaco 1

Collection including mlabonne/llama-2-7b-guanaco