Edit model card

How to use:

from transformers import TextStreamer
from unsloth import FastLanguageModel
import torch

model, tokenizer = FastLanguageModel.from_pretrained(
    model_name = "AdrienB134/French-Alpaca-Croissant-1.3B-Instruct",
    max_seq_length = 4096,
    dtype = None,
    load_in_4bit = True,
    fix_tokenizer = False,
)

alpaca_prompt = """Below is an instruction that describes a task, paired with an input that provides further context. Write a response that appropriately completes the request.

### Instruction:
{}

### Input:
{}

### Response:
{}"""

FastLanguageModel.for_inference(model) 

inputs = tokenizer(
[
    alpaca_prompt.format(
        "Continue la suite de Fibonnaci", # instruction
        "1, 1, 2, 3, 5, 8", # input
        "", # output - leave this blank for generation!
    )
], return_tensors = "pt").to("cuda")

text_streamer = TextStreamer(tokenizer)
_ = model.generate(**inputs, streamer = text_streamer, max_new_tokens = 128)

Uploaded model

  • Developed by: AdrienB134
  • License: MIT
  • Finetuned from model : croissantllm/CroissantLLMBase

This llama model was trained 2x faster with Unsloth and Huggingface's TRL library.

Downloads last month
35
Inference Examples
This model does not have enough activity to be deployed to Inference API (serverless) yet. Increase its social visibility and check back later, or deploy to Inference Endpoints (dedicated) instead.

Model tree for AdrienB134/French-Alpaca-Croissant-1.3B-Instruct

Finetuned
(9)
this model

Dataset used to train AdrienB134/French-Alpaca-Croissant-1.3B-Instruct