Mistral-Ita-7b / README.md
DeepMount00's picture
Update README.md
284719e verified
metadata
language:
  - it
license: apache-2.0
tags:
  - text-generation-inference
  - text generation

Mistral-7B-v0.1 for Italian Language Text Generation

Overview

Mistral-7B-v0.1 is a state-of-the-art Large Language Model (LLM) specifically pre-trained for generating text. With its 7 billion parameters, it's built to excel in benchmarks and outperforms even some larger models like the Llama 2 13B.

Model Architecture

The Mistral-7B-v0.1 model is a transformer-based model that can handle a variety of tasks including but not limited to translation, summarization, and text completion. It's particularly designed for the Italian language and can be fine-tuned for specific tasks.

Quantized version

DeepMount00/Mistral-Ita-7b-GGUF

Unique Features for Italian

  • Tailored Vocabulary: The model's vocabulary is fine-tuned to encompass the nuances and diversity of the Italian language.
  • Enhanced Understanding: Mistral-7B is specifically trained to grasp and generate Italian text, ensuring high linguistic and contextual accuracy.

Capabilities

  • Vocabulary Size: 32,000 tokens, allowing for a broad range of inputs and outputs.
  • Hidden Size: 4,096 dimensions, providing rich internal representations.
  • Intermediate Size: 14,336 dimensions, which contributes to the model's ability to process and generate complex sentences.

How to Use

How to utilize my Mistral for Italian text generation

import transformers
from transformers import TextStreamer
import torch

MODEL_NAME = "DeepMount00/Mistral-Ita-7b"

model = AutoModelForCausalLM.from_pretrained(MODEL_NAME, torch_dtype=torch.bfloat16).eval()
tokenizer = AutoTokenizer.from_pretrained(MODEL_NAME)

def generate_answer(prompt):
    encoded_input = tokenizer.apply_chat_template([{"role": "user", "content": prompt}], return_tensors="pt").to(device)
    generated_ids = model.generate(**encoded_input, max_new_tokens=200, do_sample=True, temperature=0.001, eos_token_id=tokenizer.eos_token_id)
    answer = tokenizer.batch_decode(generated_ids, skip_special_tokens=True)
    return answer

prompt = "Se un mattone pesa 1kg più metà di se stesso, quanto pesa il mattone? Rispondi impostando l'equazione matematica"
print(generate_answer(prompt))

Developer

[Michele Montebovi]