mgonzs13's picture
Update README.md
75d09cc verified
metadata
base_model: BarraHome/Mistroll-7B-v2.2
license: mit
language:
  - en
  - es
pipeline_tag: text-generation
tags:
  - mistral
  - unsloth
  - gguf
library_name: llama.cpp
model_creator: BarraHome
model_name: Mistroll 7B v2.2
model_type: mistral
prompt_template: |
  <|im_start|>system
  {system_message}<|im_end|>
  <|im_start|>user
  {prompt}<|im_end|>
  <|im_start|>assistant
quantized_by: mgonzs13

Mistroll-7B-v2.2-GGUF

Model creator: BarraHome
Original model: Mistroll-7B-v2.2
GGUF quantization: llama.cpp commit 6e472f58e40cd4acf6023e15c75a2700535c5f0b

Description

This model was trained 2x faster with Unsloth and Huggingface's TRL library.

This experiment serves to test and refine a specific training and evaluation pipeline research framework. Its primary objective is to identify potential optimizations, with a focus on data engineering, architectural efficiency, and evaluation performance.

The goal of this experiment is to evaluate the effectiveness of a new training and evaluation pipeline for Large Language Models (LLMs). To achieve this, we will explore adjustments in data preprocessing, model training algorithms, and evaluation metrics to test methods for improvement.

Prompt Template

Following the Mistroll chat template, the prompt template is ChatML.

<|im_start|>system
{system_message}<|im_end|>
<|im_start|>user
{prompt}<|im_end|>
<|im_start|>assistant