Chocolatine-3B-Instruct-DPO-v1.2-Q4_K_M-GGUF

Quantized q4_k_m GGUF version of the original model jpacifico/Chocolatine-3B-Instruct-DPO-v1.2
can be used on a CPU device, compatible llama.cpp
now supported architecture by LM Studio.
Also ready for Raspberry Pi 5 8Gb.

The model supports 128K context length.

Ollama

jpacifico/chocolatine-3b

Usage:

ollama run jpacifico/chocolatine-3b

Ollama Modelfile example :

FROM ./chocolatine-3b-instruct-dpo-v1.2-q4_k_m.gguf
TEMPLATE """{{ if .System }}<|system|>
{{ .System }}<|end|>
{{ end }}{{ if .Prompt }}<|user|>
{{ .Prompt }}<|end|>
{{ end }}<|assistant|>
{{ .Response }}<|end|>
"""
PARAMETER stop """{"stop": ["<|end|>","<|user|>","<|assistant|>"]}"""
SYSTEM """You are a friendly assistant called Chocolatine."""

Limitations

The Chocolatine model is a quick demonstration that a base model can be easily fine-tuned to achieve compelling performance.
It does not have any moderation mechanism.

  • Developed by: Jonathan Pacifico, 2024
  • Model type: LLM
  • Language(s) (NLP): French, English
  • License: MIT
Downloads last month
10
GGUF
Model size
3.82B params
Architecture
phi3

4-bit

Inference Examples
This model does not have enough activity to be deployed to Inference API (serverless) yet. Increase its social visibility and check back later, or deploy to Inference Endpoints (dedicated) instead.

Model tree for jpacifico/Chocolatine-3B-Instruct-DPO-v1.2-Q4_K_M-GGUF

Quantized
(5)
this model

Dataset used to train jpacifico/Chocolatine-3B-Instruct-DPO-v1.2-Q4_K_M-GGUF

Collection including jpacifico/Chocolatine-3B-Instruct-DPO-v1.2-Q4_K_M-GGUF