mlx-community/JOSIEv4o-8b-stage1-beta-4bit

The Model mlx-community/JOSIEv4o-8b-stage1-beta-4bit was converted to MLX format from Isaak-Carter/JOSIEv4o-8b-stage1-beta1 using mlx-lm version 0.14.0.

Use with mlx

pip install mlx-lm
from mlx_lm import load, generate

model, tokenizer = load("mlx-community/JOSIEv4o-8b-stage1-beta-4bit")
response = generate(model, tokenizer, prompt="hello", verbose=True)

System Prompto and Prompt template used:

<|begin_of_text|>system
You are J.O.S.I.E. which is an acronym for "Just an Outstandingly Smart Intelligent Entity", a private and super-intelligent AI assistant, created by Gökdeniz Gülmez.
<|begin_of_text|>main user "{your_name}"
{{ .Prompt }}<|end_of_text|>
<|begin_of_text|>josie
{{ .Response }}<|end_of_text|>
Downloads last month
80
Safetensors
Model size
1.25B params
Tensor type
FP16
·
U32
·
Inference Providers NEW
This model is not currently available via any of the supported Inference Providers.

Model tree for mlx-community/JOSIEv4o-8b-stage1-beta-4bit

Finetuned
(2595)
this model