mlx-community/JOSIEv4o-8b-stage1-beta-4bit
The Model mlx-community/JOSIEv4o-8b-stage1-beta-4bit was converted to MLX format from Isaak-Carter/JOSIEv4o-8b-stage1-beta1 using mlx-lm version 0.14.0.
Use with mlx
pip install mlx-lm
from mlx_lm import load, generate
model, tokenizer = load("mlx-community/JOSIEv4o-8b-stage1-beta-4bit")
response = generate(model, tokenizer, prompt="hello", verbose=True)
System Prompto and Prompt template used:
<|begin_of_text|>system
You are J.O.S.I.E. which is an acronym for "Just an Outstandingly Smart Intelligent Entity", a private and super-intelligent AI assistant, created by Gökdeniz Gülmez.
<|begin_of_text|>main user "{your_name}"
{{ .Prompt }}<|end_of_text|>
<|begin_of_text|>josie
{{ .Response }}<|end_of_text|>
- Downloads last month
- 80
Inference Providers
NEW
This model is not currently available via any of the supported Inference Providers.
Model tree for mlx-community/JOSIEv4o-8b-stage1-beta-4bit
Base model
meta-llama/Meta-Llama-3-8B
Quantized
unsloth/llama-3-8b-bnb-4bit