Transformers
PyTorch
Inference Endpoints
Edit model card

Mamba-2.8b-slimpj is a model using the Mamba architecture, with 2.8B parameters, trained for 600B tokens on the SlimPajama dataset.

Model code: https://github.com/state-spaces/mamba/tree/main

To load the model, follow the installation instruction in the code repo, and then:

from mamba_ssm.models.mixer_seq_simple import MambaLMHeadModel
model = MambaLMHeadModel.from_pretrained("state-spaces/mamba-2.8b-slimpj")
Downloads last month
3,718
Inference API
Unable to determine this model’s pipeline type. Check the docs .

Model tree for state-spaces/mamba-2.8b-slimpj

Finetunes
3 models
Merges
1 model

Space using state-spaces/mamba-2.8b-slimpj 1