Command-R 35B v1.0 - GGUF

Description

This repo contains llama.cpp GGUF format model files for Command-R 35B v1.0.

Compatibility

These quantised GGUF files are compatible with llama.cpp from Mar 16, 2024 onwards, starting from release b2440

F16 files are split and require joining

Note: Hugging face does not support uploading files larger than 50GB so I uploaded the GGUF as 2 split files.

To join the files, run the following:

Linux and macOS:

cat c4ai-command-r-v01-f16.gguf-split-* > c4ai-command-r-v01-f16.gguf

Then you can remove the split files to save space:

rm c4ai-command-r-v01-f16.gguf-split-*

Windows command line:

COPY /B c4ai-command-r-v01-f16.gguf-split-a + c4ai-command-r-v01-f16.gguf-split-b c4ai-command-r-v01-f16.gguf

Then you can remove the split files to save space:

del c4ai-command-r-v01-f16.gguf-split-a c4ai-command-r-v01-f16.gguf-split-b

You can optionally confirm the checksum of merged c4ai-command-r-v01-f16.gguf with the md5sum file:

md5sum -c md5sum
Downloads last month
1,992
GGUF
Model size
35B params
Architecture
command-r

2-bit

3-bit

4-bit

5-bit

6-bit

8-bit

Inference API
Unable to determine this model's library. Check the docs .