abacaj's picture
Update README.md
f1b6814
---
license: other
---
---
This is a [ggml](https://github.com/ggerganov/ggml/) quantized version of [Replit-v2-CodeInstruct-3B](https://huggingface.co/teknium/Replit-v2-CodeInstruct-3B). Quantized to 4bit -> q4_1.
To run inference you can use ggml directly or [ctransformers](https://github.com/marella/ctransformers).
- Memory usage of model: **2GB~**
- Repo to run the model using ctransformers: https://github.com/abacaj/replit-3B-inference