eachadea's picture
Update README.md
4d101c5
metadata
pipeline_tag: conversational
tags:
  - vicuna
  - llama
  - text-generation-inference

Converted for use with llama.cpp

  • 4-bit quantized
  • Needs ~6GB of CPU RAM
  • Won't work with alpaca.cpp or old llama.cpp (new ggml format requires latest llama.cpp)
  • 7B parameter version

Bigger 13B version can be found here: https://huggingface.co/eachadea/ggml-vicuna-13b-4bit


tags: - vicuna - llama - text-generation-inference