Edit model card

stablelm-tuned-alpha-7b-sharded-8bit

This is a sharded checkpoint (with ~4GB shards) of the stabilityai/stablelm-tuned-alpha-7b model in 8bit precision using bitsandbytes.

Refer to the original model for all details w.r.t. to the model. For more info on loading 8bit models, refer to the example repo and/or the 4.28.0 release info.

  • total model size is only ~7 GB!
  • this enables low-RAM loading, i.e. Colab :)

Basic Usage

Open In Colab

You can use this model as a drop-in replacement in the notebook for the standard sharded models.

Python

Install/upgrade transformers, accelerate, and bitsandbytes. For this to work you must have transformers>=4.28.0 and bitsandbytes>0.37.2.

pip install -U -q transformers bitsandbytes accelerate

Load the model. As it is serialized in 8bit you don't need to do anything special:

from transformers import AutoTokenizer, AutoModelForCausalLM

model_name = "ethzanalytics/stablelm-tuned-alpha-7b-sharded-8bit"
tokenizer = AutoTokenizer.from_pretrained(model_name)

model = AutoModelForCausalLM.from_pretrained(model_name)
Downloads last month
3
Inference Examples
Inference API (serverless) has been turned off for this model.