--- library_name: transformers tags: [] --- This is a quantized version of the Jais-13b model If you are using text-generator-webui Select Transformers - Compute d-type: bfloat16 - Quantization Type : nf4 - Load in 4-bit: True - Use double quantization: True ```python from transformers import AutoModelForCausalLM, AutoTokenizer, BitsAndBytesConfig import transformers import torch model_name = "jwnder/core42_jais-13b-bnb-4bit" import warnings warnings.filterwarnings('ignore') tokenizer = AutoTokenizer.from_pretrained(model_input_folder, trust_remote_code=True) model = AutoModelForCausalLM.from_pretrained(model_input_folder, trust_remote_code=True) inputs = tokenizer("Testing LLM!", return_tensors="pt") start = datetime.now() outputs = model.generate(**inputs) end = datetime.now() print(tokenizer.batch_decode(outputs, skip_special_tokens=True)) ```