Edit model card

AWQ Quantized

!pip install  git+https://github.com/huggingface/transformers.git -q
!pip install huggingface_hub
!pip install autoawq -q
from awq import AutoAWQForCausalLM
from transformers import AutoTokenizer
import torch

# Assuming your model and tokenizer are loaded
model_name_or_path = "arlineka/manbasya_2x7b_MOE"
model = AutoAWQForCausalLM.from_quantized(model_name_or_path, fuse_layer=True, trust_remote_code=False, safetensors=True)
tokenizer = AutoTokenizer.from_pretrained(model_name_or_path, trust_remote_code=False)

# Set device to CUDA if available
device = torch.device("cuda" if torch.cuda.is_available() else "cpu")

# Move model to the device
model.to(device)

# Prepare your input text and move input tensors to the same device
input_text = "Hello. Input Here"
input_ids = tokenizer.encode(input_text, return_tensors="pt").to(device)

# Now generate text with model and input tensors on the same device
output = model.generate(input_ids, max_new_tokens=2048)  # Example usage, adjust as necessary
generated_text = tokenizer.decode(output[0], skip_special_tokens=True)

print(generated_text)
Downloads last month
396
Safetensors
Model size
1.95B params
Tensor type
I32
·
FP16
·
Inference API
This model can be loaded on Inference API (serverless).