Text Generation
Transformers
GGUF
code
granite
Eval Results

Error Loading ValueError: Failed to load model from file:

#1
by Zues - opened

I am using latest llama-cpp-python and I keep getting this error

ValueError: Failed to load model from file:
I dabbled with adding n_gqa param but in vain.
Code:

llm = Llama.from_pretrained(
    repo_id="ibm-granite/granite-34b-code-instruct-GGUF",
    filename="granite-34b-code-instruct.Q4_K_M.gguf",
    verbose=False,
    n_gpu_layers=-1,
    # n_ctx=30000,
    n_gqa=4,
)

Sign up or log in to comment