Request

#1
by HR1777 - opened

Dear Maziyar,
Could you please make the GGUF version of this promising model? https://huggingface.co/Arist12/eabf-llama2-7b-chat-3.5k

Hi @HR1777

I tried, the Llama.cpp doesn't support eabf ROPE.

if rope_scaling is not None and (typ := rope_scaling.get("type")):
            rope_factor = rope_scaling.get("factor")
            f_rope_scale = rope_factor
            if typ == "linear":
                rope_scaling_type = gguf.RopeScalingType.LINEAR
            elif typ == "yarn":
                rope_scaling_type = gguf.RopeScalingType.YARN
                n_orig_ctx = rope_scaling['original_max_position_embeddings']
                rope_finetuned = rope_scaling['finetuned']
            else:
                raise NotImplementedError(f'Unknown rope scaling type: {typ}')

I think we should open a ticket on llama.cpp asking to add support for Entropy-ABF type:

    raise NotImplementedError(f'Unknown rope scaling type: {typ}')
NotImplementedError: Unknown rope scaling type: eabf

Thank you so much. I will submit it and i hope they merge it!

HR1777 changed discussion status to closed

Sign up or log in to comment