from ._base import * | |
class RWGPTQForCausalLM(BaseGPTQForCausalLM): | |
layer_type = "DecoderLayer" | |
layers_block_name = "transformer.h" | |
outside_layer_modules = ["transformer.word_embeddings", "transformer.ln_f"] | |
inside_layer_modules = [ | |
["self_attention.query_key_value"], | |
["self_attention.dense"], | |
["mlp.dense_h_to_4h"], | |
["mlp.dense_4h_to_h"] | |
] | |
__all__ = ["RWGPTQForCausalLM"] | |