AutoGPTQ/auto_gptq/modeling/rw.py
2023-05-27 07:53:39 +09:00

15 lines
428 B
Python

from ._base import *
class RWGPTQForCausalLM(BaseGPTQForCausalLM):
layer_type = "DecoderLayer"
layers_block_name = "transformer.h"
outside_layer_modules = ["transformer.word_embeddings", "transformer.ln_f"]
inside_layer_modules = [
["self_attention.query_key_value"],
["self_attention.dense"],
["mlp.dense_h_to_4h"],
["mlp.dense_4h_to_h"]
]
__all__ = ["RWGPTQForCausalLM"]