AutoGPTQ/auto_gptq/modeling/llama.py
2023-04-14 01:09:40 +08:00

15 lines
416 B
Python

from ._base import *
class LlamaGPTQForCausalLM(BaseGPTQForCausalLM):
layers_block_name = "model.layers"
outside_layer_modules = ["model.embed_tokens", "model.norm"]
inside_layer_modules = [
["self_attn.k_proj", "self_attn.v_proj", "self_attn.q_proj"],
["self_attn.o_proj"],
["mlp.up_proj", "mlp.gate_proj"],
["mlp.down_proj"]
]
__all__ = ["LlamaGPTQForCausalLM"]