commit
3fb7d1ed1c
4 changed files with 20 additions and 0 deletions
|
@ -12,3 +12,4 @@ from .gpt_bigcode import *
|
|||
from .codegen import *
|
||||
from .baichuan import *
|
||||
from .internlm import *
|
||||
from .qwen import *
|
|
@ -20,6 +20,7 @@ SUPPORTED_MODELS = [
|
|||
"RefinedWeb",
|
||||
"baichuan",
|
||||
"internlm",
|
||||
"qwen",
|
||||
]
|
||||
if compare_transformers_version("v4.28.0", op="ge"):
|
||||
SUPPORTED_MODELS.append("llama")
|
||||
|
|
|
@ -15,6 +15,7 @@ from .rw import RWGPTQForCausalLM
|
|||
from .gpt_bigcode import GPTBigCodeGPTQForCausalLM
|
||||
from .baichuan import BaiChuanGPTQForCausalLM
|
||||
from .internlm import InternLMGPTQForCausalLM
|
||||
from .qwen import QwenGPTQForCausalLM
|
||||
|
||||
GPTQ_CAUSAL_LM_MODEL_MAP = {
|
||||
"bloom": BloomGPTQForCausalLM,
|
||||
|
@ -30,6 +31,7 @@ GPTQ_CAUSAL_LM_MODEL_MAP = {
|
|||
"RefinedWeb": RWGPTQForCausalLM,
|
||||
"baichuan": BaiChuanGPTQForCausalLM,
|
||||
"internlm": InternLMGPTQForCausalLM,
|
||||
"qwen": QwenGPTQForCausalLM,
|
||||
}
|
||||
|
||||
|
||||
|
|
16
auto_gptq/modeling/qwen.py
Normal file
16
auto_gptq/modeling/qwen.py
Normal file
|
@ -0,0 +1,16 @@
|
|||
from ._base import *
|
||||
|
||||
|
||||
class QwenGPTQForCausalLM(BaseGPTQForCausalLM):
|
||||
layer_type = "QWenBlock"
|
||||
layers_block_name = "transformer.h"
|
||||
outside_layer_modules = ["transformer.wte", "transformer.wpe", "transformer.ln_f"]
|
||||
inside_layer_modules = [
|
||||
["attn.c_attn"],
|
||||
["attn.c_proj"],
|
||||
["mlp.w1", "mlp.w2"],
|
||||
["mlp.c_proj"]
|
||||
]
|
||||
|
||||
|
||||
__all__ = ["QwenGPTQForCausalLM"]
|
Loading…
Add table
Reference in a new issue