AutoGPTQ/auto_gptq/modeling/_const.py

22 lines
340 B
Python

from torch import device
CPU = device("cpu")
CUDA_0 = device("cuda:0")
SUPPORTED_MODELS = [
"bloom",
"gptj",
"gpt2",
"gpt_neox",
"opt",
"moss",
"gpt_bigcode",
"codegen",
"RefinedWebModel",
"RefinedWeb",
"baichuan",
"internlm",
"llama"
]
__all__ = ["CPU", "CUDA_0", "SUPPORTED_MODELS"]