import os from setuptools import setup, find_packages try: import torch TORCH_AVAILABLE = True except ImportError: TORCH_AVAILABLE = False version = "0.1.0" requirements = [ "accelerate>=0.18.0", "datasets", "numpy", "rouge", "torch>=1.13.0", "safetensors", "transformers>=4.26.1" ] extras_require = { "llama": ["transformers>=4.28.0"], "triton": ["triton>=2.0.0"] } include_dirs=["quant_cuda"] if TORCH_AVAILABLE: BUILD_CUDA_EXT = int(os.environ.get('BUILD_CUDA_EXT', '1')) == 1 additional_setup_kwargs = dict() if BUILD_CUDA_EXT and torch.cuda.is_available(): from torch.utils import cpp_extension from distutils.sysconfig import get_python_lib conda_cuda_include_dir=os.path.join(get_python_lib(),"nvidia/cuda_runtime/include") print(f"checking conda cuda include dir {conda_cuda_include_dir}") if os.path.isdir(conda_cuda_include_dir): include_dirs.append(conda_cuda_include_dir) print(f"appending conda cuda include dir {conda_cuda_include_dir}") extensions = [ cpp_extension.CUDAExtension( "quant_cuda", [ "quant_cuda/quant_cuda.cpp", "quant_cuda/quant_cuda_kernel.cu" ] ) ] additional_setup_kwargs = { "ext_modules": extensions, "cmdclass": {'build_ext': cpp_extension.BuildExtension} } setup( name="auto_gptq", packages=find_packages(), version=version, install_requires=requirements, extras_require=extras_require, include_dirs=include_dirs, **additional_setup_kwargs ) else: setup( name="auto_gptq", packages=find_packages(), version=version, install_requires=requirements, extras_require=extras_require, include_dirs=include_dirs )