This website requires JavaScript.
Explore
Help
Sign in
simcop2387
/
AutoGPTQ
Watch
1
Star
0
Fork
You've already forked AutoGPTQ
0
Code
Issues
Pull requests
Projects
Releases
Packages
1
Wiki
Activity
26dc6852fe
AutoGPTQ
/
auto_gptq
/
nn_modules
/
fused_modules
History
PanQiWei
26dc6852fe
support inherit one of the three fused attention class and customize attn_bias building logic
2023-08-07 18:59:04 +08:00
..
__init__.py
add FusedGeneralQuantLinear
2023-08-04 19:10:32 +08:00
attention.py
support inherit one of the three fused attention class and customize attn_bias building logic
2023-08-07 18:59:04 +08:00
linear.py
fix using wrong attribute
2023-08-06 16:23:19 +08:00
mlp.py
doing 'memory_efficient_fusion' in __init__
2023-08-06 17:23:57 +08:00