Commit graph

9 commits

Author SHA1 Message Date
PanQiWei
db63c0876a half out 2023-05-23 16:08:28 +08:00
TheBloke
898f1ef62d Rename 'quant_cuda' to 'autogptq_cuda' to avoid conflicts with existing GPTQ-for-LLaMa installations. 2023-05-20 09:33:51 +01:00
PanQiWei
4bb10fda49 groupsize -> group_size 2023-05-12 13:37:52 +08:00
qwopqwop200
3ff6ab18cb
Merge branch 'main' into faster-llama 2023-05-06 00:20:29 +09:00
TheBloke
1b3329b399 Fix 'groupsize' -> 'group_size' in all other .py files. I haven't touched any CUDA kernels in case there's any complexity there I don't understand 2023-05-05 14:44:16 +01:00
PanQiWei
6cba6e7123 reformat code 2023-05-04 22:16:08 +08:00
qwopqwop200
c359f672a8
support faster and model load strict 2023-05-04 09:04:07 +09:00
qwopqwop200
f0f37c1fe7
fix bug 2023-05-01 18:09:39 +09:00
qwopqwop200
9dfcac8e26
add qlinear_old 2023-05-01 13:03:57 +09:00