qwopqwop200
|
8c7c806d36
|
if exllama auto diable fused attention
|
2023-08-07 19:24:16 +09:00 |
|
qwopqwop200
|
11afc47f7f
|
support gqa
|
2023-08-07 19:00:05 +09:00 |
|
qwopqwop200
|
2f48780165
|
fix bug disable exlllama
|
2023-08-07 16:28:30 +09:00 |
|
Felix Marty
|
d0608b09db
|
rocm support
|
2023-08-04 13:38:02 +00:00 |
|
Felix Marty
|
38447262c0
|
fix fused attn
|
2023-07-31 13:46:32 +00:00 |
|
Felix Marty
|
179776bd1d
|
exllama kernel
|
2023-07-31 11:50:45 +00:00 |
|
PanQiWei
|
2b532f9453
|
add trainable mode
|
2023-05-26 13:11:30 +08:00 |
|
PanQiWei
|
e1c564ac0e
|
compatible with older pytorch version
|
2023-05-14 16:17:03 +08:00 |
|
PanQiWei
|
2273f9ef39
|
refactor file structure for triton kernels
|
2023-05-14 11:49:10 +08:00 |
|
PanQiWei
|
fef1a4fe4b
|
make code clean and extendable
|
2023-05-12 20:11:55 +08:00 |
|