__init__.py
|
support qwen
|
2023-08-08 19:27:43 +09:00 |
_const.py
|
expose api to set exllama max length
|
2023-08-24 11:22:15 +00:00 |
_utils.py
|
fix type
|
2023-08-31 14:07:16 -04:00 |
auto.py
|
expose disable_exllama argument
|
2023-08-09 12:03:31 +08:00 |
codegen.py
|
Add support for CodeGen/2
|
2023-05-08 17:34:00 +03:00 |
gpt_bigcode.py
|
Add support for GPTBigCode
|
2023-05-08 12:28:29 +03:00 |
gptj.py
|
add GPTJ fused attention module
|
2023-05-14 16:17:21 +08:00 |
internlm.py
|
Add support for InternLM
|
2023-07-07 09:25:40 -07:00 |
llama.py
|
make compatible with older transformers version
|
2023-05-15 13:26:18 +08:00 |
qwen.py
|
Update qwen.py for Qwen-VL
|
2023-08-30 16:29:55 +08:00 |
rw.py
|
support falcon
|
2023-05-27 07:53:39 +09:00 |