LaaZa
9fd558f2ba
Add support for Baichuan
2023-06-18 20:13:29 +03:00
Angainor Development
e75611e1b7
Fix error message
2023-06-05 22:19:09 +02:00
lunar
618a5f50ee
Add transpose operator when replace Conv1d with qlinear_cuda_old
2023-06-05 23:11:18 +08:00
潘其威(William)
023bb1c593
Merge pull request #125 from PanQiWei/support-32dim
...
Support 32dim
2023-06-03 19:08:29 +08:00
qwopqwop200
f4820f2988
change qlinear cuda support 64dim
2023-06-03 07:30:34 +09:00
潘其威(William)
b4fdd8d264
Merge branch 'main' into peft_integration
2023-06-02 19:11:59 +08:00
qwopqwop200
2df7d7105d
support 64 cuda dim
2023-06-02 19:54:37 +09:00
qwopqwop200
b03f53294f
support 64dim cuda
2023-06-02 19:53:50 +09:00
qwopqwop200
0891ea4036
support 32dim triton]
2023-06-02 19:05:55 +09:00
qwopqwop200
b3654a68c3
support 32dim triton kernel
2023-06-02 19:04:12 +09:00
PanQiWei
ec6603d0ab
support older version python
2023-05-31 22:11:16 +08:00
qwopqwop200
b1a8cc28e8
remove raise
2023-05-31 00:03:51 +09:00
qwopqwop200
c381958a5f
add warning
2023-05-30 23:53:33 +09:00
qwopqwop200
0f2841cb13
remove log
2023-05-30 23:51:55 +09:00
qwopqwop200
33809a8e59
remove log
2023-05-30 23:51:39 +09:00
qwopqwop200
dfd9dc0e6b
change if trainable backend pytorch
2023-05-30 23:43:55 +09:00
qwopqwop200
5274313067
change if trainable backend pytorch
2023-05-30 23:40:58 +09:00
潘其威(William)
defc96ff04
Merge pull request #91 from TheBloke/TheBloke_support-HF-download
...
Add support for HF Hub download, and `push_to_hub`
2023-05-30 07:37:15 +08:00
潘其威(William)
2245fad095
Update auto.py
...
fix None type error
2023-05-30 07:35:15 +08:00
潘其威(William)
15db2cdc44
Update _base.py
...
fix problem that recursively adding file extension to model_base_name
2023-05-30 07:26:42 +08:00
潘其威(William)
cfa7271617
Update _base.py
...
fix variable not found error
2023-05-30 07:22:10 +08:00
潘其威(William)
e5771fb206
Update _base.py
...
fix key mismatch
2023-05-30 06:44:45 +08:00
潘其威(William)
61a4ea035f
Update auto.py
...
add back save_dir for backward compatible
2023-05-30 06:43:00 +08:00
潘其威(William)
ea74e15199
Update _base.py
...
add model_name_or_path and model_file_base_name to BaseQuantizeConfig for better model file management; add back save_dir to .from_quantized() for backward compatible
2023-05-30 06:40:31 +08:00
PanQiWei
6c64b0b361
raise NotImplementedError when model with fused attention injected try to use ADAPTION_PROMPT peft type
2023-05-28 22:35:34 +08:00
PanQiWei
def084bf0e
reset value of AdaptionPromptConfig.adapter_layers to number of model's hidden layers when exceeds
2023-05-28 22:11:02 +08:00
PanQiWei
ad10c13d40
support AdaLora
2023-05-28 21:30:45 +08:00
PanQiWei
3ee2daa73c
make GPTQLoraModel to inherit from LoraModel to simplify code
2023-05-28 17:36:18 +08:00
PanQiWei
22d1d8dcaa
add 'auto_find_all_linears' argument to get_gptq_peft_model function
2023-05-28 17:04:38 +08:00
PanQiWei
83132a663a
add warning to guide users interact with lora properly
2023-05-28 16:57:31 +08:00
PanQiWei
86f060c74b
Merge branch 'main' into peft_integration
2023-05-28 16:23:38 +08:00
PanQiWei
491da62402
fix signature at import time
2023-05-27 17:49:58 +08:00
潘其威(William)
23998345f5
Merge branch 'main' into falcon
2023-05-27 16:23:16 +08:00
Bill Cai
0729760234
Update auto.py
2023-05-27 11:16:43 +08:00
潘其威(William)
269ef7335c
Merge branch 'main' into falcon
2023-05-27 08:15:52 +08:00
潘其威(William)
3c3b0e1e79
Merge branch 'main' into GPTBigCode
2023-05-27 08:03:03 +08:00
潘其威(William)
eab728b263
Merge branch 'main' into Codegen
2023-05-27 08:00:19 +08:00
潘其威(William)
f6fd314d5a
Merge branch 'main' into GPTBigCode
2023-05-27 07:57:25 +08:00
qwopqwop200
277809381b
fix bug
2023-05-27 08:53:47 +09:00
PanQiWei
5bc5325920
add find_all_linear_names help function, make customized lora module more general
2023-05-27 07:49:17 +08:00
PanQiWei
eb9c0b140f
update FusedLlamaMLPForQuantizedModel for general usage purpose
2023-05-27 07:47:20 +08:00
qwopqwop200
bcb345fb35
support falcon
2023-05-27 07:53:39 +09:00
qwopqwop200
4d5b4fa5c6
add dtype
2023-05-27 07:49:28 +09:00
qwopqwop200
c14b4c1567
change find layer algorithm
2023-05-27 07:48:50 +09:00
PanQiWei
f7e705848a
move peft compatible model injection to the last step
2023-05-26 14:29:33 +08:00
PanQiWei
8bf21a7e4c
set xavier_uniform_ as lora_A's init function
2023-05-26 14:06:53 +08:00
PanQiWei
2b532f9453
add trainable mode
2023-05-26 13:11:30 +08:00
PanQiWei
fe5f5d12ed
Merge branch 'main' into peft_integration
2023-05-26 09:48:06 +08:00
PanQiWei
69609c4bc7
support faster vecquant4matmul cuda kernel
2023-05-26 08:55:05 +08:00
PanQiWei
cfd27e8caa
refactor file structure of qlinears
2023-05-26 07:18:16 +08:00