mirror of
https://github.com/oobabooga/text-generation-webui.git
synced 2025-06-07 14:17:09 -04:00
ExLlamaV3_HF: Change max_chunk_size to 256
This commit is contained in:
parent
df7bb0db1f
commit
f3da45f65d
1 changed files with 1 additions and 1 deletions
|
@ -119,7 +119,7 @@ class Exllamav3HF(PreTrainedModel, GenerationMixin):
|
||||||
reset = True
|
reset = True
|
||||||
|
|
||||||
# Maximum number of tokens to process in a single forward pass
|
# Maximum number of tokens to process in a single forward pass
|
||||||
max_chunk_size = 2048
|
max_chunk_size = 256
|
||||||
|
|
||||||
# Make the forward call
|
# Make the forward call
|
||||||
if labels is None:
|
if labels is None:
|
||||||
|
|
Loading…
Add table
Reference in a new issue