mirror of
https://github.com/oobabooga/text-generation-webui.git
synced 2025-06-07 14:17:09 -04:00
ExLlamaV3_HF: Change max_chunk_size to 256
This commit is contained in:
parent
df7bb0db1f
commit
f3da45f65d
1 changed files with 1 additions and 1 deletions
|
@ -119,7 +119,7 @@ class Exllamav3HF(PreTrainedModel, GenerationMixin):
|
|||
reset = True
|
||||
|
||||
# Maximum number of tokens to process in a single forward pass
|
||||
max_chunk_size = 2048
|
||||
max_chunk_size = 256
|
||||
|
||||
# Make the forward call
|
||||
if labels is None:
|
||||
|
|
Loading…
Add table
Reference in a new issue