Model: Fix no flash attention
Was being called wrong from config. Signed-off-by: kingbri <bdashore3@proton.me>
This commit is contained in:
parent
ad8807a830
commit
95fd0f075e
1 changed files with 1 additions and 1 deletions
2
model.py
2
model.py
|
|
@ -94,7 +94,7 @@ class ModelContainer:
|
|||
)
|
||||
|
||||
# Turn off flash attention?
|
||||
self.config.no_flash_attn = unwrap(kwargs.get("no_flash_attn"), False)
|
||||
self.config.no_flash_attn = unwrap(kwargs.get("no_flash_attention"), False)
|
||||
|
||||
# low_mem is currently broken in exllamav2. Don't use it until it's fixed.
|
||||
"""
|
||||
|
|
|
|||
Loading…
Add table
Add a link
Reference in a new issue