tabbyAPI-ollama/common
kingbri 949248fb94 Config: Add experimental torch cuda malloc backend
This option saves some VRAM, but does have the chance to error out.
Add this in the experimental config section.

Signed-off-by: kingbri <bdashore3@proton.me>
2024-02-14 21:45:56 -05:00
..
args.py Config: Add experimental torch cuda malloc backend 2024-02-14 21:45:56 -05:00
auth.py Auth: Create keys on different exception 2024-02-04 01:56:42 -05:00
config.py Launch: Make exllamav2 requirement more friendly 2024-02-02 23:36:17 -05:00
gen_logging.py Tree: Refactor code organization 2024-01-25 00:15:40 -05:00
generators.py Tree: Refactor code organization 2024-01-25 00:15:40 -05:00
logger.py Tree: Refactor code organization 2024-01-25 00:15:40 -05:00
sampling.py Sampling: Make validators simpler 2024-02-11 15:28:43 -05:00
templating.py API: Add template switching and unload endpoints 2024-01-25 00:15:40 -05:00
utils.py Sampling: Cleanup and update 2024-02-02 23:36:17 -05:00