tabbyAPI-ollama/common
kingbri 79f9c6e854 Model: Remove num_experts_per_token
This shouldn't even be an exposed option since changing it always
breaks inference with the model. Let the model's config.json handle
it.

Signed-off-by: kingbri <8082010+kingbri1@users.noreply.github.com>
2025-03-19 11:52:10 -04:00
..
actions.py Downloader: log errors when downloading 2025-02-19 23:16:17 -05:00
args.py Args: Expose api-servers to subcommands 2025-02-10 23:39:46 -05:00
auth.py Tree: Format 2024-09-18 20:36:17 -04:00
concurrency.py API + Model: Add blocks and checks for various load requests 2024-05-25 21:16:14 -04:00
config_models.py Model: Remove num_experts_per_token 2025-03-19 11:52:10 -04:00
downloader.py Downloader: log errors when downloading 2025-02-19 23:16:17 -05:00
gen_logging.py Logging: Remove preferences global 2024-09-14 21:49:44 -04:00
health.py Add health check monitoring for EXL2 errors (#206) 2024-09-22 21:40:36 -04:00
logger.py Logger: Switch to ISO timestamp formatting 2025-02-19 21:48:23 -05:00
model.py Model: Fix load packets 2024-11-21 18:06:47 -05:00
multimodal.py Tree: Format 2025-02-07 18:03:33 -05:00
networking.py remove unused imports 2024-09-11 18:00:29 +01:00
optional_dependencies.py Dependencies: Remove outlines from optional check 2024-12-18 11:56:40 -05:00
sampling.py Sampling: Add max_completion_tokens 2024-12-13 01:02:37 -05:00
signals.py Signals: Split signal handler between sync and async 2024-09-19 23:31:29 -04:00
tabby_config.py Cleanup config file loader (#208) 2024-09-23 21:42:01 -04:00
templating.py Tree: Format 2025-02-13 22:42:59 -05:00
transformers_utils.py Refactor the sampling class (#199) 2024-10-27 11:43:41 -04:00
utils.py Cleanup config file loader (#208) 2024-09-23 21:42:01 -04:00