tabbyAPI-ollama/OAI/types
kingbri ce2602df9a Model: Fix max seq len handling
Previously, the max sequence length was overriden by the user's
config and never took the model's config.json into account.

Now, set the default to 4096, but include config.prepare when
selecting the max sequence length. The yaml and API request
now serve as overrides rather than parameters.

Signed-off-by: kingbri <bdashore3@proton.me>
2023-12-19 23:37:52 -05:00
..
chat_completion.py Templating: Add generation prompt appending 2023-12-18 23:53:47 -05:00
common.py OAI: Add logit bias support 2023-12-18 23:53:47 -05:00
completion.py Adding token usage support 2023-11-27 20:05:05 -08:00
lora.py OAI: Add ability to specify fastchat prompt template 2023-12-10 15:43:58 -05:00
model.py Model: Fix max seq len handling 2023-12-19 23:37:52 -05:00
token.py Api: Add ban_eos_token and add_bos_token support 2023-11-15 00:55:15 -05:00