No description
Find a file
TerminalMan 7d18d2e2ca
Refactor the sampling class (#199)
* improve validation

* remove to_gen_params functions

* update changes for all endpoint types

* OAI: Fix calls to generation

Chat completion and completion need to have prompt split out before
pushing to the backend.

Signed-off-by: kingbri <bdashore3@proton.me>

* Sampling: Convert Top-K values of -1 to 0

Some OAI implementations use -1 as disabled instead of 0. Therefore,
add a coalesce case.

Signed-off-by: kingbri <bdashore3@proton.me>

* Sampling: Format and space out

Make the code more readable.

Signed-off-by: kingbri <bdashore3@proton.me>

* Sampling: Fix mirostat

Field items are nested in data within a Pydantic FieldInfo

Signed-off-by: kingbri <bdashore3@proton.me>

* Sampling: Format

Signed-off-by: kingbri <bdashore3@proton.me>

* Sampling: Fix banned_tokens and allowed_tokens conversion

If the provided string has whitespace, trim it before splitting.

Signed-off-by: kingbri <bdashore3@proton.me>

* Sampling: Add helpful log to dry_sequence_breakers

Let the user know if the sequence errors out.

Signed-off-by: kingbri <bdashore3@proton.me>

* Sampling: Apply validators in right order

Validators need to be applied in order from top to bottom, this is why
the after validator was not being applied properly.

Set the model to validate default params for sampler override purposes.
This can be turned off if there are unclear errors.

Signed-off-by: kingbri <bdashore3@proton.me>

* Endpoints: Format

Cleanup and semantically fix field validators

Signed-off-by: kingbri <bdashore3@proton.me>

* Kobold: Update validators and fix parameter application

Validators on parent fields cannot see child fields. Therefore,
validate using the child fields instead and alter the parent field
data from there.

Also fix badwordsids casting.

Signed-off-by: kingbri <bdashore3@proton.me>

* Sampling: Remove validate defaults and fix mirostat

If a user sets an override to a non-default value, that's their
own fault.

Run validator on the actual mirostat_mode parameter rather than
the alternate mirostat parameter.

Signed-off-by: kingbri <bdashore3@proton.me>

* Kobold: Rework badwordsids

Currently, this serves to ban the EOS token. All other functionality
was legacy, so remove it.

Signed-off-by: kingbri <bdashore3@proton.me>

* Model: Remove HuggingfaceConfig

This was only necessary for badwordsids. All other fields are handled
by exl2. Keep the class as a stub if it's needed again.

Signed-off-by: kingbri <bdashore3@proton.me>

* Kobold: Bump kcpp impersonation

TabbyAPI supports XTC now.

Signed-off-by: kingbri <bdashore3@proton.me>

* Sampling: Change alias to validation_alias

Reduces the probability for errors and makes the class consistent.

Signed-off-by: kingbri <bdashore3@proton.me>

* OAI: Use constraints for validation

Instead of adding a model_validator, use greater than or equal to
constraints provided by Pydantic.

Signed-off-by: kingbri <bdashore3@proton.me>

* Tree: Lint

Signed-off-by: kingbri <bdashore3@proton.me>

---------

Co-authored-by: SecretiveShell <84923604+SecretiveShell@users.noreply.github.com>
Co-authored-by: kingbri <bdashore3@proton.me>
2024-10-27 11:43:41 -04:00
.github fix issues with optional dependencies (#204) 2024-09-19 22:24:55 -04:00
backends Refactor the sampling class (#199) 2024-10-27 11:43:41 -04:00
colab Colab: Update 2024-03-24 21:48:48 -04:00
common Refactor the sampling class (#199) 2024-10-27 11:43:41 -04:00
docker debloat docker build 2024-09-08 00:02:00 +01:00
endpoints Refactor the sampling class (#199) 2024-10-27 11:43:41 -04:00
loras Implement lora support (#24) 2023-12-08 23:38:08 -05:00
models Tree: Update documentation and configs 2023-11-16 02:30:33 -05:00
sampler_overrides Sampling: Add XTC support 2024-09-24 18:10:52 -04:00
templates Templates: Remove whitespace from metadata 2024-09-08 12:36:36 -04:00
tests Tree: Format 2024-03-13 00:02:55 -04:00
update_scripts Start: Make linux scripts executable 2024-08-03 15:19:31 -04:00
.dockerignore debloat docker build 2024-09-08 00:02:00 +01:00
.gitignore Update .gitignore 2024-09-16 22:48:13 -04:00
api_tokens_sample.yml Improve docker deployment configuration (#163) 2024-08-18 15:19:18 -04:00
config_sample.yml Tree: Remove fasttensors 2024-09-30 00:18:47 -04:00
formatting.bat feat: workflows for formatting/linting (#35) 2023-12-22 16:20:35 +00:00
formatting.sh feat: workflows for formatting/linting (#35) 2023-12-22 16:20:35 +00:00
LICENSE Create LICENSE 2023-11-16 17:43:23 -05:00
main.py Model: Fix inline loading and draft key (#225) 2024-10-24 23:35:05 -04:00
pyproject.toml Dependencies: Update ExllamaV2 2024-09-30 00:17:12 -04:00
README.md Update README 2024-08-17 00:35:42 -04:00
start.bat Start: Fix startup with new argparser 2024-09-21 14:36:21 -04:00
start.py Tree: Format 2024-09-21 14:37:01 -04:00
start.sh Start: Fix startup with new argparser 2024-09-21 14:36:21 -04:00

TabbyAPI

Python 3.10, 3.11, and 3.12 License: AGPL v3 Discord Server

Developer facing API documentation

Support on Ko-Fi

Important

In addition to the README, please read the Wiki page for information about getting started!

Note

Need help? Join the Discord Server and get the Tabby role. Please be nice when asking questions.

A FastAPI based application that allows for generating text using an LLM (large language model) using the Exllamav2 backend

TabbyAPI is also the official API backend server for ExllamaV2.

Disclaimer

This project is marked as rolling release. There may be bugs and changes down the line. Please be aware that you might need to reinstall dependencies if needed.

TabbyAPI is a hobby project made for a small amount of users. It is not meant to run on production servers. For that, please look at other solutions that support those workloads.

Getting Started

Important

This README does not have instructions for setting up. Please read the Wiki.

Read the Wiki for more information. It contains user-facing documentation for installation, configuration, sampling, API usage, and so much more.

Features

  • OpenAI compatible API
  • Loading/unloading models
  • HuggingFace model downloading
  • Embedding model support
  • JSON schema + Regex + EBNF support
  • AI Horde support
  • Speculative decoding via draft models
  • Multi-lora with independent scaling (ex. a weight of 0.9)
  • Inbuilt proxy to override client request parameters/samplers
  • Flexible Jinja2 template engine for chat completions that conforms to HuggingFace
  • Concurrent inference with asyncio
  • Utilizes modern python paradigms
  • Continuous batching engine using paged attention
  • Fast classifer-free guidance
  • OAI style tool/function calling

And much more. If something is missing here, PR it in!

Supported Model Types

TabbyAPI uses Exllamav2 as a powerful and fast backend for model inference, loading, etc. Therefore, the following types of models are supported:

  • Exl2 (Highly recommended)

  • GPTQ

  • FP16 (using Exllamav2's loader)

In addition, TabbyAPI supports parallel batching using paged attention for Nvidia Ampere GPUs and higher.

Contributing

Use the template when creating issues or pull requests, otherwise the developers may not look at your post.

If you have issues with the project:

  • Describe the issue in detail

  • If you have a feature request, please indicate it as such.

If you have a Pull Request

  • Describe the pull request in detail, what, and why you are changing something

Acknowldgements

TabbyAPI would not exist without the work of other contributors and FOSS projects:

Developers and Permissions

Creators/Developers: