Start: Migrate options from cu121/118 to cu12

This encapsulates more cuda versions and makes install easier for
new users.

Signed-off-by: kingbri <8082010+kingbri1@users.noreply.github.com>
This commit is contained in:
kingbri 2025-08-19 22:25:30 -04:00
parent 1344726936
commit 30a3cd75cf
6 changed files with 35 additions and 41 deletions

View file

@ -56,7 +56,7 @@
"%cd tabbyAPI\n", "%cd tabbyAPI\n",
"\n", "\n",
"# Install cuda requirements\n", "# Install cuda requirements\n",
"!pip install .[cu121] -q\n", "!pip install .[cu12] -q\n",
"!pip install huggingface-hub -q\n", "!pip install huggingface-hub -q\n",
"\n", "\n",
"# Download cloudflared tunnel\n", "# Download cloudflared tunnel\n",

View file

@ -26,8 +26,8 @@ WORKDIR /app
# Get requirements # Get requirements
COPY pyproject.toml . COPY pyproject.toml .
# Install packages specified in pyproject.toml cu121, extras # Install packages specified in pyproject.toml cu12, extras
RUN pip install --no-cache-dir .[cu121,extras] RUN pip install --no-cache-dir .[cu12,extras]
RUN rm pyproject.toml RUN rm pyproject.toml

View file

@ -47,7 +47,7 @@ To get started, make sure you have the following installed on your system:
1. On Windows: `.\venv\Scripts\activate` 1. On Windows: `.\venv\Scripts\activate`
2. On Linux: `source venv/bin/activate` 2. On Linux: `source venv/bin/activate`
3. Install the pyproject features based on your system: 3. Install the pyproject features based on your system:
1. Cuda 12.x: `pip install -U .[cu121]` 1. Cuda 12.x: `pip install -U .[cu12]`
2. ROCm 5.6: `pip install -U .[amd]` 2. ROCm 5.6: `pip install -U .[amd]`
4. Start the API by either 4. Start the API by either
1. Run `start.bat/sh`. The script will check if you're in a conda environment and skip venv checks. 1. Run `start.bat/sh`. The script will check if you're in a conda environment and skip venv checks.
@ -98,7 +98,7 @@ There are a couple ways to update TabbyAPI:
These scripts exit after running their respective tasks. To start TabbyAPI, run `start.bat` or `start.sh`. These scripts exit after running their respective tasks. To start TabbyAPI, run `start.bat` or `start.sh`.
2. **Manual** - Install the pyproject features and update dependencies depending on your GPU: 2. **Manual** - Install the pyproject features and update dependencies depending on your GPU:
1. `pip install -U .[cu121]` = CUDA 12.x 1. `pip install -U .[cu12]` = CUDA 12.x
2. `pip install -U .[amd]` = ROCm 6.0 2. `pip install -U .[amd]` = ROCm 6.0
If you don't want to update dependencies that come from wheels (torch, exllamav2, and flash attention 2), use `pip install .` or pass the `--nowheel` flag when invoking the start scripts. If you don't want to update dependencies that come from wheels (torch, exllamav2, and flash attention 2), use `pip install .` or pass the `--nowheel` flag when invoking the start scripts.
@ -121,7 +121,7 @@ NOTE:
Here are ways to install exllamav2: Here are ways to install exllamav2:
1. From a [wheel/release](https://github.com/turboderp/exllamav2#method-2-install-from-release-with-prebuilt-extension) (Recommended) 1. From a [wheel/release](https://github.com/turboderp/exllamav2#method-2-install-from-release-with-prebuilt-extension) (Recommended)
1. Find the version that corresponds with your cuda and python version. For example, a wheel with `cu121` and `cp311` corresponds to CUDA 12.1 and python 3.11 1. Find the version that corresponds with your cuda and python version. For example, a wheel with `cu12` and `cp311` corresponds to CUDA 12.1 and python 3.11
2. From [pip](https://github.com/turboderp/exllamav2#method-3-install-from-pypi): `pip install exllamav2` 2. From [pip](https://github.com/turboderp/exllamav2#method-3-install-from-pypi): `pip install exllamav2`
2. This is a JIT compiled extension, which means that the initial launch of tabbyAPI will take some time. The build may also not work due to improper environment configuration. 2. This is a JIT compiled extension, which means that the initial launch of tabbyAPI will take some time. The build may also not work due to improper environment configuration.
3. From [source](https://github.com/turboderp/exllamav2#method-1-install-from-source) 3. From [source](https://github.com/turboderp/exllamav2#method-1-install-from-source)

View file

@ -151,7 +151,7 @@ def entrypoint(
"Or you can manually run a requirements update " "Or you can manually run a requirements update "
"using the following command:\n\n" "using the following command:\n\n"
"For CUDA 12.1:\n" "For CUDA 12.1:\n"
"pip install --upgrade .[cu121]\n\n" "pip install --upgrade .[cu12]\n\n"
"For ROCm:\n" "For ROCm:\n"
"pip install --upgrade .[amd]\n\n" "pip install --upgrade .[amd]\n\n"
) )

View file

@ -56,7 +56,7 @@ extras = [
dev = [ dev = [
"ruff == 0.11.10" "ruff == 0.11.10"
] ]
cu121 = [ cu12 = [
# Torch (Extra index URLs not support in pyproject.toml) # Torch (Extra index URLs not support in pyproject.toml)
"torch @ https://download.pytorch.org/whl/cu128/torch-2.8.0%2Bcu128-cp313-cp313-win_amd64.whl ; platform_system == 'Windows' and python_version == '3.13'", "torch @ https://download.pytorch.org/whl/cu128/torch-2.8.0%2Bcu128-cp313-cp313-win_amd64.whl ; platform_system == 'Windows' and python_version == '3.13'",
"torch @ https://download.pytorch.org/whl/cu128/torch-2.8.0%2Bcu128-cp312-cp312-win_amd64.whl ; platform_system == 'Windows' and python_version == '3.12'", "torch @ https://download.pytorch.org/whl/cu128/torch-2.8.0%2Bcu128-cp312-cp312-win_amd64.whl ; platform_system == 'Windows' and python_version == '3.12'",

View file

@ -41,14 +41,13 @@ def get_user_choice(question: str, options_dict: dict):
def get_install_features(lib_name: str = None): def get_install_features(lib_name: str = None):
"""Fetches the appropriate requirements file depending on the GPU""" """Fetches the appropriate requirements file depending on the GPU"""
install_features = None install_features = None
possible_features = ["cu121", "cu118", "amd"] possible_features = ["cu12", "amd"]
if not lib_name: if not lib_name:
# Ask the user for the GPU lib # Ask the user for the GPU lib
gpu_lib_choices = { gpu_lib_choices = {
"A": {"pretty": "NVIDIA Cuda 12.x", "internal": "cu121"}, "A": {"pretty": "NVIDIA Cuda 12.x", "internal": "cu12"},
"B": {"pretty": "NVIDIA Cuda 11.8 (Unsupported)", "internal": "cu118"}, "B": {"pretty": "AMD", "internal": "amd"},
"C": {"pretty": "AMD", "internal": "amd"},
} }
user_input = get_user_choice( user_input = get_user_choice(
"Select your GPU. If you don't know, select Cuda 12.x (A)", "Select your GPU. If you don't know, select Cuda 12.x (A)",
@ -79,7 +78,7 @@ def get_install_features(lib_name: str = None):
if platform.system() == "Windows": if platform.system() == "Windows":
print( print(
"ERROR: TabbyAPI does not support AMD and Windows. " "ERROR: TabbyAPI does not support AMD and Windows. "
"Please use Linux and ROCm 6.0. Exiting." "Please use Linux and ROCm 6.4. Exiting."
) )
sys.exit(0) sys.exit(0)
@ -139,24 +138,17 @@ def add_start_args(parser: argparse.ArgumentParser):
) )
def migrate_gpu_lib(): def migrate_start_options(start_options: dict):
gpu_lib_path = pathlib.Path("gpu_lib.txt") migrated = False
if not gpu_lib_path.exists(): # Migrate gpu_lib key
return gpu_lib = start_options.get("gpu_lib")
if (gpu_lib == "cu121" or gpu_lib == "cu118"):
print("GPU lib key is legacy, migrating to cu12")
start_options["gpu_lib"] = "cu12"
migrated = True
print("Migrating gpu_lib.txt to the new start_options.json") return migrated
with open("gpu_lib.txt", "r") as gpu_lib_file:
start_options["gpu_lib"] = gpu_lib_file.readline().strip()
start_options["first_run_done"] = True
# Remove the old file
gpu_lib_path.unlink()
print(
"Successfully migrated gpu lib options to start_options. "
"The old file has been deleted."
)
if __name__ == "__main__": if __name__ == "__main__":
@ -183,6 +175,7 @@ if __name__ == "__main__":
add_start_args(parser) add_start_args(parser)
args, _ = parser.parse_known_args() args, _ = parser.parse_known_args()
script_ext = "bat" if platform.system() == "Windows" else "sh" script_ext = "bat" if platform.system() == "Windows" else "sh"
do_start_options_write = False
start_options_path = pathlib.Path("start_options.json") start_options_path = pathlib.Path("start_options.json")
if start_options_path.exists(): if start_options_path.exists():
@ -190,6 +183,7 @@ if __name__ == "__main__":
start_options = json.load(start_options_file) start_options = json.load(start_options_file)
print("Loaded your saved preferences from `start_options.json`") print("Loaded your saved preferences from `start_options.json`")
do_start_options_write = migrate_start_options(start_options)
if start_options.get("first_run_done"): if start_options.get("first_run_done"):
first_run = False first_run = False
else: else:
@ -198,9 +192,6 @@ if __name__ == "__main__":
"Getting things ready..." "Getting things ready..."
) )
# Migrate from old setting storage
migrate_gpu_lib()
# Set variables that rely on start options # Set variables that rely on start options
first_run = not start_options.get("first_run_done") first_run = not start_options.get("first_run_done")
@ -240,15 +231,7 @@ if __name__ == "__main__":
start_options["first_run_done"] = True start_options["first_run_done"] = True
# Save start options on first run # Save start options on first run
with open("start_options.json", "w") as start_file: do_start_options_write = True
start_file.write(json.dumps(start_options))
print(
"Successfully wrote your start script options to "
"`start_options.json`. \n"
"If something goes wrong, editing or deleting the file "
"will reinstall TabbyAPI as a first-time user."
)
if args.update_deps: if args.update_deps:
print( print(
@ -262,6 +245,17 @@ if __name__ == "__main__":
"inside the `update_scripts` folder." "inside the `update_scripts` folder."
) )
if do_start_options_write:
with open("start_options.json", "w") as start_file:
start_file.write(json.dumps(start_options))
print(
"Successfully wrote your start script options to "
"`start_options.json`. \n"
"If something goes wrong, editing or deleting the file "
"will reinstall TabbyAPI as a first-time user."
)
# Expand the parser if it's not fully created # Expand the parser if it's not fully created
if not has_full_parser: if not has_full_parser:
from common.args import init_argparser from common.args import init_argparser