-
-
Notifications
You must be signed in to change notification settings - Fork 726
Open
Description
Describe the bug
Docker command: docker run --gpus all -p 8880:8880 ghcr.io/remsky/kokoro-fastapi-gpu
Kokoro falls back to using CPU. This is fine, but I wanted to use my GPU. As I understand it, CUDA is somewhat backwards compatible, so my CUDA 13.0 system should run a CUDA 12.8 program fine, yes?
I was able to install and use Open-WebUI and its bundled Ollama with my GPU just fine.
Screenshots or console output
2025-10-30 01:23:43.851 | INFO | __main__:download_model:60 - Model files already exist and are valid
/app/.venv/lib/python3.10/site-packages/torch/cuda/__init__.py:734: UserWarning: Can't initialize NVML
warnings.warn("Can't initialize NVML")
INFO: Started server process [30]
Branch / Deployment used
docker local on a headless machine
Docker command: docker run --gpus all -p 8880:8880 ghcr.io/remsky/kokoro-fastapi-gpu
Operating System
me@my-box:~> cat /etc/os-release
NAME="openSUSE Leap"
VERSION="15.6"
ID="opensuse-leap"
ID_LIKE="suse opensuse"
#Linux kernel version
me@my-box:~> uname -rp
6.4.0-150600.23.73-default x86_64
me@my-box:~> sudo nvidia-smi -q
==============NVSMI LOG==============
Timestamp : Wed Oct 29 18:52:11 2025
Driver Version : 580.95.05
CUDA Version : 13.0
Attached GPUs : 1
GPU 00000000:01:00.0
Product Name : Quadro K620
Product Brand : Quadro
Product Architecture : Maxwell
me@my-box:~> docker --version
Docker version 28.3.3-ce, build bea959c7b
Additional context
Installed NVIDIA proprietary driver and NVIDIA-container-toolkit using https://en.opensuse.org/SDB:NVIDIA_drivers
Metadata
Metadata
Assignees
Labels
No labels