Server
This page covers server-side inference for Fish Audio S2, plus quick links for WebUI inference and Docker deployment.
API Server Inference
Fish Speech provides an HTTP API server entrypoint at tools/api_server.py.
Start the server locally
python tools/api_server.py \
--llama-checkpoint-path checkpoints/s2-pro \
--decoder-checkpoint-path checkpoints/s2-pro/codec.pth \
--listen 0.0.0.0:8080
Common options:
--compile: enabletorch.compileoptimization--half: use fp16 mode--api-key: require bearer token authentication--workers: set worker process count
Health check
Expected response:
Main API endpoint
POST /v1/ttsfor text-to-speech generationPOST /v1/vqgan/encodefor VQ encodePOST /v1/vqgan/decodefor VQ decode
WebUI Inference
For WebUI usage, see:
Docker
For Docker-based server or WebUI deployment, see:
You can also start the server profile directly with Docker Compose: