انتقل إلى المحتوى

Server

This page covers server-side inference for Fish Audio S2, plus quick links for WebUI inference and Docker deployment.

API Server Inference

Fish Speech provides an HTTP API server entrypoint at tools/api_server.py.

Start the server locally

python tools/api_server.py \
  --llama-checkpoint-path checkpoints/s2-pro \
  --decoder-checkpoint-path checkpoints/s2-pro/codec.pth \
  --listen 0.0.0.0:8080

Common options:

  • --compile: enable torch.compile optimization
  • --half: use fp16 mode
  • --api-key: require bearer token authentication
  • --workers: set worker process count

Health check

curl -X GET http://127.0.0.1:8080/v1/health

Expected response:

{"status":"ok"}

Main API endpoint

  • POST /v1/tts for text-to-speech generation
  • POST /v1/vqgan/encode for VQ encode
  • POST /v1/vqgan/decode for VQ decode

WebUI Inference

For WebUI usage, see:

Docker

For Docker-based server or WebUI deployment, see:

You can also start the server profile directly with Docker Compose:

docker compose --profile server up