AI Voice system with Whisper transcription.
Before running the server, you need to download the Whisper model. You can use the provided downloader tool:
#Install uv from https://docs.astral.sh/uv/getting-started/installation/
cd tools/model_downloader
# Install requirements (uv is recommended)
uv sync
# Download the default small model
uv run main.py --model openai/whisper-smallThe model will be downloaded to the root directory as whisper-small.
The bolt.sh script provides a convenient way to manage the whole project.
Install all necessary dependencies (Rust, Protobuf, etc.):
./bolt.sh setup./bolt.sh serve./bolt.sh launchThe frontend will be available at http://localhost:8080.
- Docker and Docker Compose installed.
- NVIDIA Container Toolkit (for GPU support).
By default, the backend uses whisper-tiny.en for fast startup.
docker-compose up --buildYou can specify the model using the MODEL_NAME environment variable. Ensure the model is already downloaded in your root directory.
# Example: Running with whisper-small
MODEL_NAME=whisper-small docker-compose up --buildThe system will be accessible via Caddy at http://localhost.
Modify Caddyfile to enable HTTPS and point to the domain.