| (docker_images)= |
| ## Docker images |
| We provide docker images to be able to test TTS without having to setup your own environment. |
| |
| ### Using premade images |
| You can use premade images built automatically from the latest TTS version. |
| |
| #### CPU version |
| ```bash |
| docker pull ghcr.io/coqui-ai/tts-cpu |
| ``` |
| #### GPU version |
| ```bash |
| docker pull ghcr.io/coqui-ai/tts |
| ``` |
| |
| ### Building your own image |
| ```bash |
| docker build -t tts . |
| ``` |
| |
| ## Basic inference |
| Basic usage: generating an audio file from a text passed as argument. |
| You can pass any tts argument after the image name. |
| |
| ### CPU version |
| ```bash |
| docker run --rm -v ~/tts-output:/root/tts-output ghcr.io/coqui-ai/tts-cpu --text "Hello." --out_path /root/tts-output/hello.wav |
| ``` |
| ### GPU version |
| For the GPU version, you need to have the latest NVIDIA drivers installed. |
| With `nvidia-smi` you can check the CUDA version supported, it must be >= 11.8 |
| |
| ```bash |
| docker run --rm --gpus all -v ~/tts-output:/root/tts-output ghcr.io/coqui-ai/tts --text "Hello." --out_path /root/tts-output/hello.wav --use_cuda true |
| ``` |
| |
| ## Start a server |
| Starting a TTS server: |
| Start the container and get a shell inside it. |
| |
| ### CPU version |
| ```bash |
| docker run --rm -it -p 5002:5002 --entrypoint /bin/bash ghcr.io/coqui-ai/tts-cpu |
| python3 TTS/server/server.py --list_models #To get the list of available models |
| python3 TTS/server/server.py --model_name tts_models/en/vctk/vits |
| ``` |
| |
| ### GPU version |
| ```bash |
| docker run --rm -it -p 5002:5002 --gpus all --entrypoint /bin/bash ghcr.io/coqui-ai/tts |
| python3 TTS/server/server.py --list_models #To get the list of available models |
| python3 TTS/server/server.py --model_name tts_models/en/vctk/vits --use_cuda true |
| ``` |
| |
| Click [there](http://[::1]:5002/) and have fun with the server! |