Try Me! [https://chat.syntithenai.com]
This repository provides a web service that integrates whisper speech to text, coqui text to speech and a language model compatible with GPT4All.
The default web page provides a voice enabled chat web interface for talking to the model.
Streaming is used throughout to minimise latency.
The default docker-compose.yml file is setup to use GPU resources.
https://github.com/syntithenai/whisper-websocket-streaming
https://github.com/syntithenai/coqui-tts-ssl
The UI is hosted on https://syntithenai.github.io/voice2llm/.
It can be configured with an OpenAI api key or URLs to locally hosted services.
The UI is also hosted locally by the service suite that can be started with Docker.
-
(Windows Users) https://docs.docker.com/desktop/gpu/
-
install docker
-
copy the .env.sample file to .env and edit to provide configuration for preferred language model (open ai key, ...)
-
docker-compose up
- map a domain name to your IP address and configure port forwarding from your router to ports 443 (web and STT), 444 (LLM) and 5002(TTS)
- copy the .env.sample file to .env and edit to update the domain name and email address
docker-compose up
cd voice2llm-ui
npm run build
To develop the UI with live updates
cd voice2llm-ui
npm run start
and visit http://localhost:3000
The build script moves the resulting files to the docs file in the root of the project ready for hosting on github.