diff --git a/TECH-STACK.md b/TECH-STACK.md index 72ee526..a140c7a 100644 --- a/TECH-STACK.md +++ b/TECH-STACK.md @@ -39,7 +39,7 @@ All AI inference runs on a unified Ray Serve endpoint with fractional GPU alloca |---------|-------|----------|----------|------------| | `/llm` | [vLLM](https://vllm.ai) (Llama 3.1 70B) | khelben | AMD Strix Halo 64GB | 0.95 GPU | | `/whisper` | [faster-whisper](https://github.com/guillaumekln/faster-whisper) v3 | elminster | NVIDIA RTX 2070 8GB | 0.5 GPU | -| `/tts` | [XTTS](https://github.com/coqui-ai/TTS) | elminster | NVIDIA RTX 2070 8GB | 0.5 GPU | +| `/tts` | [XTTS](https://github.com/idiap/coqui-ai-TTS) | elminster | NVIDIA RTX 2070 8GB | 0.5 GPU | | `/embeddings` | [BGE-Large](https://huggingface.co/BAAI/bge-large-en-v1.5) | drizzt | AMD Radeon 680M 12GB | 0.8 GPU | | `/reranker` | [BGE-Reranker](https://huggingface.co/BAAI/bge-reranker-large) | danilo | Intel Arc 16GB | 0.8 GPU |