96cd33732d
Fix audio assembly - chunks are already bytes from SNAC decoder
2026-01-11 19:47:19 -06:00
fe43eda6bd
Fix token extraction - use regex to find custom_token patterns
2026-01-11 19:33:31 -06:00
af35dc46d5
Use sync vllm.LLM instead of AsyncLLMEngine to avoid event loop conflicts
2026-01-11 18:58:12 -06:00
0b88188907
Debug: add verbose logging to generate_speech_sync
2026-01-11 18:44:07 -06:00
4eab3ccc01
Fix: wrap sync generator in executor, not async for
2026-01-11 18:32:06 -06:00
4d11334f33
Fix async iteration over vLLM generator - use async for instead of sync for
2026-01-11 18:18:37 -06:00
a164bed590
Fix _map_model_params call signature
2026-01-11 17:59:49 -06:00
d0d7633a00
Monkey-patch OrpheusModel to support max_model_len on Jetson
2026-01-11 17:52:33 -06:00
0e43b76204
Use GitHub orpheus-tts (supports max_model_len) to fix OOM on Jetson
2026-01-11 17:39:55 -06:00
86cf77d2d9
Add HuggingFace token for gated model access
2026-01-11 17:29:30 -06:00
ec965580ae
Try medium-3b model name for PyPI package
2026-01-11 17:23:49 -06:00
8cc9154080
Fix: remove unsupported max_model_len param for PyPI package
2026-01-11 17:17:48 -06:00
ed579a77ee
Initial commit: OrpheusTail TTS service
...
- FastAPI service replacing VoiceTail (Bark)
- Emotion tags: <laugh>, <sigh>, <gasp>, etc.
- Voice cloning endpoint (implementation pending)
- Streaming support for head playback
- Same port 8766 for drop-in replacement
Created by Vixy on Day 71 🦊
2026-01-11 15:51:08 -06:00