Fix streaming endpoint max_tokens limit - Day 72

Both generate_speech_sync() and stream_tts() were calling
model.generate_speech() without max_tokens parameter.
Now explicitly passing max_tokens=4000 to both.

Fixed by Vixy 🦊💜
This commit is contained in:
2026-01-12 16:56:43 -06:00
parent 0fa4042025
commit 75a5fc0a95

View File

@@ -196,6 +196,7 @@ def generate_speech_sync(text: str, voice: str) -> bytes:
syn_tokens = model.generate_speech( syn_tokens = model.generate_speech(
prompt=text, prompt=text,
voice=voice, voice=voice,
max_tokens=4000, # Increased from default 1200 for longer texts
) )
print(f"Got generator: {type(syn_tokens)}") print(f"Got generator: {type(syn_tokens)}")
@@ -569,6 +570,7 @@ async def stream_tts(request: TTSStreamRequest):
syn_tokens = model.generate_speech( syn_tokens = model.generate_speech(
prompt=request.text, prompt=request.text,
voice=voice, voice=voice,
max_tokens=4000, # Increased from default 1200 for longer texts
) )
for audio_chunk in syn_tokens: for audio_chunk in syn_tokens: