Thanks for your comment, Ashish!
As I mentioned in the "Notes and Recommendations" section, I used the ElevenLabs eleven_turbo_v2 model since it’s more suitable for tasks demanding low latency, intentionally set the max_tokens OpenAI parameter to 100, and used a custom prompt to reduce the number of characters returned to 100. This worked perfectly for my use case.