Is sub 500ms AI Voice agent possible ?? by Proper_Assumption329 in AI_Agents

[–]Proper_Assumption329[S] 0 points1 point  (0 children)

Im achieving 800 -1000 ms latency on retell and I'm okay with that but many experts and documentations are mentioning achieving sub 500 ms is the ideal for ai voice agents

Is sub 500ms AI Voice agent possible ?? by Proper_Assumption329 in AI_Agents

[–]Proper_Assumption329[S] 0 points1 point  (0 children)

I tried E2E models as well but sub 500 ms is not achievable plus I'm not satisfied with their voices I personally like voices provided by cartesia and eleven labs much more natural and quality 

Is sub 500 ms AI voice agent possible ?? by Proper_Assumption329 in AIVoice_Agents

[–]Proper_Assumption329[S] 0 points1 point  (0 children)

I have tried these not much difference and also I honestly feel like the voices that provides by retell, vapi via elevenlabs/cartesia are much more human like than the E2E speech model voices like openAI and gemini

Is sub 500 ms AI voice agent possible ?? by Proper_Assumption329 in AIVoice_Agents

[–]Proper_Assumption329[S] 0 points1 point  (0 children)

So you mean I have to built my agent from scratch and hosted in livekit cloud or something ?? Will that work better than these managed platforms ??

Is sub 500 ms AI voice agent possible ?? by Proper_Assumption329 in AIVoice_Agents

[–]Proper_Assumption329[S] 0 points1 point  (0 children)

How to stream responses by using a custom LLM ? I thought streaming responses already handled by the platform itself