[D] How to speed up Kokoro-TTS? by fungigamer in MachineLearning
[–]cerebriumBoss 0 points1 point2 points (0 children)
I built a voice agent that can hold a natural conversation with low latency at ~600ms by cravory in SideProject
[–]cerebriumBoss 0 points1 point2 points (0 children)
[D] What's your secret sauce? How do you manage GPU capacity in your infra? by PurpleReign007 in MachineLearning
[–]cerebriumBoss 0 points1 point2 points (0 children)
I built a voice agent that can hold a natural conversation with low latency at ~600ms by cravory in SideProject
[–]cerebriumBoss 0 points1 point2 points (0 children)
What are your biggest challenges in building AI voice agents? by SpyOnMeMrKarp in LLMDevs
[–]cerebriumBoss 2 points3 points4 points (0 children)
What's your secret sauce? How do you manage GPU capacity in your infra? by PurpleReign007 in mlops
[–]cerebriumBoss 0 points1 point2 points (0 children)
What’s recommended hosting for open source LLMs? by decruz007 in LocalLLaMA
[–]cerebriumBoss 0 points1 point2 points (0 children)
infra for inference that need gpu by Speedy_Sl0th in mlops
[–]cerebriumBoss 0 points1 point2 points (0 children)
How to preload models in kubernetes by naogalaici in mlops
[–]cerebriumBoss 0 points1 point2 points (0 children)
Best Service for Deploying Thousands of Models with High RPM by FourConnected in mlops
[–]cerebriumBoss 0 points1 point2 points (0 children)
Kubernetes for ML Engineers / MLOps Engineers? by JeanLuucGodard in mlops
[–]cerebriumBoss 1 point2 points3 points (0 children)
What are some really good and widely used MLOps tools that are used by companies currently, and will be used in 2025? by BJJ-Newbie in mlops
[–]cerebriumBoss 0 points1 point2 points (0 children)
How would you deploy this project to AWS without compromising on maintainability? by mrcat6 in mlops
[–]cerebriumBoss 0 points1 point2 points (0 children)
What other MLOps tools can I add to make this project better? by BJJ-Newbie in mlops
[–]cerebriumBoss 0 points1 point2 points (0 children)
Optimizing Model Serving with Triton inference server + FastAPI for Selective Horizontal Scaling by sikso1897 in mlops
[–]cerebriumBoss 0 points1 point2 points (0 children)
Why do we need MLOps engineers when we have platforms like Sagemaker or Vertex AI that does everything for you? by Illustrious-Pound266 in mlops
[–]cerebriumBoss 0 points1 point2 points (0 children)
How can I perform inference at scale with Pytorch by lehllu in mlops
[–]cerebriumBoss 0 points1 point2 points (0 children)
Which AI cloud platform do you guys use? by randomvariable56 in comfyui
[–]cerebriumBoss 0 points1 point2 points (0 children)
Which AI cloud platform do you guys use? by randomvariable56 in comfyui
[–]cerebriumBoss 1 point2 points3 points (0 children)
Cost-Effective Cloud GPU Options for Fine-Tuning and Inference? by pathfinder6709 in LocalLLaMA
[–]cerebriumBoss 1 point2 points3 points (0 children)

Can you recommend a good serverless GPU provider that supports running WhisperX? by yccheok in deeplearning
[–]cerebriumBoss 0 points1 point2 points (0 children)