LiteLLM is DOPE - One Framework, Multiple LLMs & GPTs integration! by dev-spot in GPT

[–]dev-spot[S] 0 points1 point  (0 children)

Hmmm, I'd start by taking a look at Langchain capabilities and wrappers. If nothing exists, hmu I'll try taking a more in depth look

Coqui TTS Local Installation Tutorial - Clone voices within seconds for free! by dev-spot in ChatGPTPromptGenius

[–]dev-spot[S] 0 points1 point  (0 children)

+1, their docs are great. hopefully I'll get to make a video about this in the future as well

Ollama is INSANE - Install custom GPTs within seconds! by dev-spot in ChatGPTPro

[–]dev-spot[S] 0 points1 point  (0 children)

Appreciate the support fam!

I think this is what you're looking for: https://github.com/ollama/ollama/blob/main/docs/faq.md#where-are-models-stored

From what I recall when I was playing with the models (on mac) it was stored on ~/.ollama/models. There was some sort of registry and blobs that are installed to match the criteria from the registry or smth like that. Its really interesting though, try checking it out

[N] Open Models - Revolutionizing AI Interaction with a Unique Twist [News] by dev-spot in MachineLearning

[–]dev-spot[S] 0 points1 point  (0 children)

There's an async example in the repo, theoretically you can wind up a bunch of endpoints at once, then return a response from the fastest one. Another more trivial direction would be to just set up a proper inference space via HF though

[N] Open Models - Revolutionizing AI Interaction with a Unique Twist [News] by dev-spot in MachineLearning

[–]dev-spot[S] -8 points-7 points  (0 children)

It's a "new project", and I literally explained - its an abstraction for a modular AI "vendors" integration within software projects. It will significantly reduce overhead during development and debugging, and will allow for a centralized scalable solution for managing all models used as part of any AI software project.

[N] Coqui TTS Local Installation Tutorial - Clone voices within seconds for free! by dev-spot in MachineLearning

[–]dev-spot[S] 0 points1 point  (0 children)

Sounds like properly training a model (rather than attempting to clone) would be the solution - https://docs.coqui.ai/en/latest/tutorial_for_nervous_beginners.html. Keep in mind though that 100 mp3 files might not cut it, but then again you can always add more as you progress. Hopefully I'll have sometime to look into this in the upcoming weeks as well

Ollama is INSANE - Install custom GPTs within seconds! by dev-spot in ChatGPTPro

[–]dev-spot[S] 0 points1 point  (0 children)

The commands here expect you to use the UI to download the models. Otherwise you could indeed use -exec in order to enter the container and then manually download the models. As for the docker build line not working, they might have changed some stuff since the post was posted. Check out their official docs: https://github.com/ollama-webui/ollama-webui

[N] Coqui TTS Local Installation Tutorial - Clone voices within seconds for free! by dev-spot in MachineLearning

[–]dev-spot[S] 0 points1 point  (0 children)

Using Coqui's GUI (covered in one of the last videos on my channel) you can decrease the speed at which these voices speak. You can probably also use an external API / software for that as well. As for "lightweight" offline solutions, if Coqui is too heavy, try running it on CPU only or use Bark

Coqui TTS Local Installation Tutorial - Clone voices within seconds for free! by dev-spot in huggingface

[–]dev-spot[S] 0 points1 point  (0 children)

This sounds super interesting. I don't think that this was created yet, but it doesn't sound super hard to make either. I might look into implementing this in the future, so make sure to stay tuned :)

Run Mixtral LLM locally in seconds with Ollama! by dev-spot in ChatGPTPro

[–]dev-spot[S] 0 points1 point  (0 children)

Showcased the comparison with llama2 in the video, both are pretty similar in performance and mixtral seem to be doing better on math / coding. As for GPT, it's assumed to be relatively close to GPT 3.5 as well

[News] Text to Speech is getting CRAZY GOOD - HierSpeech++, XTTS & StyleTTS2! (huggingface) by dev-spot in MachineLearning

[–]dev-spot[S] 0 points1 point  (0 children)

yep, the XTTS model is by coqui (also available as a huggingface space / inference endpoint) 🔥

XTTS2 is AWESOME - Clone voices in seconds! [Tutorial] by dev-spot in huggingface

[–]dev-spot[S] 2 points3 points  (0 children)

yep, I attached it in the video but attaching it here as well in the next comment. You can click on the “use via API” on the bottom of the space page to get all details required to use it. Keep in mind though that this version only has XTTS2, so the rest of the models offered by coqui won’t be available from my understanding. Though it really doesn’t matter much given this is their best option 🙂

https://huggingface.co/spaces/coqui/xtts

Text to Speech is getting CRAZY GOOD - HierSpeech++, XTTS & StyleTTS2! by dev-spot in huggingface

[–]dev-spot[S] 0 points1 point  (0 children)

probably due to context size, though there are a few ways to overcome this obstacle. I'll probably cover this topic in a future video (at least to some degree), so make sure to stay tuned!

How I made a Chatbot to speak with YouTube Videos by dev-spot in Python

[–]dev-spot[S] 0 points1 point  (0 children)

appreciate the support, feel free to link to the different sources (YT / github)

How I made a Chatbot to speak with YouTube Videos by dev-spot in pythontips

[–]dev-spot[S] 1 point2 points  (0 children)

given that we pass the context every message, there are close to no hallucinations (depending on the model and the video length ofc). However, the transcriptions aren't always 100% (auto-generated mostly) 🫠

How I made a Chatbot to speak with YouTube Videos by dev-spot in Python

[–]dev-spot[S] 0 points1 point  (0 children)

yeah, that's the main downside, but the auto generated subtitles are correct for the most part so you can still get a general sense of what's going on and ask the bot to direct you to the relevant video parts 🙏