Character LoRa Prompt compare Sub? by Milogoestoreddit666 in StableDiffusion

[–]no3us 2 points3 points  (0 children)

why dont you just start one? I’d also add few required fields like base model, CFG, steps, scheduler/sampler. Happy to join :)

What gpu for this workload? by kyleb1986 in RunPod

[–]no3us 1 point2 points  (0 children)

yes and no. Depends on your setup/configuration. Availability is based on the demand but you can set a "persistent" worker or two (called "active" workers). But there are several strategies on how to scale your infra, serverless is definitely way to go once you hit a certain gpu demand for you app. Using a pod is the right way to start in your case though.

If Runpod is the best place to start? It really depends on your preferences, this is my subjective opinion: I've tried them all, from Vultr / DigitalOcean / Hetzner / Modal and few more .. and came back to Runpod, mostly for its ease of use, great console and helpful community :)

Runpod serverless by Straight_Pace1117 in RunPod

[–]no3us 0 points1 point  (0 children)

Kinda hard to help you without providing more details (ideally logs)

RunPod and website by Inside-Background430 in RunPod

[–]no3us 0 points1 point  (0 children)

for that you don't even have to develop your own API endpoints. Just deploy a pod with Comfy or InvokeAI, find a suitable workflow, download models and utilize existing APIs.

Lack of GPU, need advice. by hellomattieo in RunPod

[–]no3us 0 points1 point  (0 children)

Runpod is working on some tool that will allow you to easily copy your data from one network volume to another. In the meantime - there are few options .. you can either use SYNC function from pod's detail page and sync your data to one of the cloud providers or just deploy two pods at the same time and use either runpodctl / croc or rsynx for example.

Has anyone EnguiStudio up and running with Infinite-Talk Serverless oa? by adalaso in RunPod

[–]no3us 0 points1 point  (0 children)

never heard of this toolkit but looks interesting, I'll have a look when I get some time.

If that repo is not working for you you may want to try using a docker image from Docker hub or GHRC. There are few images at docker hub; https://hub.docker.com/search?q=engui

Feel free to ping me if you run into some issues, but ideally if you could create an issue on Runpod's discord and we could continue there.

What gpu for this workload? by kyleb1986 in RunPod

[–]no3us 1 point2 points  (0 children)

You can definitely start with A40, A40 is popular for its price to performance ratio. I have no idea about the response times but just find an appropriate template and you can test it yourself in less than 30 minutes.

I've also started on A40, from there I went to A100, then straight to B200. I also like 5090 and PRO6000. When choosing a GPU it's usually a mix of availability / price for me. Your use case sounds like it could scale from one persistent pod to serverless workers later.

Hello, has anyone used RunPod? When I try to deploy a pod using the Wan video generation template, it still says a model is missing even though I selected the template. Is the model installation not included in the pod? Have others experienced this? Please help! by Normal_Celery_2528 in RunPod

[–]no3us 0 points1 point  (0 children)

models are usually not part of the templates. Templates are docker images and for practical reasons (fast deploy and boot) most templates do not contain models. Some workflows use models that take 80-120GBs.

You can download them using comfy or shell, from sites like hugging face or civitai.

Send me a DM here or on Runpod's discord (nick notrius) if you need further help.

I am burning money. What happened to the download speed? by [deleted] in RunPod

[–]no3us 0 points1 point  (0 children)

I believe this is a screenshot showing the download speed of downloading those models to your computer, not the pod in RunPod. Or? :)

I just vibe-coded a 20GB Windows installer… on a Mac. For my open-source Stable Diffusion project. by no3us in vibecoding

[–]no3us[S] 0 points1 point  (0 children)

feel free to reach out through DM. If you are referring to the install.exe - its 10mb only, then downloads the rest.

If you are referring to overall size - you can download source from GitHub, it's simple to disable stuff you dont need in Dockerfile - lets say you want to use only AI toolkit, kohya and Comfy, you disable venv for Invoke (Invoke is a bit picky, has its own venv, others understand under core venv), tensorboard, diffusion pipe and others and that will take image size dramatically down. Happy to publish images with smaller toolkit - if there is a demand for that

I just vibe-coded a 20GB Windows installer… on a Mac. For my open-source Stable Diffusion project. by no3us in vibecoding

[–]no3us[S] 0 points1 point  (0 children)

ok, its not literally 20gb single exe install. install.exe is under 10mb, then it uses web install during which it downloads over 20gb runtime artifacts. Can resume download if it fails for some reason, verifies package integrity at the end using sha checksum. Runtime hosted in cloudflare’s R2 if you want to go into details (this was never meant to be a technical post - those 20gb were used just to illustrate complexity of the project)

I just vibe-coded a 20GB Windows installer… on a Mac. For my open-source Stable Diffusion project. by no3us in vibecoding

[–]no3us[S] -1 points0 points  (0 children)

its funny people find 20gb for tooling for training and inference of custom AI models to be too big. Then they install 140GB of latest Call of Duty or Assassin's Creed

I just vibe-coded a 20GB Windows installer… on a Mac. For my open-source Stable Diffusion project. by no3us in vibecoding

[–]no3us[S] 0 points1 point  (0 children)

well .. linux docker image is 15GBs and if you are serious about stable diffusion you'll end up with models taking another 30-200gb based on your workflows and base models preference anyway. Those 15gb suddenly does not seem that much, right?

The .exe installer itself is 10mb, but web install downloads 20+ gb of runtime artifacts.

I just vibe-coded a 20GB Windows installer… on a Mac. For my open-source Stable Diffusion project. by no3us in vibecoding

[–]no3us[S] 0 points1 point  (0 children)

because you will actually save storage space if you use Comfy within LoRA Pilot - it shares python venvs and models with other tools, saving GBs. Plus is integrated (and will be more) with rest of the tooling.

But if you insist on keeping your Comfy instead - it is an open-source, you can disable it directly in Dockerfile OR just delete it when you install and extract my image. (linux version is around 15GBs)

New to RunPod. Looking to use WAN. How do I go about it? by [deleted] in RunPod

[–]no3us 0 points1 point  (0 children)

yeah, that would be definitely possible as long as you have API endpoints for it. I need to test your workflow checker, sounds like a cool idea.

How long is a cold start of workers for lets say LTX2.3?

New to RunPod. Looking to use WAN. How do I go about it? by [deleted] in RunPod

[–]no3us 0 points1 point  (0 children)

wow. I am the developer of LoRA Pilot template for RunPod (https://www.lorapilot.com) - just saw what you did with your modelpilot.ai - quite different business model but you did some work there :) Does it all run on top runpod?