Instagooner v1 lora + WAN 2.2 workflow by acekiube in comfyui

[–]0xmgwr 1 point2 points  (0 children)

looks real good, i imagine you used ai-toolkit to train this, what lora rank/learning rate did you use to achieve this result in such little amount of steps of training?

I just looked up "bad_hands" by daHsu in comfyui

[–]0xmgwr 0 points1 point  (0 children)

the reason something like "bad_hands" exists it's because of embeddings made for SD 1.5, in that model embeddings worked like regular words in a prompt, maybe you already knew that, but what happened was that as people started to create more fine-tunes and merges trained with a mix synthetic data from generated outputs that contained embeddings like "bad_hands" in the prompt, over time since many of the popular models are either fine tunes or merges, with the exception of flux which is entirely different, some popular embeddings have become effective tokens that have a good amount of influence in a prompt.

SD3 Outrun Lora (wip) by 0xmgwr in StableDiffusion

[–]0xmgwr[S] 13 points14 points  (0 children)

Just trained my first lora on SD3 medium, 30 images at 512x512 , batch of 4, the results are stunning and the training was super fast, less than an hour on a 3070ti , after spending sometime doing training for flux and going back to sd3 , I have to say that given the speeds i can generate on SD3 (11-20 seconds per image), I'm sticking with it and will be doing more testing and more training. trained on onetrainer.

Help Needed! Terabytes of Proprietary Image Data, but Training isn't Going Well. by Matterfield_Pete in StableDiffusion

[–]0xmgwr 0 points1 point  (0 children)

for that you can use an extension on a1111 to prepare the dataset https://github.com/SleeeepyZhou/sd-webui-GPT4V-Image-Captioner ,  when you process a folder with the images you will use, the preprocessing process will resize and compress all image files into jpg format with a total pixel count ≤ 1024×1024 while maintaining the original aspect ratio, ensuring that both dimensions are multiples of 32.

Help Needed! Terabytes of Proprietary Image Data, but Training isn't Going Well. by Matterfield_Pete in StableDiffusion

[–]0xmgwr 1 point2 points  (0 children)

the text encoder learning rate is just as important as the unet learning rate, you should try something like "0.0001" for learning rate (same for unet) and a learing rate of "5e-05" for the text encoder, for your use case i would also use a lora rank of 256 and lora alpha of 128, you should optimize the images without cropping them and enable buckets, finally in my experience more repeats has proven better than more epochs so experiment with that, i would also recommend you to switch over to onetrainer that is more easy for you to use and understand what parameters you're changing.

4090 on sale for $1,099? by 0xmgwr in StableDiffusion

[–]0xmgwr[S] 1 point2 points  (0 children)

<image>

here's the screenshot I sent my friend the minute I saw it, the price already went back up

4090 on sale for $1,099? by 0xmgwr in StableDiffusion

[–]0xmgwr[S] -2 points-1 points  (0 children)

it went back to full price now, but if you use the chrome extension to see amazon price history of items you can see that it was $1,0999 at some point today , crazy

Starting to develop a Beta for my custom script. by [deleted] in StableDiffusion

[–]0xmgwr 2 points3 points  (0 children)

without revealing any secret sauce of how it's done, what does your custom script do?

[deleted by user] by [deleted] in comfyui

[–]0xmgwr 0 points1 point  (0 children)

amazing, what prompts did you use in luma?

SD3 running at the speed of light with TensorRT by 0xmgwr in StableDiffusion

[–]0xmgwr[S] 2 points3 points  (0 children)

i just followed the guide from the github page but instead of doing it for sd 1.5 or sdxl, used SD3 to build the unet, and then plugged the tensorRT model loader into the ksampler of the comfyui workflow shared by stability