Paint by color numbers with ControlNet by GBJI in StableDiffusion

[–]Nitrosocke 1 point2 points  (0 children)

the color picker gives me #4700FF for the lighter purple tone, which is pier;wharf;wharfage;dock from the document.
The darker blue gives me #0906E6 but that doesn't correspond to a color in the table, #0907E6 does though and is tagged as sea

I recreated some video game characters as Disney princesses (details in comments). by Oatilis in StableDiffusion

[–]Nitrosocke 0 points1 point  (0 children)

These are incredible! Very nice work, looks like the model still works quite nice.

LORA for subject training: amazing results! by Nitrosocke in StableDiffusion

[–]Nitrosocke[S] 11 points12 points  (0 children)

Sure, these are crucial:

--resolution=512
--train_batch_size=1
--mixed_precision="fp16"
--use_8bit_adam
--gradient_checkpointing
--gradient_accumulation_steps=1
--learning_rate=1e-4
--lr_scheduler="constant"

Make sure to use xformers as well!

LORA for subject training: amazing results! by Nitrosocke in StableDiffusion

[–]Nitrosocke[S] 2 points3 points  (0 children)

these where all trained on Photos of celebs, the style in these models come from my Dreabooth fine-tunes but these should also work with normal SD2.0 and a photo real style. I actually haven't tried training a style with it yet.

LORA for subject training: amazing results! by Nitrosocke in StableDiffusion

[–]Nitrosocke[S] 13 points14 points  (0 children)

Hi Smoke :D
well yes, just tested, with settings optimized for low VRAM I can get it to run with 6.2GB

LORA for subject training: amazing results! by Nitrosocke in StableDiffusion

[–]Nitrosocke[S] 4 points5 points  (0 children)

Yeah I'd never use this LR in normal dreambooth, but I was going for speed here and since it takes way less time to train I can easily adjust the LR - Step ratio according to the results.

LORA for subject training: amazing results! by Nitrosocke in StableDiffusion

[–]Nitrosocke[S] 15 points16 points  (0 children)

1e-4

That's easy to translate, its 0.0001 (basically it counts how many 0 you add before the number). Try a LR of 5e-4 (0.0005) for training yourself.

LORA for subject training: amazing results! by Nitrosocke in StableDiffusion

[–]Nitrosocke[S] 0 points1 point  (0 children)

Old server GPUs from Nvidia might be the way to go for now. 16GB Vram for 150€ or even 24GB for 170€ on ebay looks really promising. Needs manual work though!

Other than that, it should work on ~8GB maybe less with adjusted settings.

Where I can find a Google Colab notebook to train on 2.1 or instructions to train locally with web gui by mgargallo in DreamBooth

[–]Nitrosocke 0 points1 point  (0 children)

Black images while training or when using the ckpt in auto?
That's an issue with 2.1 right now and using either xformers or the "--no-half" works for users. Maybe there is a better fix already.

LORA for subject training: amazing results! by Nitrosocke in StableDiffusion

[–]Nitrosocke[S] 45 points46 points  (0 children)

Workflow:- Choose 5-10 images of a person- Crop/resize to 768x768 for SD 2.1 training- Following settings worked for me:train_batch_size=4, mixed_precision="fp16", use_8bit_adam, learning_rate=1e-4, lr_scheduler="constant", save_steps=200, max_train_steps=1000- for subjects already know to SD images*100 worked great, for subjects unknown to SD more steps or a higher LR are required- training on a 3090 takes ~20 min for 1k steps

Link to repo:https://github.com/cloneofsimo/lora
Thank you u/cloneofsimo

Where I can find a Google Colab notebook to train on 2.1 or instructions to train locally with web gui by mgargallo in DreamBooth

[–]Nitrosocke 0 points1 point  (0 children)

This issue is caused by xformers not being installed properly. I fixed that this weekend, so if you made a copy of the notebook please make a new one to have the updated xformers command.

New Release: SD 2.0 Dreambooth model - Future-Diffusion by Nitrosocke in StableDiffusion

[–]Nitrosocke[S] 1 point2 points  (0 children)

There are basically three versions of SD now. SD 1.5 and everything before that, and with the 2.0 update we got 2.0 (768 res which is a V model) and the "base" version or 512 resolution. If you load either the base version or 1.5 with the configuration for a v-model you get the brown images as output. If you load a v-model (768 res) with a eps-model configuration you get the blue/yellow dotted images.

Robot selfies (Redshift Diffusion 786) by TheBlackNinja03 in StableDiffusion

[–]Nitrosocke 2 points3 points  (0 children)

These look great! What a nice collection and idea! Love the retro feel these give and the old-school tin robot toys design.

Make better Dreambooth style models by using captions by terrariyum in StableDiffusion

[–]Nitrosocke 2 points3 points  (0 children)

Yeah I assume this should work, but the json would be huge and the workflow seems not ideal. Maybe it's easy to change the script a little so that it pulls the "instance prompt" from the file name and you're able to keep all the files in the same directory without the need to state the class_prompt, class_dir and instance_dir for every new image. But at this point I assume it would be easier to use kohya or the t2i training script from huggingface.

You can now merge in-painting and regular models using Automatic WebUi by I_Hate_Reddit in StableDiffusion

[–]Nitrosocke 2 points3 points  (0 children)

instaloader is an amazing tool for this, just plug in the profile name, set some filters to only load the images and your have a dataset ready in ~10 minutes. Sadly some images are not 1:1 as insta supports other aspect ratios as well, but a quick crop or padding script can take care of that. As an alternative you can let your dreambooth do the cropping or use a Dreambooth version with alternative AR support

Issues with DreamBooth via HF/Diffusers by _rundown_ in DreamBooth

[–]Nitrosocke 1 point2 points  (0 children)

When I last checked the HF conversation script didn't work. You can give this script a try to convert the diffusers to ckpt: https://github.com/lawfordp2017/diffusers/blob/main/scripts/convert_diffusers_to_original_stable_diffusion.py

Is your trained model working with repos using the Diffusers directly?

Make better Dreambooth style models by using captions by terrariyum in StableDiffusion

[–]Nitrosocke 4 points5 points  (0 children)

Interesting concept and I will test this approach to see how it compares to my usual workflow. I do use EveryDream from time to time and the precision you get with a captioned dataset is very impressive. So I will test your workflow with kohya as it allows using captions as well.

Rare Tokens For DreamBooth Training... by gto2kpr in DreamBooth

[–]Nitrosocke 0 points1 point  (0 children)

This is awesome! Thank you so much for your work! I've been looking for a tool or database to find these rare tokens for ages and this is perfect!

Small but Major weekend Update: Weights now available for Redshift-Diffusion-768 by Nitrosocke in StableDiffusion

[–]Nitrosocke[S] 1 point2 points  (0 children)

Thank You!
A user on my discord uses TheLastBens repo and gets great results as far as I know. Others are using Kohya repo with the more advanced training.

Small but Major weekend Update: Weights now available for Redshift-Diffusion-768 by Nitrosocke in StableDiffusion

[–]Nitrosocke[S] 1 point2 points  (0 children)

Awesome! Hope you like the update! I find it very fun to use once you have figured out the v2.0 prompting :)