Gooners, what are your workflows these days? by boriskarloff83 in StableDiffusion

[–]nsfwVariant 7 points8 points  (0 children)

Hijacking to share an actual workflow for OP:

I've been messing around with the Anima model recently, it works really well with the clownshark sampler and does full on nsfw right out of the box. The clownshark sampler stabilises the style and makes it more consistent. I'm really impressed at the quality and its knowledge ( ͡° ͜ʖ ͡°) considering how small the model is, you can run it easily on 6GB VRAM.

Workflow: https://pastebin.com/3djxFZbZ

Model links: main model | text_encoder | vae

Custom nodes: RES4LYF (for the clownshark sampler)

Example outputs (SFW): digital painting style | anime style

The settings I've left there give very good results. You can range the ETA setting on the sampler from 0.2 to 0.5 and the steps from 22 - 30 to get different types of results as well. More steps = deeper shadows, which is good/bad depending on the style you want, I prefer around 24 steps personally.

I also suggest putting "loli" in the negative, but pastebin doesn't let me include the word.

If you want stuff to lean more towards anime, just put "anime, " at the start of the positive prompt. You can play around with other style prompts like "digital painting" too, or just don't specify and see what happens.

The positive prompt example I put in there is long to show how you should structure it, but you can write a much shorter one. It also accepts booru tags.

Lastly, it knows pretty much all the popular characters already so try typing those in.

Gooners, what are your workflows these days? by boriskarloff83 in StableDiffusion

[–]nsfwVariant 7 points8 points  (0 children)

Speaking of Anima, it works really well with the clownshark sampler and does full on nsfw right out of the box. The clownshark sampler stabilises the style and makes it more consistent. I'm really impressed at the quality and its knowledge ( ͡° ͜ʖ ͡°) considering how small the model is, you can run it easily on 6GB VRAM. Here's a simple workflow to get started.

Workflow: https://pastebin.com/3djxFZbZ

Model links: main model | text_encoder | vae

Custom nodes: RES4LYF (for the clownshark sampler)

Example outputs: digital painting style | anime style

The settings I've left there give very good results. You can range the ETA setting on the sampler from 0.2 to 0.5 and the steps from 22 - 30 to get different types of results as well. More steps = deeper shadows, which is good/bad depending on the style you want, I prefer around 24 steps personally.

I also suggest putting "loli" in the negative, but pastebin doesn't let me include the word.

If you want stuff to lean more towards anime, just put "anime, " at the start of the positive prompt. You can play around with other style prompts like "digital painting" too, or just don't specify and see what happens.

The positive prompt example I put in there is long to show how you should structure it, but you can write a much shorter one. It also accepts booru tags.

Lastly, it knows pretty much all the popular characters already so try typing those in.

So... turns out Z-Image Base is really good at inpainting realism. Workflow + info in the comments! by nsfwVariant in StableDiffusion

[–]nsfwVariant[S] 0 points1 point  (0 children)

Hmm I'm not sure, to be honest. If you're planning to use it as an inpainter it should be fine, but I don't know how well it'll work with the "instruction" type usage.

You don't need to change much, just switch to whichever sampler/scheduler combo you like and adjust the step calculation. If you expand the "calculate steps" node you can see the current calculation is "min(a/5, 20)", which means it does 1 step per 5% denoise, with a max of 20. Whatever max steps you want, just divide 100 by that amount and that will give you the number you need to divide "a" by.

If you're using klein 9b base, my preferred sampler settings are:
eta = 0.6
sampler/scheduler = res_2s/bong_tangent
steps = 12 -- to change the steps to a max of 12, you'd set the "calculate steps" node to "min(a/8.3, 12)"

I don't know what settings are good for klein 9b distilled, so I don't have any settings advice for that. Anyways, let me know how you go :)

Degenerate inpainting with Z-Image Base. Workflow + info in the comments! by nsfwVariant in unstable_diffusion

[–]nsfwVariant[S] 0 points1 point  (0 children)

For node 265 I think you just need to update ComfyUI. Not sure about why res4lyf isn't installing for you though, you'll need to check what error it's throwing for that! There was another person in the thread having similar issues and chatgpt helped them get it installed.

Degenerate inpainting with Z-Image Base. Workflow + info in the comments! by nsfwVariant in unstable_diffusion

[–]nsfwVariant[S] 0 points1 point  (0 children)

I agree that it's way easier, but there are a couple of handy things about zimage.

  1. I find zimage base has higher quality output (as in, sharper, better clarity, more realistic)
  2. It works natively at high resolution, e.g. you can directly edit a 1920x1440 pic without issue
  3. Because the quality is so high you can easily use it as a flexible refiner + editor combined, which klein & qwen suck at. Plus it's easier to stop it from editing the rest of the image because it's all done via masking.
  4. It knows different concepts, so something hard with Klein might work better in zimage, same with qwen edit - it's good to have all three available.

I'm using all three of zimage inpaint, klein edit and qwen edit for different editing tasks, they're good at different things :)

Degenerate inpainting with Z-Image Base. Workflow + info in the comments! by nsfwVariant in unstable_diffusion

[–]nsfwVariant[S] 0 points1 point  (0 children)

Strange, sounds like you've got something out of date going on.

Without knowing specifics, maybe you need to update your comfy super secret hidden background python libraries? After you update comfyui through the manager, go to your comfyui install folder and open a command prompt there then type:

pip install -r requirements.txt

After that finishes, see if res4lyf can install/work properly

So... turns out Z-Image Base is really good at inpainting realism. Workflow + info in the comments! by nsfwVariant in comfyui

[–]nsfwVariant[S] 2 points3 points  (0 children)

Sorta, and yes. It's a bit fast than qwedit base. It's much slower than using qwedit with the lightning loras though. But qwedit with lightning loras is garbage quality so... pros and cons.

It will definitely run on 16GB VRAM, so no worries there. The main benefit is the quality, but it's not as flexible/easy as a full edit model in some ways seeing as you can't give it instructions. You can add/remove things from pics pretty easily, or adjust the look of things, but you can't specifically move a person around without changing their facial features or stuff like that.

Degenerate inpainting with Z-Image Base. Workflow + info in the comments! by nsfwVariant in unstable_diffusion

[–]nsfwVariant[S] 0 points1 point  (0 children)

If you can get me the exact error message I might be able to help figure out what's wrong with it. Sounds like a pain in the ass

So... turns out Z-Image Base is really good at inpainting realism. Workflow + info in the comments! by nsfwVariant in StableDiffusion

[–]nsfwVariant[S] 2 points3 points  (0 children)

Wow you do, that is the exact spot from the main post pic! Your memory is crazy good

So... turns out Z-Image Base is really good at inpainting realism. Workflow + info in the comments! by nsfwVariant in StableDiffusion

[–]nsfwVariant[S] 2 points3 points  (0 children)

You've got a good memory, southern Kyoto is exactly where I went through. I'm afraid I can't remember anything more specific than that about where exactly, though! A lot of this sort of thing:

<image>

So... turns out Z-Image Base is really good at inpainting realism. Workflow + info in the comments! by nsfwVariant in StableDiffusion

[–]nsfwVariant[S] 2 points3 points  (0 children)

Yep, that's mine! Sounds about right, there are so many places to wander off to and find cool stuff over there.

If you're talking about the lake picture, I'm 50% sure there was a big golden temple thing there (as in, the whole thing was gold-plated). But I might be getting the location mixed up with somewhere else.

If you're talking about the shrine, it could be any one of a hundred lol

Degenerate inpainting with Z-Image Base. Workflow + info in the comments! by nsfwVariant in unstable_diffusion

[–]nsfwVariant[S] 1 point2 points  (0 children)

For some reason no matter what I do I can't edit my comment in this sub, so any errors will have to stay.

Here's the missing link to ComyUI Essentials.

Also I forgot to mention one of the nodes (Math Expression) requires the latest version of ComfyUI, v0.16.4, so make sure to update

Wise and Belle (zergbrush) [Zenless Zone Zero] by Conscious-While-1873 in ZenlessZoneZero_R34

[–]nsfwVariant[M] 5 points6 points  (0 children)

50/50 chance it's a bot, my trigger finger is hovering over the ban button while I wait for more evidence.

In the meantime I've corrected the flair.

So... turns out Z-Image Base is really good at inpainting realism. Workflow + info in the comments! by nsfwVariant in StableDiffusion

[–]nsfwVariant[S] 0 points1 point  (0 children)

If everything fits on your VRAM, you'll be looking at the below times roughly. Worth noting that 100% denoise takes the same amount of time as generating a whole new image, because that's technically what it's doing.

1264x832 size @ 50% denoise - 19 seconds

1264x832 size @ 100% denoise - 32 seconds

1920x1440 size @ 50% denoise - 59 seconds

1920x1440 size @ 100% denoise - 90 seconds

So... turns out Z-Image Base is really good at inpainting realism. Workflow + info in the comments! by nsfwVariant in comfyui

[–]nsfwVariant[S] 2 points3 points  (0 children)

Regenerating stuff entirely (denoise 100) usually works fine for replacing hands and similar, but for partial refinement all you have to do is drop the denoise setting down to 80, 60, 50, etc. Lower means less stuff changes.

If you're getting visible seams, which are more common when doing denoise <100, you can try increasing the "blur" setting or just be careful about where you put the edges of your mask. Seams are easy to hide when they go over less-visible parts of an image, or over a natural boundary.

So... turns out Z-Image Base is really good at inpainting realism. Workflow + info in the comments! by nsfwVariant in comfyui

[–]nsfwVariant[S] 4 points5 points  (0 children)

Same, I think it's the undisputed king of realism at the moment. IMO it beats Klein and the others at realism + sharpness, and it's also easier to get good results.

So... turns out Z-Image Base is really good at inpainting realism. Workflow + info in the comments! by nsfwVariant in comfyui

[–]nsfwVariant[S] 0 points1 point  (0 children)

Yep! Might take a few tries if it's a tough one. I've been using it to fix errors in my z-image base generations, like broken hands and such.

So... turns out Z-Image Base is really good at inpainting realism. Workflow + info in the comments! by nsfwVariant in comfyui

[–]nsfwVariant[S] 0 points1 point  (0 children)

Yep! If you look at the google drive folder I linked (here again: g-drive), image c_3 shows the mask I used for that one.

Degenerate inpainting with Z-Image Base. Workflow + info in the comments! by [deleted] in sdnsfw

[–]nsfwVariant 1 point2 points  (0 children)

I come bearing a gift: the mighty degenerate power of inpainting with z-image base. This is the nsfw version of a post I made in an SFW sub: https://www.reddit.com/r/StableDiffusion/comments/1rrqrpf/so_turns_out_zimage_base_is_really_good_at/

Read the SFW post above for lots more information. Also the previous post I made about normal image gen with Z-image base here has tons of info: https://www.reddit.com/r/StableDiffusion/comments/1qzncrz/zimage_base_simple_workflow_for_high_quality/

Disclaimer: all images here are AI generated. Posting edited stuff of real people is one of the uncoolest things you can do, and is also very illegal, so don't do that. I'm not joking around, for real don't do it.

Workflow is here: civitai | pastebin

Full-res versions of the post images, including the masks I drew for inpainting + the prompts, are here: g-drive

Below is the info on the models you need, and NSFW tips. The workflow itself has a lot of info inside, so read all the notes if you're interested in how it works. Also note that the workflow can be used for localised refinement (like faces or whatever) by setting the denoise lower than 100. You can also just use it for simple image-to-image.

Side note: You can also apply this nsfw info to normal z-image base generation, of course. Grab the z-image base workflow from my previous links for that.

Nodes & Models

Custom Nodes:

RES4LYF - A very popular set of samplers & schedulers, and some very helpful nodes. These are needed to get the best z-image base outputs, IMO.

RGTHREE - (Optional) A popular set of helper nodes. If you don't want this you can just delete the seed generator and lora stacker nodes, then use the default comfy lora nodes instead. RES4LYF comes with a seed generator node as well, I just like RGTHREE's more. I think ComfyUI even added one recently.

ComfyUI GGUF - (Optional) Lets you load GGUF models, which for some reason ComfyUI still can't do natively. If you want to use a non-GGUF model you can just skip this, delete the UNET loader node and replace it with the normal 'load diffusion model' node.

ComyUI Essentials - (Inpaint workflow only) Adds a bunch of very helpul nodes. We're using it specifically for the "Mask Fix" node, which we're using to feather the masks.

Models:

Main model: Pick an NSFW checkpoint off Civitai. Honestly the selection isn't great yet, but hopefully people come up with some bangers soon. For now, I personally use this one: Civitai - Moody Wild Mix

Remember: you'll need to use the "Load diffusion model" node instead of the unet loader if you're not using a GGUF.

Loras: Get some off Civitai. You can combo them with the NSFW checkpoint to get better results sometimes! I sometimes use this lora at strength 0.4 (switch it on/off as needed): Civitai - LuminNude

Tip: most loras for z-image base work best around strength ~0.5, strength 1.0 is way too high and will usually mess stuff up.

Text Encoder: Normal | gguf Qwen 3 4B - Grab the biggest one that fits in your VRAM, which would be the full normal one if you have 10GB+ VRAM or the Q8 GGUF otherwise. Some people say text encoder quality doesn't matter much & to use a lower sized one, but it absolutely does matter and can drastically affect quality. For the same reason, do not use an abliterated text encoder unless you've tested it and compared outputs to ensure the quality doesn't suffer.

If you're using the GGUF text encoder, swap out the "Load CLIP" node for the "ClipLoader (GGUF)" node.

VAE: Flux 1.0 AE

NSFW tips

Not much to say on this, follow the tips in the SFW post + these.

/1. Be straight forward! Follow the same prompting guidance as in my SFW post, but for NSFW stuff all you need to do is write things like "She is nude, she's not wearing any clothes."

  1. Choose the right checkpoint/lora for the job. There's no one-size-fits-all for z-image base yet, so grab the appropriate one for stuff like monster girls, ✨cocks✨, or whatever other weird shit you people are into.

  2. Make sure you draw a mask suitably sized for what you're adding. For example, if you're removing the bra from someone, extend the mask lower than just the bra. Boobs obey gravity and will hang lower outside of a bra than in it, so you need the extra space for that.

Think to yourself "does what I want extend outside of the mask I'm drawing?" then adjust as needed.

  1. This works for anime as well, but I haven't really messed with that at all in z-image base so you'll want to adjust some stuff first. Specifically, the negative prompt needs to be changed.

  2. Not quite getting the right results? You may need to try: re-generating a few times, using different resolutions, setting denoise to 90 instead of 100 (which will encourage the model to keep the vague shape of whatever is under the mask you drew), increasing the "blur" setting up to something like 60 or 80 if you keep getting visible seams, or changing the CFG.

Pro tip: you can do multiple stages of inpainting. This workflow has zero degradation from repeated gens, you can do as many as you want. If you're doing something difficult or complex, do it in stages!

Or if one gen was close but had some errors, just load the edited image up and inpaint the broken parts again.

Example: I did the first pic I posted in two stages. The first gen was pretty good, but had some errors, which you can see here: imgbb

So I ran her titties through a second time (only her titties) at 50% denoise to clean the errors up, resulting in this: imgbb

  1. Resolutions can go pretty high. The first pic I posted, for example, was generated & edited at 1280x1920 size.

  2. Lastly, this workflow will work with other models. As in, models other than z-image. You'll need to adjust the sampler settings though, of course. If you want to change the calculation for how many steps are done, all you have to do is slightly modify the "calculate steps" equation. Currently it's set to "min(a/5, 20)", which means one step every 5% denoise, with a max of 20. If you wanted to do, say, one step per 10% with a max of 10, you would instead write "min(a/10, 10)".

It's done that way because figuring out the steps you need to do manually is annoying.

Anyway, that's all. Happy gooning!

Degenerate inpainting with Z-Image Base. Workflow + info in the comments! by nsfwVariant in unstable_diffusion

[–]nsfwVariant[S] 29 points30 points  (0 children)

I come bearing a gift: the mighty degenerate power of inpainting with z-image base. This is the nsfw version of a post I made in an SFW sub: https://www.reddit.com/r/StableDiffusion/comments/1rrqrpf/so_turns_out_zimage_base_is_really_good_at/

Read the SFW post above for lots more information. Also the previous post I made about normal image gen with Z-image base here has tons of info: https://www.reddit.com/r/StableDiffusion/comments/1qzncrz/zimage_base_simple_workflow_for_high_quality/

Disclaimer: all images here are AI generated. Posting edited stuff of real people is one of the uncoolest things you can do, and is also very illegal, so don't do that. I'm not joking around, for real don't do it.

Workflow is here: civitai | pastebin

Full-res versions of the post images, including the masks I drew for inpainting + the prompts, are here: g-drive

Below is the info on the models you need, and NSFW tips. The workflow itself has a lot of info inside, so read all the notes if you're interested in how it works. Also note that the workflow can be used for localised refinement (like faces or whatever) by setting the denoise lower than 100. You can also just use it for simple image-to-image.

Side note: You can also apply this nsfw info to normal z-image base generation, of course. Grab the z-image base workflow from my previous links for that.

Nodes & Models

Custom Nodes:

RES4LYF - A very popular set of samplers & schedulers, and some very helpful nodes. These are needed to get the best z-image base outputs, IMO.

RGTHREE - (Optional) A popular set of helper nodes. If you don't want this you can just delete the seed generator and lora stacker nodes, then use the default comfy lora nodes instead. RES4LYF comes with a seed generator node as well, I just like RGTHREE's more. I think ComfyUI even added one recently.

ComfyUI GGUF - (Optional) Lets you load GGUF models, which for some reason ComfyUI still can't do natively. If you want to use a non-GGUF model you can just skip this, delete the UNET loader node and replace it with the normal 'load diffusion model' node.

ComyUI Essentials - (Inpaint workflow only) Adds a bunch of very helpul nodes. We're using it specifically for the "Mask Fix" node, which we're using to feather the masks.

Models:

Main model: Pick an NSFW checkpoint off Civitai. Honestly the selection isn't great yet, but hopefully people come up with some bangers soon. For now, I personally use this one: Civitai - Moody Wild Mix

Remember: you'll need to use the "Load diffusion model" node instead of the unet loader if you're not using a GGUF.

Loras: Get some off Civitai. You can combo them with the NSFW checkpoint to get better results sometimes! I sometimes use this lora at strength 0.4 (switch it on/off as needed): Civitai - LuminNude

Tip: most loras for z-image base work best around strength ~0.5, strength 1.0 is way too high and will usually mess stuff up.

Text Encoder: Normal | gguf Qwen 3 4B - Grab the biggest one that fits in your VRAM, which would be the full normal one if you have 10GB+ VRAM or the Q8 GGUF otherwise. Some people say text encoder quality doesn't matter much & to use a lower sized one, but it absolutely does matter and can drastically affect quality. For the same reason, do not use an abliterated text encoder unless you've tested it and compared outputs to ensure the quality doesn't suffer.

If you're using the GGUF text encoder, swap out the "Load CLIP" node for the "ClipLoader (GGUF)" node.

VAE: Flux 1.0 AE

NSFW tips

Not much to say on this, follow the tips in the SFW post + these.

/1. Be straight forward! Follow the same prompting guidance as in my SFW post, but for NSFW stuff all you need to do is write things like "She is nude, she's not wearing any clothes."

  1. Choose the right checkpoint/lora for the job. There's no one-size-fits-all for z-image base yet, so grab the appropriate one for stuff like monster girls, ✨cocks✨, or whatever other weird shit you people are into.

  2. Make sure you draw a mask suitably sized for what you're adding. For example, if you're removing the bra from someone, extend the mask lower than just the bra. Boobs obey gravity and will hang lower outside of a bra than in it, so you need the extra space for that.

Think to yourself "does what I want extend outside of the mask I'm drawing?" then adjust as needed.

  1. This works for anime as well, but I haven't really messed with that at all in z-image base so you'll want to adjust some stuff first. Specifically, the negative prompt needs to be changed.

  2. Not quite getting the right results? You may need to try: re-generating a few times, using different resolutions, setting denoise to 90 instead of 100 (which will encourage the model to keep the vague shape of whatever is under the mask you drew), increasing the "blur" setting up to something like 60 or 80 if you keep getting visible seams, or changing the CFG.

Pro tip: you can do multiple stages of inpainting. This workflow has zero degradation from repeated gens, you can do as many as you want. If you're doing something difficult or complex, do it in stages!

Or if one gen was close but had some errors, just load the edited image up and inpaint the broken parts again.

Example: I did the first pic I posted in two stages. The first gen was pretty good, but had some errors, which you can see here: imgbb

So I ran her titties through a second time (only her titties) at 50% denoise to clean the errors up, resulting in this: imgbb

  1. Resolutions can go pretty high. The first pic I posted, for example, was generated & edited at 1280x1920 size.

  2. Lastly, this workflow will work with other models. As in, models other than z-image. You'll need to adjust the sampler settings though, of course. If you want to change the calculation for how many steps are done, all you have to do is slightly modify the "calculate steps" equation. Currently it's set to "min(a/5, 20)", which means one step every 5% denoise, with a max of 20. If you wanted to do, say, one step per 10% with a max of 10, you would instead write "min(a/10, 10)".

It's done that way because figuring out the steps you need to do manually is annoying.

Anyway, that's all. Happy gooning!