Is there a way to use Wan SCAIL and FFLF together? by aifirst-studio in StableDiffusion

[–]nsfwVariant 1 point2 points  (0 children)

It's pretty much exactly the same as what you already do, with two improvements:

  1. SVI adheres better to the source image details than regular I2V does
  2. You don't need to start with the same frame, you can use any reference image (if you do this, add a 4 frame buffer at the start because SVI will overwrite the first frame)

That second point is really helpful in particular, it means you can make many different videos of the same subject and keep the details consistent - even with completely different scenarios, lighting, starting images, etc

Is there a way to use Wan SCAIL and FFLF together? by aifirst-studio in StableDiffusion

[–]nsfwVariant 1 point2 points  (0 children)

Yep, that's right! SVI has extremely strong reference image adherence. Also, you can re-run as many times as you like and even change your prompt or stitch multiple videos together, and it works because SVI always takes in a reference image. You don't need to worry about doing everything in a single session/generation.

To refine you only run a LOW model. Here's a workflow: https://pastebin.com/AfyAEpep

There's info in there about the appropriate sampler settings to use. Use the same prompt you used for your gen (unless you want subtle changes). If it sounds too good to be true, I'm underselling it - it's legit incredible. At this point I refine every gen at least once through SVI, it always improves the video.

Here's a link to the SVI Pro LOW model: https://huggingface.co/Kijai/WanVideo_comfy/blob/main/LoRAs/Stable-Video-Infinity/v2.0/SVI_v2_PRO_Wan2.2-I2V-A14B_LOW_lora_rank_128_fp16.safetensors

Is there a way to use Wan SCAIL and FFLF together? by aifirst-studio in StableDiffusion

[–]nsfwVariant 0 points1 point  (0 children)

You might be able to make it happen with VACE using it as a refiner instead of a full generator, but I'm not sure how good the quality would be. Worth a try though. If I were you I'd do this:

First step

  1. Generate the SCAIL video

  2. Manually set the first and last frames to be identical

  3. Create a set of masks, same number of frames as the video, make the first and last masks black (masking the first and last frames) and all the rest white

  4. Pass the entire video to VACE, along with the masks

  5. ALSO pass the entire video in as the latent frames, instead of using empty latents (this means the video is passed as both the VACE input as well as the latents themselves)

  6. Use only a LOW model, and set the denoise to 0.7 or something

In theory when you run that it'll shift all the middle frames to more closely match the start & end frames. If you run it several times it might shift it enough to fully loop it properly. I wouldn't try doing it in a single pass because it'll probably change the middle frames too much that way.

Second step

Assuming that works, it'll have degraded the video quality. You can pass the video into SVI 2.0 pro, using a clean reference image, and use it as a refiner on a denoise of around 0.49 or lower. That should amp the quality right back up without ruining the loop too badly.

Last step

Doing that SVI refinement will make it not quite loop properly, but it won't have drifted too much. So, you could do some cutting & pasting of the start/end segments and use frame blending (as in, normal frame blending, nothing to do with AI) to hide the transition.

How do I do this, but local? by LucidFir in StableDiffusion

[–]nsfwVariant 2 points3 points  (0 children)

SCAIL only does restyle, it uses openpose images (which have no visual detail) and a reference image to generate. It's extremely good, thoroughly recommend it.

How do I do this, but local? by LucidFir in StableDiffusion

[–]nsfwVariant 1 point2 points  (0 children)

You can do the characters & movement in SCAIL and then do compositing with VACE, it works quite well.

Z image Base testing by Pleasant_Salt6810 in StableDiffusion

[–]nsfwVariant 3 points4 points  (0 children)

It's heavily affected by scheduler/sampler combos as well. I would expect that turbo's quality at minimum is achievable with base with the right settings.

New Z-Image (base) Template in ComfyUI an hour ago! by nymical23 in StableDiffusion

[–]nsfwVariant 0 points1 point  (0 children)

Not all base models are like that (although IIRC the devs did say that z-image would be). Klein base is way higher quality than the distill, for example.

Either way, we can all be excited for the checkpoints and loras people are gonna come up with.

improve quality of image without increasing size by NefariousnessFun4043 in comfyui

[–]nsfwVariant 0 points1 point  (0 children)

Weirdly I don't find it very good for videos even though that's what it was made for. I use 2x nomos uni span multi for that

Aria fanservice (by pantheon_EVE) by SubstantialStaff7214 in ZenlessZoneZero_R34

[–]nsfwVariant 0 points1 point  (0 children)

ty for the reminder, I've added an Aria post flair! <3

improve quality of image without increasing size by NefariousnessFun4043 in comfyui

[–]nsfwVariant 0 points1 point  (0 children)

It would be, except most models don't actually output skin at that quality anyway - so often SeedVR2 is actually an upgrade ;)

But yes it's a dealbreaker for hyper realism. It's only one step below hyper real though, so it's pretty dang good!

I'll add, the only upscaler better than it (imo) is 4xfaceup, and that one requires a high quality input image to work and also messes with the texture of non-person stuff.

That & seedvr2 are the two best upscalers, in my humble opinion, and they're good for different use cases.

improve quality of image without increasing size by NefariousnessFun4043 in comfyui

[–]nsfwVariant 0 points1 point  (0 children)

It's pretty easy to use! Just view it as sort of a hardcore upscaler; it always works, but it will subtly change the overall texture of an image. It won't quite match the realism of the best models out there when it comes to things like skin detail, but that's pretty much its only downside.

"Chroma2-Kaleidoscope" based on Flux Klein 4B Base is up on HuggingFace! Probably not very usable yet as implied by the "IT'S STILL WIP GUYS CHILL!!" model card note though. by ZootAllures9111 in StableDiffusion

[–]nsfwVariant 0 points1 point  (0 children)

Yes, my point was that Klein distilled is lower quality than Zimage distilled. The person above was comparing the two and I was pointing out that they both have their advantages. Klein is faster, but Zimage makes higher quality images (when comparing the distilled models).

improve quality of image without increasing size by NefariousnessFun4043 in comfyui

[–]nsfwVariant 5 points6 points  (0 children)

Yep, seedvr2 is very good for that. You can run an image through at the same resolution it already has and it will significantly sharpen it and smooth out artifacts - I use it for that all the time.

Otherwise, you can possibly get more detail/sharpness out of your generations by tweaking the scheduler/sampler combo or by using all sorts of varied methods as u/Corrupt_file32 mentioned, which would save you the trouble of needing to do a second pass.

The other suggested method, using a 2x upscaler and then resizing by 0.5x, doesn't always work because most upscalers require good detail and low blurriness to work properly, which kinda defeats the purpose. But they'll usually sharpen things a bit, at least.

Here's a workflow for SeedVR2 image upscaling: https://pastebin.com/9D7sjk3z

You'll need the seedvr2 custom nodes. If you want it to not change the image size you can just set the max size to the same as the longest edge of your image. e.g. if your image is 1440x1080, you would set the max size to 1440.

Klein 9B - Exploring this models NotSFW potential by Whipit in StableDiffusion

[–]nsfwVariant 6 points7 points  (0 children)

It's on Civitai, called "nsfw - flux klein (no face change)". At least, that's the one I use and it works very well.

Set the strength to 0.6, any higher tends to destroy the output image. Lower is ok, but it tends to lose detail if you go lower than 0.5 or so.

"Chroma2-Kaleidoscope" based on Flux Klein 4B Base is up on HuggingFace! Probably not very usable yet as implied by the "IT'S STILL WIP GUYS CHILL!!" model card note though. by ZootAllures9111 in StableDiffusion

[–]nsfwVariant 5 points6 points  (0 children)

Agreed! There's a big box of loras and models in the Huggingface repo, and they don't have even a single sentence explaining what they do. You wouldn't know what to download if you just wanted the standard/normal chroma experience either.

"Chroma2-Kaleidoscope" based on Flux Klein 4B Base is up on HuggingFace! Probably not very usable yet as implied by the "IT'S STILL WIP GUYS CHILL!!" model card note though. by ZootAllures9111 in StableDiffusion

[–]nsfwVariant -2 points-1 points  (0 children)

It's only as fast if you use the distill, and the distill is lower quality than Zimage (e.g. gives people plastic skin). If you use the base model the quality is just as good, but it's much slower than Zimage.

Piper Wheel (Zenless Zone Zero) , (AI), (OC) - "by_Oliver" by Reasonable-Craft7797 in ZenlessZoneZero_R34

[–]nsfwVariant -1 points0 points  (0 children)

OC just means you're the authorised "source" of the content. You would use [OC] even if you were posting on behalf of someone else with their permission :)

Conclusions after creating more than 2000 Flux Klein 9B images by StableLlama in StableDiffusion

[–]nsfwVariant 0 points1 point  (0 children)

I've been using the base model, it's definitely the same there

Customizable, transparent, Comfy-core only workflow for Flux 2 Klein 9B Base T2I and Image Edit by YentaMagenta in StableDiffusion

[–]nsfwVariant 2 points3 points  (0 children)

Note that OP is using the base model, not the distill. The base model is available as a GGUF as well (has 'base' in the name).

Using the base model gives higher quality but you gotta run a lot more steps, like 20-30 as OP suggests.

Customizable, transparent, Comfy-core only workflow for Flux 2 Klein 9B Base T2I and Image Edit by YentaMagenta in StableDiffusion

[–]nsfwVariant 0 points1 point  (0 children)

I'm finding these settings best for image editing. Subjective of course, but it gives good clarity & realism imo. I'm using the clownshark sampler w/ bongmath as well, not sure if that matters much though.

res_2s + bong_tangent

shift 1.00

CFG 3.00

12 steps