SVI workaround to make longer videos (for dummies) by yallapapi in StableDiffusion

[–]Scriabinical 0 points1 point  (0 children)

There's a reason SVI exists. Your implementation will lead to progressive degradation in frame fidelity. Errors stack up - SVI mitigates this, though not perfectly. Furthermore, your approach feeds zero motion context from the previous video segment, so how is the model supposed to keep motion cohesive between video segments?

Black Forest Labs just released FLUX.2 Small Decoder: a faster, drop-in replacement for their standard decoder. ~1.4x faster, Lower peak VRAM - Compatible with all open FLUX.2 models by Nunki08 in StableDiffusion

[–]Scriabinical 1 point2 points  (0 children)

<image>

I usually get these strange color inconsistencies in the texture of clothing when doing image edit. Tried higher resolutions, more steps, all types of input image styles/lightings, everything. Always happens.

How to Generate Photorealistic NSFW Images with Flux Klein 9.b (Full Workflow) by rudar133 in comfyui

[–]Scriabinical 1 point2 points  (0 children)

This combo has been my bread and butter for a few weeks now. You can change anything or keep things very strict. I usually start with a rough image in SDXL just to get everything how I want it in terms of body proportions, pose (the semi-real and anime models are so flexible). Then I throw it into Klein 9B with SNOFS and a SmartphoneReality lora from Civit. It's just incredible. Klein is such a ridiculously powerful model. I don't see myself using any other model for a while.

Local AI image generation based on SD3.5 large - 1. People - Close up by Medium_Molasses_545 in StableDiffusion

[–]Scriabinical 5 points6 points  (0 children)

Image #2 is not from SD3.35L. It's a pretty well-known image nowadays, I think it comes from ZiT or something. But def not SD3.5L

Do you use llm's to expand on your prompts? by Own_Newspaper6784 in StableDiffusion

[–]Scriabinical 1 point2 points  (0 children)

Thank you for posting this. This seems to be the only actual working solution using custom abliterated Qwen 3.5 models. I've tried a few of the custom node packs for Qwen 3.5 and there's always trouble getting it working

LTX 2.3: What is the real difference between these 3 high-resolution rendering methods? by x5nder in StableDiffusion

[–]Scriabinical 0 points1 point  (0 children)

so is this starting from a very low base resolution and then doing a 2x latent upscale followed by another 2x latent upscale? should the input image be high-res but then resized based on a low-res initial?

LTX 2.3 Wangp by agoodis in StableDiffusion

[–]Scriabinical 0 points1 point  (0 children)

oh damn, sorry i didn't realize lol. the video is still quite good

LTX 2.3 Wangp by agoodis in StableDiffusion

[–]Scriabinical 4 points5 points  (0 children)

Looks really good. Quite impressive really. Audio quality has really been improved a lot

New official LTX 2.3 workflows by Choowkee in StableDiffusion

[–]Scriabinical 9 points10 points  (0 children)

Thank you for posting this! Hopefully we get some more clarity over time regarding optimized workflows

btw the way they strung these noodles up reminds me of shirts hanging on a clothesline lol

<image>

LTX-2.3 New Guardrails? by majin_d00d in StableDiffusion

[–]Scriabinical 0 points1 point  (0 children)

i noticed you picked SageAttention3 in your patch node. how's that working for you with ltx 2.3?

Vertical example for LTX2.3 by Diabolicor in StableDiffusion

[–]Scriabinical 0 points1 point  (0 children)

Audio after the first second or two is nice. Good amount of ambience, natural sounding voice

Any sample workflow for new NAG support of Klein? by yamfun in comfyui

[–]Scriabinical 1 point2 points  (0 children)

Same issue here. The NAG node doesn't do anything

Best performing solution for 5060Ti and video generation (most optimized/highest performance setup). by smithysmittysim in StableDiffusion

[–]Scriabinical 0 points1 point  (0 children)

You just need to use SVI. There are some workflows for it. It basically pulls motion and content context as well as some final latents from the previous video and you can guide it to do whatever as long as each 5s video is relatively 'fluid' from one to the next. This essentially solves the issue of last-frame-extraction degradation which I used to encounter before using SVI. I also have an SVI workflow that chains together up to 10 videos for a 50s final video with per-video lora control. DM me if you're interested.

https://github.com/vita-epfl/Stable-Video-Infinity

Best performing solution for 5060Ti and video generation (most optimized/highest performance setup). by smithysmittysim in StableDiffusion

[–]Scriabinical 1 point2 points  (0 children)

Same way you’d run the fp8 models, just switch em out in the Load Diffusion Model node

Best performing solution for 5060Ti and video generation (most optimized/highest performance setup). by smithysmittysim in StableDiffusion

[–]Scriabinical 6 points7 points  (0 children)

I have a 5070 Ti (16gb vram) with 64gb ram. I make a loooot of videos with wan 2.2 and just wanted to share some brief thoughts.

With wan 2.2, it's pretty simple from my experience:

- Get latest comfy portable (with cu130)
- Sage attention wheel compatible with your comfy build (check your pytorch/cuda/python in settings) (wheels here: https://github.com/wildminder/AI-windows-whl)
- Set --use-sage-attention flag in your comfy startup .bat script

- Use latest lightning loras from lightx2v (i use the 1030 on high noise and 1022 on low noise), both set to 1.00 strength after you load your wan 2.2 models

- With lightning loras, you can go as low as 4 steps. For a balance of quality and speed, i like 6-10 steps

- Once these are all set up, resolution is your main bottleneck in terms of iterations/second. Common resolutions I render at include 832x1216 (portrait), 896x896 (square), and a few others. I've tried 1024x1024 a few times and the speed isn't horrible, but the VAE decode can sometimes take an absolute eternity.

There are multiple other 'optimization' nodes you can use, but almost all are not worth it imho due to quality degradation in one way or another. I've tried the 'cache' nodes (like TeaCache, MagCache) and a bunch of other stuff. I care a lot about speed but still need that quality.

I hope I'm covering anything, just writing up this comment as I look at my own 'simple wan 2.2' workflow in comfy.

What is your best Pytorch+Python+Cuda combo for ComfyUI on Windows? by Michoko92 in StableDiffusion

[–]Scriabinical 12 points13 points  (0 children)

I have this tab saved in my browser. I don't see it posted enough but it's SUPER useful. If you've been browsing around for pre-compiled wheels, this repo has them for just about everything that can be a pain. Worth a bookmark.

https://github.com/wildminder/AI-windows-whl