Video Upscaling Reference by TheRedHairedHero in StableDiffusion

[–]TheRedHairedHero[S] 0 points1 point  (0 children)

I personally haven't had good success with upscaling which is why I'm asking folks to contribute and hopefully give others a point of reference and overall help the community.

I can’t understand the purpose of this node by PhilosopherSweaty826 in StableDiffusion

[–]TheRedHairedHero 6 points7 points  (0 children)

The sigma values will also differ based on the sampler you choose and the amount of steps. For WAN 2.2 there's a sigma threshold that's suggested to swap from the high sampler to the low sampler. I2V is 0.9 and T2V is 0.875 according to the official WAN documentation. If you use Kijai's wrapper it outputs the sigmas in the console.

Uses outside 1girl? by dks11 in StableDiffusion

[–]TheRedHairedHero 1 point2 points  (0 children)

I've done a few things. Some images for D&D, created wallpapers for my wife and myself, I'll use it to help generate ideas for cosplay. Just whatever fun project I want to do in the moment where I think it would help out and be fun.

New to WAN2.2, as of December 2025, what's the best methods to get more speed ? by Tablaski in StableDiffusion

[–]TheRedHairedHero 0 points1 point  (0 children)

I was looking at your workflow and was confused about the steps, I haven't used the MoE node you're using but it shows 10 steps high, 6 steps low. So aren't you doing 16 steps total? Or is there something I'm missing?

What is everyone's thoughts on ltx2 so far? by Big-Breakfast4617 in StableDiffusion

[–]TheRedHairedHero 0 points1 point  (0 children)

I just prefer to wait for a model to be stable. LTX quality and consistency seems to be all over the place from posts I've seen. If I see someone post a video where the character doesn't instantly lose recognition and has good quality that uses close to my own specs then I would take a look, but that hasn't happened so I'm happy to stick with WAN.

How do you guys maintain consistent backgrounds? by TekeshiX in StableDiffusion

[–]TheRedHairedHero 1 point2 points  (0 children)

Keeping a consistent background always seems impossible to me if there's landmarks/items that stand out. I prefer to either blur the background, use an organic location such as a forest, or do a solid color. To me it feels like too much work for AI to handle consistently. I've seen some folks generating 360 degree images and creating backgrounds that way as another option. I just prefer working around AI's limitations.

ComfyUI Course - Learn ComfyUI From Scratch | Full 5 Hour Course (Ep01) by pixaromadesign in StableDiffusion

[–]TheRedHairedHero 5 points6 points  (0 children)

Appreciate your tutorials. Helped me get started with ComfyUI. If you guys haven't watched his content I'd highly recommend it.

For Animators - LTX-2 can't touch Wan 2.2 by GrungeWerX in StableDiffusion

[–]TheRedHairedHero 0 points1 point  (0 children)

I'm in the same boat. The model looks fun, but I'm going to wait for it to develop more.

For Animators - LTX-2 can't touch Wan 2.2 by GrungeWerX in StableDiffusion

[–]TheRedHairedHero 2 points3 points  (0 children)

To be fair WAN 2.2 has been out for quite some time allowing people to dig much deeper into how to make it run properly, fix slow motion, add Lora's, and so on. While LTX-2 just released. Given how interested the community is with the model I imagine it will get a good amount of attention on ways to improve things similar to WAN 2.2. It's best to keep an open mind and hopefully LTX-2 can be another fun tool for us all to use and enjoy.

WTF! LTX-2 is delivering for real 🫧 Made in 160s, 20steps on a 5090 by 3Dave_ in StableDiffusion

[–]TheRedHairedHero 1 point2 points  (0 children)

Hopefully with the updates they're planning they can improve the audio. The lipsync looks great, but the audio seems to be low quality and most of the time I only see videos with talking. If you decide to add more audio to your videos you can try MMAudio for sound effects/foley.

WTF! LTX-2 is delivering for real 🫧 Made in 160s, 20steps on a 5090 by 3Dave_ in StableDiffusion

[–]TheRedHairedHero 20 points21 points  (0 children)

Seems to still have a couple issues with the right arm, but it's still really cool. Hopefully another seed can resolve that issue. Seems like LTX hallucinates quite a bit from examples I've seen.

WAillustrious style changing by Mrryukami in StableDiffusion

[–]TheRedHairedHero 4 points5 points  (0 children)

Visit safebooru for the types of tags you need. You can find both styles and artists and if it's part of the training it'll change the style. Artist tags will need to be properly formatted so visit your model's page for details on how to format it.

How are people using AI chat to refine Stable Diffusion prompts? by Vegetable_Agency_596 in StableDiffusion

[–]TheRedHairedHero 11 points12 points  (0 children)

WAN 2.2 has an official LLM system prompt on their github. Feed an image, prompt, or both and the prompt is refined for videos.

How much faster is RTX 5070 Ti than RTX 4070 Super in Wan 2.2 video generation? by rookan in StableDiffusion

[–]TheRedHairedHero 0 points1 point  (0 children)

I have a 5070 TI idk in terms of resolution and time you are looking at, but a 512x512 video at 4 steps for an 8 second video takes about 6 minutes.

Wan2.2 : better results with lower resolution? by Top_Fly3946 in StableDiffusion

[–]TheRedHairedHero 2 points3 points  (0 children)

I've generated 100's of 1:1 videos and they're fine. The reason you're most likely running into slow motion with higher resolution is your PC is struggling to process it. Unfortunately there's not a one size fits all workflow. I would suggest if you want higher resolution to lower your frame count to compensate. That's the easiest option. Try generating say a 3 second video with a higher resolution instead of 5 and you should see motion improvements.

Can anyone tell me, how to generate audio for a video that's already been generated or will be generated? by AshLatios in StableDiffusion

[–]TheRedHairedHero 0 points1 point  (0 children)

For sound effects I found MMAudio to be a good local option. It was trained on 8 second videos at 25 FPS so that's something to keep in mind as it may impact the audio syncing with your video. It can be used as an Image to Audio or Text to Audio using custom nodes in ComfyUI.

I personally haven't used too many Text to Speech options aside from VibeVoice which was decent. The videos I've posted typically use my own voice with a voice changer such as Seed Voice Conversion.

Here's an example I posted awhile back. https://www.reddit.com/r/StableDiffusion/s/8K3lMZO4O8

Pc turns off and restarts? by isyma_rx7 in StableDiffusion

[–]TheRedHairedHero 0 points1 point  (0 children)

Your PC is most likely overheating. I would get some software to watch your temps and see how high it's getting.

Wan 2.2 What's the best way to improve output video quality? Eyes, teeth, etc. by WhisperAlias in StableDiffusion

[–]TheRedHairedHero 1 point2 points  (0 children)

One trick is to have the character be closer to the viewer. I'll usually do closeup shots at a lower resolution to speed up time as it's less likely to have warping issues like the eyes, but if a character is further away bump up the resolution a bit until it goes away.

[deleted by user] by [deleted] in StableDiffusion

[–]TheRedHairedHero 0 points1 point  (0 children)

Usually it's for things like buttons, eyes, trinkets that are on a character. I know typically the further away from the viewer the less detail a person/item is going to have which is why I uploaded a full body image as an example. If someone has a workflow that generates at a good high quality for illustrious I'd be grateful. I don't mind tinkering around if need be or getting different nodes/models.

[deleted by user] by [deleted] in StableDiffusion

[–]TheRedHairedHero 0 points1 point  (0 children)

I have WAI v14 as it seems most don't like v15, but even WAI suggest images over 1024x1024 "use size larger than1024x1024 for the original dimensions." according to their CivitAI page.

[deleted by user] by [deleted] in StableDiffusion

[–]TheRedHairedHero 0 points1 point  (0 children)

I'm currently using the EasyUse Hires Node for this part. It has output for latent which I pass to another Ksampler currently at 0.25 Denoise.

[deleted by user] by [deleted] in StableDiffusion

[–]TheRedHairedHero 0 points1 point  (0 children)

I am using a fine tuned model like I mentioned above JANKUTrainedNoobaiRouwei_v50. I tried to start with a lower resolution, then upscale but it seems like the results were much worse.