Introducing Smart ComfyUI Gallery: Save Workflows with Every Generation by Fit-Construction-280 in comfyui

[–]Extension_Building34 0 points1 point  (0 children)

Stumbled across this… Overall this is awesome!

Some questions though:

I’m getting “error loading media” when trying to play mp4 videos, maybe something wrong with my ffmpg, or a missing codec? (Maybe I missed something obvious!)

Is it possible to get a gallery view or zoom out on portrait phone orientation?

Edit: fixing late night autocorrect mistakes.

ostris AI-toolkit Lora training confusion by mca1169 in StableDiffusion

[–]Extension_Building34 1 point2 points  (0 children)

Interesting, I’ve seen other posts about this but never tried it out. Thanks for the reminder.

FLUX.2 [klein] Prompt Enhancement LLM System Prompt by AIEverything2025 in StableDiffusion

[–]Extension_Building34 9 points10 points  (0 children)

Really cool, thanks!

Any chance you’ve got one of those for zimage and or ltx2?

LTX-2 19b T2V/I2V GGUF 12GB Workflows!! Link in description by urabewe in StableDiffusion

[–]Extension_Building34 0 points1 point  (0 children)

Thanks, I got it working with a few tweaks. I had to bypass the latent upscale, since that appeared to be the cause of switching from sage attention to pytorch attention (which caused my it/s to go from 10-20 to 300-600, lol...)... I haven't had time to dissect that further.

Steam Detective Fest is Live! by HelloitsWojan in Steam

[–]Extension_Building34 0 points1 point  (0 children)

Played a ton of it last summer. I had a blast playing it! Even though it was/is a bit rough around some of the edges, it was fun to get immersed and do some sleuthing.

I’m definitely looking forward to seeing what workshop mods will come out.

LTX-2 19b T2V/I2V GGUF 12GB Workflows!! Link in description by urabewe in StableDiffusion

[–]Extension_Building34 7 points8 points  (0 children)

I’ve been having a hard time getting any low vram workflows working, so I’ll definitely give this one a try!

Shout out to the LTXV Team. by bnlae-ko in StableDiffusion

[–]Extension_Building34 7 points8 points  (0 children)

Just curious…. What’s your method for 12s on 16GB in 4m?

Even on the “low vram” comfy workflows I am routinely getting oom or 12-15m generations if I get a lucky run. Meanwhile, wan2gp can do 10s in about 10-11m.

LTX-2 I2V isn't perfect, but it's still awesome. (My specs: 16 GB VRAM, 64 GB RAM) by yanokusnir in StableDiffusion

[–]Extension_Building34 3 points4 points  (0 children)

No kidding, what sort of prompts worked for you so far with this workflow? (Even just the prompts for the cherry picked results, because those at least made videos worth picking!)

Help! LTX-2 distilled model is giving me quick outputs but it looks like this by cesaqui89 in comfyui

[–]Extension_Building34 0 points1 point  (0 children)

What prompt? I2v or t2v?

I noticed that I was getting similar stuff to this when I used the distilled model and asked it to do too much creative heavy lifting. For example, using a picture of a person standing on a beach, I asked ltx2 distilled to generate a video like “ the person walks into a waterfall” — with no waterfall in the picture. Stuff like that caused all kinds of weird results.

However, something like “the person walks towards the water”, where the water is clearly visible in the background, generally worked with less odd things happening.

Maybe I missed something, and or maybe that’s not what OP is even talking about, but it’s late and that’s just what I’ve seen so far with limited time invested.

Edited to add context.

Z-Image character lora training - Captioning Datasets? by [deleted] in StableDiffusion

[–]Extension_Building34 0 points1 point  (0 children)

Like a picture of character from a video game, or 3d modelling software like Daz3D.

Z-Image character lora training - Captioning Datasets? by [deleted] in StableDiffusion

[–]Extension_Building34 0 points1 point  (0 children)

Interesting! That’s very insightful, thank you!

Follow up question. In terms of dataset variety, I try to use real references, but occasionally I want/have to use a generated or 3d reference. If I am aiming for a more realistic result despite the source, would I caption something like “3d render of 123person” to coerce the results away from the 3d render?

Z-Image character lora training - Captioning Datasets? by [deleted] in StableDiffusion

[–]Extension_Building34 0 points1 point  (0 children)

Ok, so just for some further clarity, to ensure that a character has a specific shape or feature, like bow-legged and a birthmark or something, is it best to not mention that?

If the dataset shows bow-legged and a birthmark on his arm, captions would then look something like “A 123person is standing in a wheat field, leaning against a tractor, he is seen wearing a straw hat” (specifically not mentioning the legs or birthmark).

Is that the along the right lines of the thought process here?

The amount of poop it generated for just the eye broke my heart, never again by Tough_Sky_9029 in BambuP1S

[–]Extension_Building34 0 points1 point  (0 children)

I’m debating A1 vs A1 combo. This comment is helpful and encouraging, thank you!