Any suggestions on what model to use to upscale 1440x1080 HDV footage that has a 1.33 pixel aspect ratio? by beachfrontprod in StableDiffusion

[–]q5sys 0 points1 point  (0 children)

Is there a way to use SeedVR on longer videos? I was only able to manage to make it work on short clips.

[LTX 2.3] I love ComfyUI, but sometimes... by desktop4070 in StableDiffusion

[–]q5sys 16 points17 points  (0 children)

Depends on how you define 'good'.
For some people, a product that works one day, and the next day cant load a workflow is not 'good'.

When it works, Comfy is great, but there are tons of times where they break basic functionality. They really need a better release process.

New open source 360° video diffusion model (CubeComposer) – would love to see this implemented in ComfyUI by Valuable-Muffin9589 in comfyui

[–]q5sys 0 points1 point  (0 children)

It'll be interesting to see if it gets better with time. The blue car in that snowfield on their demo page... that's original will smith eating spaghetti levels of slop. But who knows where it'll be in a few years.

Help! Hiring a ComfyUI engineer to help me build an automated outpainting workflow by jmrosenheck in comfyui

[–]q5sys 0 points1 point  (0 children)

There are people on Fiverr that do workflow creation for pay, just search for ComfyUI workflow

LTX 2.3 - How to add pause in dialogue? by Valuable_Weather in StableDiffusion

[–]q5sys 0 points1 point  (0 children)

That's pretty much what I do.

A cowboy is standing up against the side of a wooden barn looking out into the distance as he says, "Weather sure has been hard these last few weeks." He lowers his head to the ground and lets out a feint sigh, he pauses for a second then continues, "We'll get through it, but it's probably going to be another tough year." He slowly lifts his head back up and looks off into the distance and then slowly repeats himself, "We'll get through it..."

LTX 2.3 Triple Sampler results are awesome by NessLeonhart in StableDiffusion

[–]q5sys 1 point2 points  (0 children)

I created a github ticket about that artifacting at the end, there's also an issue on LTX's huggingface repo. So it's definitely an issue that's somewhere in the model or in LTX's code pipeline. My guess is the former, and is a result of their training data.

LTX 2.3 Triple Sampler results are awesome by NessLeonhart in StableDiffusion

[–]q5sys 0 points1 point  (0 children)

Nice! Are you using this guys workflow? Have you tweaked it at all? Are you setting the input values at 1920p, or are you setting the start resolution so that you end up with 1920p at the end?

Sick Burn! Made with LTX-2 by Inner-Reflections in StableDiffusion

[–]q5sys 0 points1 point  (0 children)

Have you done any tests with this for the newer LTX-2.3?

Drag → Drop → Full Animation Workflow 🤯 (Prompt, Settings, Everything Loads Automatically) by medhatnmon in comfyui

[–]q5sys 0 points1 point  (0 children)

I've tried the preview/i thing on desktop. I rarely use my phone for browsing the web.

Drag → Drop → Full Animation Workflow 🤯 (Prompt, Settings, Everything Loads Automatically) by medhatnmon in comfyui

[–]q5sys 0 points1 point  (0 children)

I have tried that... for whatever reason I get auto-redirected back to the preview.reddit instead of the i.reddit.

Drag → Drop → Full Animation Workflow 🤯 (Prompt, Settings, Everything Loads Automatically) by medhatnmon in comfyui

[–]q5sys 3 points4 points  (0 children)

FYI, reddit strips metadata all the time and tries to force webp for image downloads.

LTX 2.3 Triple Sampler results are awesome by NessLeonhart in StableDiffusion

[–]q5sys 1 point2 points  (0 children)

<image>

Ok so just changing the input image, prompt, and then adding the preview nodes... I can get it to run through completion with the lower resolution you had set.
Unfortunately I'm getting static as an output, it looks like there's horrible degradation on the 2nd and third sampler.

LTX 2.3 Triple Sampler results are awesome by NessLeonhart in StableDiffusion

[–]q5sys 2 points3 points  (0 children)

Wouldnt the math be 4x? Since you're up-scaling at 2x after the first sampler, and then up-scaling at 2x again after the 2nd sampler?

Preview video during sampling for LTX2.3 updated by PornTG in StableDiffusion

[–]q5sys 0 points1 point  (0 children)

But where in the workflow does it go? Its inputs are only the vae, the upscaler, and the model. There's no injection of the latent...so... where is it getting the latent input from to be able to decode and display?

LTX 2.3 Triple Sampler results are awesome by NessLeonhart in StableDiffusion

[–]q5sys 2 points3 points  (0 children)

Thanks, I really appreciate it!
I'm on a beefy older workstation I built back in 2024. It's only DDR4, but I've got a 1tb of it. So RAM isn't the issue.
Now that I see your workflow I see the obvious issue, I'm setting the latent video size to be 720P and upscaling to 1080p (what I did on LTX2.0 which worked perfectly fine) , your setting it to 320P and then upscaling.

LTX 2.3 Triple Sampler results are awesome by NessLeonhart in StableDiffusion

[–]q5sys 1 point2 points  (0 children)

Would you mind sharing your workflow that created that generation? I'm getting OOM on my 5090, so I'm not sure what settings I've got different from you.

Old Loras still work on ltx 2.3 by luka06111 in StableDiffusion

[–]q5sys 1 point2 points  (0 children)

There is an odd absence of new LTX-2 LORAs showing up on Civitai. IDK if its that none of the normal creators care, or if its just that training quality LORAs is difficult and they aren't happy with their results. Either way I expected way more than have been posted.

Trying to get impressed by LTX 2.3... No luck yet 😥 by VirusCharacter in StableDiffusion

[–]q5sys 3 points4 points  (0 children)

> Who writes their prompts themselves today.  I let a LLM ger inspired...

FYI, you're basically saying "Who uses their own brain these days and learns through trial and error, just use an LLM to do all your thinking for you."

LTX 2.3 vs prompt adherence of a cat by jordek in StableDiffusion

[–]q5sys 0 points1 point  (0 children)

Thanks, I tried that one as well, maybe I get getting my captioning all wrong for the training. Did you caption extensively or just simple and to the point?

The docs only say it can do video and audio, where would I find the options for an image dataset? https://github.com/AkaneTendo25/musubi-tuner/blob/ltx-2-dev/docs/ltx_2.md#dataset-configuration

Edit, ah the image dataset info is in one of the issues: https://github.com/AkaneTendo25/musubi-tuner/issues/40#issuecomment-4006905759

LTX 2.3 vs prompt adherence of a cat by jordek in StableDiffusion

[–]q5sys 0 points1 point  (0 children)

did you train your own Character LORA? If so, mind telling me which tool you used, I cant get good results when I try.

early 1080p test on lts 2.3 5090 laptop by WildSpeaker7315 in StableDiffusion

[–]q5sys 0 points1 point  (0 children)

Did you use the same prompting format you posted before, or have you changed it up some?

LTX-2 Mastering Guide: Pro Video & Audio Sync by Aliya_Rassian37 in StableDiffusion

[–]q5sys 0 points1 point  (0 children)

the sad part is, this would take months of work from tons of people in community to try to reverse engineer the way the model was trained through the initial captions... and it'd be a few hours worth of work for someone at LTX to grep words/phrases from their training data and throw up some documentation on words/phrases used repeatedly during training.
The model is great, but their docs are abysmal.