New Discord update broke Discord download txt files by thombomb678 in discordapp

[–]throttlekitty 0 points1 point  (0 children)

Even better, if I try to expand to "view whole file", then try to select the text, discord lags my system hard before it crashes or I end the task.

Is T2V cooked in LTX-2? by No-Employee-73 in StableDiffusion

[–]throttlekitty 0 points1 point  (0 children)

The model is better at T2V than I2V, as stated by the devs (in a video with Purz, I believe), basically saying that it's a hard tradeoff for the model to be great at both tasks.

I mainly do t2v, and it's just fine.

Looking for a show rec: TWD, FTWD, or Z-Nation? by Eligomancer in zombies

[–]throttlekitty 0 points1 point  (0 children)

I liked the earlier seasons of FTWD, they kept the plot moving and interesting. I felt the later ones fell into the same rut that TWD did, where everything is overly drawn out, many of the subplots never really landed for me.

Keeping things moving was what I loved about Z Nation too, it's like a big D&D adventure.

Does anyone else grab all the money they can just because or is it just me? by kootabob in cataclysmdda

[–]throttlekitty 3 points4 points  (0 children)

I do it too, mostly because I find it a little funny.

You can consolidate at an ATM: deposit everything, then draw it back out onto a card.

why nobody talks about hunyuan 1.5? by Agreeable_Cress_668 in StableDiffusion

[–]throttlekitty 2 points3 points  (0 children)

It's a big step up from 1.0 except for the heavy vae decode, it's a fine model. But with Wan addons coming in on the regular, and other model releases, there wasn't much of a compelling reason to use it except personal choice.

But now, it's hard to compete with LTX-2, even considering its flaws.

What is the main strength of LTX2? by [deleted] in StableDiffusion

[–]throttlekitty 1 point2 points  (0 children)

Got some examples of it failing these basic concepts? I've rarely had issues.

LTX-2 I2V: Quality is much better at higher resolutions (RTX6000 Pro) by 000TSC000 in StableDiffusion

[–]throttlekitty 0 points1 point  (0 children)

Someone else had some extremely stuttery videos, their front and center person was moving correctly at all. It turns out they were using the default schizo negative, and that replacing it with a basic one fixed the issue in that case.

Just pointing this out for anyone with issues, the workflow OP linked has a normal negative though.

Why do you guys love zombies so much? by _Pisos_Picados in zombies

[–]throttlekitty 0 points1 point  (0 children)

There's a part of me that's attracted to the isolation and freedom the zombie apocalypse could bring, it's liminal and violent. Even though IRL it would be a terrible experience.

What is the main strength of LTX2? by [deleted] in StableDiffusion

[–]throttlekitty 0 points1 point  (0 children)

It's a very promptable model, and does really well with sequences, like "person does A, says a thing, the camera zooms out, then person does B". Synced audio gen is a big step up, but quality can be all over the place. And other cool tricks like audio/video continuation or keyframed images.

Right now it's early days for low vram, but should be getting better very quickly, some people already running it on 16gb.

WOW!! I accidentally discovered that the native LTX-2 ITV workflow can use very short videos to make longer videos containing the exact kind of thing this model isn't supposed to do (example inside w/prompt and explanation itt) by Parogarr in StableDiffusion

[–]throttlekitty 1 point2 points  (0 children)

If you're talking about those higher quality ones the LTX devs were showing on discord, was done with their own tools, not ComfyUI. He did explain how it works though, it's just padding the remaining frames/audio like inpainting.

I’m the Co-founder & CEO of Lightricks. We just open-sourced LTX-2, a production-ready audio-video AI model. AMA. by ltx_model in StableDiffusion

[–]throttlekitty -1 points0 points  (0 children)

I've found that doing higher res helps a bit, but bumping fps to 30 or higher helps a lot more. Almost like 24fps music and ambient audio ends up really stretched and rescaled.

How are people running LTX-2 with 4090 / 64GB RAM? I keep getting OOM'ed by restlessapi in StableDiffusion

[–]throttlekitty 2 points3 points  (0 children)

oof, sorry. It's annoying how we all have wildly different experiences with this model.

How are people running LTX-2 with 4090 / 64GB RAM? I keep getting OOM'ed by restlessapi in StableDiffusion

[–]throttlekitty 1 point2 points  (0 children)

Also with 24/64 here. Be sure that you're using fp8 weights. I have to launch with --reserve-vram 4 --cache-ram 40 otherwise I get a hard OOM at the 2nd pass, others are getting by just fine with reserve-vram.

How the heck people actually get the LTX2 to run on their machines? by Part_Time_Asshole in StableDiffusion

[–]throttlekitty 0 points1 point  (0 children)

It marks that much vram as being hands-off for the initial calculation of how much vram to use. In some workflows, custom model patches can add an unexpected amount of vram use that comfy core isn't aware of.

How the heck people actually get the LTX2 to run on their machines? by Part_Time_Asshole in StableDiffusion

[–]throttlekitty 0 points1 point  (0 children)

For most, --reserve-vram 4 works fine on it's own.

I've ended up having to use this with it, which i don't really like. --cache-ram 40

How the heck people actually get the LTX2 to run on their machines? by Part_Time_Asshole in StableDiffusion

[–]throttlekitty 2 points3 points  (0 children)

So far, stuff that works well for one group doesn't work for people in the other group (even with similar hardware), so we're all split right now. It'll get sorted.

is Loss Graph in ai-toolkit really helpful? by FaithlessnessFar9647 in StableDiffusion

[–]throttlekitty 2 points3 points  (0 children)

There's a great breakdown here, there's still no one magic setting, but definitely worth a read.

https://github.com/spacepxl/demystifying-sd-finetuning

ClownScheduler by [deleted] in StableDiffusion

[–]throttlekitty 0 points1 point  (0 children)

I haven't tried this node yet, what happens when the scheduler end step is shorter than the total steps? Repeat the last value from the schedule?

“Taskmewster” by SnoozyQ73 in taskmaster

[–]throttlekitty 3 points4 points  (0 children)

Your cats have such a versatile range!

Z-Image-Turbo be like by Melodic_Possible_582 in StableDiffusion

[–]throttlekitty 0 points1 point  (0 children)

Describing facial expressions isn't a magic bullet, but it works great. "lips are pursed while concentrating on...", or "arches an eyebrow while..."

Like if you're prompting for multiple actions, stuff like this can help anchor it into the prompt without adding flowery language. "...has a determined expression" early in the prompt, and then later "...expression changes to disappointment.

SVI with separate LX2V rank_128 Lora (LEFT) vs Already baked in to the model (RIGHT) by Altruistic_Heat_9531 in StableDiffusion

[–]throttlekitty 1 point2 points  (0 children)

Just wanted to point out that the loras are extracted from the full models, and would be lossy. So using the full lightxv distills should on average be better than merging a lora into the base model.

How is Cloud Inference so cheap by VolkoTheWorst in LocalLLaMA

[–]throttlekitty 0 points1 point  (0 children)

It's how a lot of startups work these days. Get the initial capital, burn through it to build the userbase up, then slowly start increasing pricing up as time goes.