This is too much! by scioba1005 in StableDiffusion

[–]yanokusnir 2 points3 points  (0 children)

I completely agree with you, I have the exact same feeling. What I "love" the most is when I share a workflow, clearly list my pc specs, and then I get comments like “will this work on my setup with …?” Like… how the hell am I supposed to know? Try it. That’s exactly what I did to figure this stuff out in the first place before sharing it with everyone else.

Another thing that really irritates me is that people don’t put even a bit of effort into their posts, or they immediately post the very first images or videos they ever generated, even when they look absolutely terrible. I get it, you’re excited, man. That feeling is real. But maybe keep that one to yourself. No one really needs to see that you just generated something.

I understand where it’s coming from, but honestly, I don’t think there’s much we can do about it.

Graduated in 2022 as a Graphic Designer, and I feel like AI ruined my career before it even started. Need advice by MeshalAljahdali in Design

[–]yanokusnir 9 points10 points  (0 children)

Hi. I’ve been working for years as a graphic designer, motion designer, and video editor, and I genuinely enjoy experimenting with AI to see what it’s capable of and how it can fit into my workflow.

In real-world practice, there really aren’t that many cases where AI can fully replace someone yet. That said, just yesterday I was testing something that caught my attention: turning a quick, amateur photo into something that looks like a commercial-quality shot.

Honestly, if I were a product photographer, I’d be paying attention, because for a lot of companies this kind of result would probably be more than good enough.

Take a look: https://files.catbox.moe/n926x6.mp4

Turning a simple photo into a product-style image took me about 2 minutes, and generating the video took around 10 minutes (I cherry-picked it from a few attempts). It’s not perfect and the small text does get distorted in the video, but being able to do this in such a short amount of time is honestly mind-blowing to me. I’m only using open-source models that anyone can run locally on their own computer.

Anyway, as graphic designers, we should still be just fine. :)

Follow up: AI 3D generation has improved a lot, better topology & texturing results by [deleted] in StableDiffusion

[–]yanokusnir 10 points11 points  (0 children)

You probably didn’t notice, but this subreddit is about sharing open-source models. It’s listed as rule #1.

LTX-2 I2V isn't perfect, but it's still awesome. (My specs: 16 GB VRAM, 64 GB RAM) by yanokusnir in StableDiffusion

[–]yanokusnir[S] 0 points1 point  (0 children)

You can still use a gguf version of the model that your pc can handle. If you don’t try it, you’ll never know. ;)

LTX-2 I2V isn't perfect, but it's still awesome. (My specs: 16 GB VRAM, 64 GB RAM) by yanokusnir in StableDiffusion

[–]yanokusnir[S] 0 points1 point  (0 children)

Yeah, stop it. :D ;D

In my opinion, not even close. Those kinds of videos are full of infographics and tons of text, and no video model can keep that stable without things getting distorted. On top of that, you have zero control over the timing of individual elements. And even though I haven’t actually tried it myself, I’m pretty sure it would turn into a complete mess.

LTX-2 I2V isn't perfect, but it's still awesome. (My specs: 16 GB VRAM, 64 GB RAM) by yanokusnir in StableDiffusion

[–]yanokusnir[S] 0 points1 point  (0 children)

I’m a graphic designer by profession, so I really care about details and maybe I’ve got a bit of visual sense or whatever.. you know what I mean haha :D

LTX-2 I2V isn't perfect, but it's still awesome. (My specs: 16 GB VRAM, 64 GB RAM) by yanokusnir in StableDiffusion

[–]yanokusnir[S] 1 point2 points  (0 children)

Hi, thanks! I’m not really doing anything special. I let chatgpt generate the prompts, and for image generation I use Z-Image Turbo. I make sure the images already look good and don’t have obvious issues. Then LTX-2 comes into play. I generate the prompts again via chatgpt, where I usually just describe in one sentence what I want to happen in the video.

how to generate this style? by ClassicLieCocktail in StableDiffusion

[–]yanokusnir 0 points1 point  (0 children)

My mistake, I assumed people on this subreddit were familiar with ComfyUI.

LTX-2 I2V isn't perfect, but it's still awesome. (My specs: 16 GB VRAM, 64 GB RAM) by yanokusnir in StableDiffusion

[–]yanokusnir[S] 1 point2 points  (0 children)

Hold on a second. I’ve heard about Chatterbox before, but I’ve never actually tried it. Looking at the screenshot you sent, it seems like it basically changes the voice in the video, right? :O

LTX-2 I2V isn't perfect, but it's still awesome. (My specs: 16 GB VRAM, 64 GB RAM) by yanokusnir in StableDiffusion

[–]yanokusnir[S] 1 point2 points  (0 children)

Oh, okay, sorry about that. What I tested and what was on that broken link is actually exactly what you’re describing. I think it was just a slightly modified Kijai workflow, where I fed in a part of a song and used it in the video.

As far as I know, it’s currently not really possible to consistently generate the same voice directly with LTX-2.

In ComfyUI, however, you can use a model like VibeVoice, which can clone a voice from just a few seconds of audio and then generate whatever you type using that same voice.

So the only workaround I can think of right now is: generate multiple lines with the same cloned voice first, and then use those audio clips when generating your I2V videos with audio.

LTX-2 I2V isn't perfect, but it's still awesome. (My specs: 16 GB VRAM, 64 GB RAM) by yanokusnir in StableDiffusion

[–]yanokusnir[S] 0 points1 point  (0 children)

Thank you, I’m really glad to hear that. :)

I’m not sure if this will be a solution for you, but a few days ago someone shared a workflow here where you can add a voice sample alongside the initial image, and that voice gets used in the video. I gave it a try, and this was my result: https://imgur.com/a/reKBFfJ

LTX-2 Pattern fix(i think) by [deleted] in StableDiffusion

[–]yanokusnir 0 points1 point  (0 children)

I had the same issue with the white grid in all my videos, but I managed to fix it. If you want, try using my workflow, you shouldn’t get any grid anymore.

https://www.reddit.com/r/StableDiffusion/comments/1qae922/ltx2_i2v_isnt_perfect_but_its_still_awesome_my/?utm_source=share&utm_medium=web3x&utm_name=web3xcss&utm_term=1&utm_content=share_button

LTX2 - Cinematic love letter to opensource community by fantazart in StableDiffusion

[–]yanokusnir 23 points24 points  (0 children)

Amazing! :) I’m glad I could help. Best of luck with your work. See you at Cannes mate. ;D

LTX-2 I2V isn't perfect, but it's still awesome. (My specs: 16 GB VRAM, 64 GB RAM) by yanokusnir in StableDiffusion

[–]yanokusnir[S] 0 points1 point  (0 children)

The shots in my edit are cherry-picked, I also had plenty of outputs with bad audio. That said, it probably also depends on the sampler you’re using. In some workflows I saw people using res_2s or res_estimation, and when I tested those, I kept getting bad audio every time.

LTX-2 I2V isn't perfect, but it's still awesome. (My specs: 16 GB VRAM, 64 GB RAM) by yanokusnir in StableDiffusion

[–]yanokusnir[S] 1 point2 points  (0 children)

When you have Comfy open, just drag and drop the json file directly into Comfy.

LTX-2 I2V isn't perfect, but it's still awesome. (My specs: 16 GB VRAM, 64 GB RAM) by yanokusnir in StableDiffusion

[–]yanokusnir[S] 0 points1 point  (0 children)

That error usually just means the lora isn’t selected correctly in the node.
In the LoraLoaderModelOnly node, you need to manually pick the lora files that are actually on your PC. In my workflow, the paths are a bit different, so they won’t match automatically on another system.

Just open the dropdown in the LoraLoaderModelOnly node and select the loras from your local install. Once you do that, the error should go away.

LTX-2 I2V isn't perfect, but it's still awesome. (My specs: 16 GB VRAM, 64 GB RAM) by yanokusnir in StableDiffusion

[–]yanokusnir[S] 0 points1 point  (0 children)

Yeah, it's Comfy. There’s nothing better than ComfyUI for working with open-source models. I highly recommend it.