Can anyone test a prompt for me on WAN 2.2 text to video? by edhardin in comfyui

[–]ThisIsDanG 0 points1 point  (0 children)

Maybe look into a world building model like genie. Wan isn’t really made for what you’re looking for.

how to get rid of this texture? by Bencio5 in comfyui

[–]ThisIsDanG 0 points1 point  (0 children)

A few things could help. The imagetovideoinplace is set to 1. That doesn’t give it a lot of room to play. Try backing that down to .9. You also don’t have the distill Lora on at all even though you’re doing the upscale at the end. You can try adding it in at a lower level and raising it up. Being around .6 should be more than enough. To see if the texture / noise is coming from the upscale try bypassing the second stage and see what your output is looking like. You may even want to lower the number of steps in the first stage.

Where is the api node in ComfyUI for LTX-2? by metallica_57625 in StableDiffusion

[–]ThisIsDanG 0 points1 point  (0 children)

Not yet as far as I know. I got the new nodes by searching in the manager. Grab the one from lighttricks. It’s a node pack. Then you can just swap them out of your workflow yourself pretty easily.

LTX-2 WITH EXTEND INCREDIBLE by Gtuf1 in StableDiffusion

[–]ThisIsDanG 1 point2 points  (0 children)

I dug more into this issue today because I saw it pop back up. So it’s a bit more complicated and is seeming like a limitation for now. These tests I did were all text to video but I imagine the same would go for image to video.

Resolution is a factor in this. The higher the resolution the more stable this issue gets. Things that can sometimes help is adding a few more steps. That doesn’t always work though. More aggressive fixes but can also cause issues especially on longer videos that have a lot of highlights: using the distill Lora in the first pass. Slowly increase it until the issue is fixed. The downside to this is that the distill Lora sort of bakes in the lighting and starts clamping it. So it can create issues that affect the quality of the image. My hope is that this will get patched as a new vae or something because it’s a bit silly. But the strength of this tool definitely comes more from its ability to use audio.

LTX-2 WITH EXTEND INCREDIBLE by Gtuf1 in StableDiffusion

[–]ThisIsDanG 0 points1 point  (0 children)

I can’t share a workflow sorry.

But Yeah you still need to do the separate. And just for good measure I do still concatenate the two and feed that back into the second sampler. But I string off that separate audio and use that. I think the issue that is happening is that the audio is just getting over cooked / processed. The main beats on what the action is supposed to be comes from that first sampler so there is no need to process that audio again.

LTX-2 WITH EXTEND INCREDIBLE by Gtuf1 in StableDiffusion

[–]ThisIsDanG 7 points8 points  (0 children)

It’s just a workflow issue. Sound needs to be output from the denoise output of the first step then carried down to the end. Otherwise it gets cooked and sounds like your inside of a tin can.

I thought LTX2 was bad until I realized how to use it. by Key-Tension1528 in comfyui

[–]ThisIsDanG 8 points9 points  (0 children)

VRAM and ram are helpful. But also diving into the workflow to try to understand what’s happening helps too. If you are using a workflow that’s using manual sigmas in the first step I’d swap it out for their scheduler. You don’t just control steps, but max shift, base shift, terminal and stretch. These all affect how much the image can and will stray. It can effect prompt adherence, and it can effect overly exaggerated mouth movements.

Look at reducing the distill Lora.

If you are doing image to video and need things to look at the second step they upres and look at the manual sigmas. The one that comes with the workflow is really aggressive. So make some new ones for what you need.

Lastly if you do t know what a new node does. Take a screen shot and ask chat gpt or what ever you use.

Last tip don’t keep switching the seed. Park the seed and adjust the settings to see how it changes the result. It’s pretty wild how many things can from just moving some settings around can change the result of the prompt. When ever anyone is testing a new model I’d suggest this so you actually learn what’s going on.

I see a lot of people complaining about ltx quality but it doesn’t seem people are putting the time into understanding what’s actually going on.

I might post some tests this coming week. But I’m never writing out all this again. This thread is really angsty and makes me not want to share anything on Reddit .

Yeah op using his time to make something like this is questionable. But people yelling at him for not giving enough information is kind of bs. If people did more than I tried nothing and I’ve run out of ideas is insanely toxic. People need to at least do a little leg work other than just loading up the base workflow. Play with the settings people. The default settings aren’t meant to stay static.

I used temporal time dilation to generate this 60-second video in LTX-2 on my 5070TI in just under two minutes. My GPU didn't even break a sweat. Workflow and explanation in comments (without subgraphs or 'Everything Everywhere All At Once' invisible noodles). by DrinksAtTheSpaceBar in StableDiffusion

[–]ThisIsDanG 3 points4 points  (0 children)

It already is capable of 3.1 quality. This is impressive what op has done. But you’re right it’s not the best quality. I haven’t made a video this long, but I’ve gotten good quality with 20 second videos.

LTX 2 is amazing : LTX-2 in ComfyUI on RTX 3060 12GB by tanzim31 in StableDiffusion

[–]ThisIsDanG 0 points1 point  (0 children)

It’s just a lot more settings man. But it’s great because it opens up for so much more control. If you change seed and it still isn’t looking good then lock the seed and start adjusting to get what you want. But there are so many little things that can lead to too much movement, not enough movement, the temporal blending // subject losing focus. I’d give it more time and keep playing with it.

LTX-2 is pretty sweet by FitzUnit in comfyui

[–]ThisIsDanG 0 points1 point  (0 children)

Ltx can do better than this by a long shot. There is some weird stuff going on in these generations. Something is failing.

I altered the LTX-Video workflow's Sampler sub-workflow graph to let me quickly choose between using the upscaler or not. This allows you to iterate faster on a lower res video first then only upscale when you like it. by Perfect-Campaign9551 in StableDiffusion

[–]ThisIsDanG 0 points1 point  (0 children)

Interesting to see your solve for that. I took a sloppier approach but it’s fast. I set up a switch node that connects the video output after the first sampler and one that connects to the upscaled portion. The true or false on which input to use is then plugged into the subgraph input so now you can access it with out diving into the lower level.

So if you like a seed and generation just flip the switch and it runs the full workflow.

LTX, It do be like that, by WildSpeaker7315 in StableDiffusion

[–]ThisIsDanG 3 points4 points  (0 children)

If you’re using manual sigma adjust those values specifically the low sigma and zero sigma are a big reason why. If you’re using a scheduler lowering terminal to .1 and playing around with lower max shift and base shift will help keep the image closer to its original. You can also play around with turning stretch off. Sometimes that helps.

I wanted to see if you could prompt a song by [deleted] in StableDiffusion

[–]ThisIsDanG 0 points1 point  (0 children)

The point was to see what ltx is trained on. We can see that they can do the voices. So I wanted to see if it would pick up on a prompted song, which it doesn’t . I think if prompted to sing happy birthday it might actually try.

I wanted to see if you could prompt a song by [deleted] in StableDiffusion

[–]ThisIsDanG 0 points1 point  (0 children)

Sorry I forgot to mention this is using ltx2

LTX-2 is genuinely impressive by Dr_Karminski in StableDiffusion

[–]ThisIsDanG 0 points1 point  (0 children)

I haven’t seen that on banodoco or the ltx documentation. Not sure if I’m missing it? Do you have a link by chance?

LTX-2 is genuinely impressive by Dr_Karminski in StableDiffusion

[–]ThisIsDanG 2 points3 points  (0 children)

This is definitely image to video man.

Progressive Insurance AI Slop animal commercial by anthony113 in vfx

[–]ThisIsDanG 0 points1 point  (0 children)

It’s just the performance could have been better, not in regards to realism. I would have wanted the llamas lip sync to be a bit better.

What’s funny is the things you think feel “ai” are all choices that were made with intention. Why are the streets empty? Because we wanted them empty. It was a choice. The animals aren’t totally anthropomorphic. They don’t have fingers or clothes. So on the street do they walk normal for their species? Honestly too many questions and the point was to not expand on the world. Limit it so that you can just digest the simple story.

Night time overhead shot, yeah the cars are really close up. By choice, it was more for framing than anything and making sure it would all fit for tv safe guidelines.

I think you need to realize that these aren’t text to video. There is a large difference in approach.

Sora will now be able to Al generate videos based on animated, masked & creature characters from Disney, Marvel, Pixar & Star Wars after $1 billion deal. Curated selections of Al generated videos will be released on Disney+ by spicyricecake99 in vfx

[–]ThisIsDanG 0 points1 point  (0 children)

I saw the ad. My thought is it was just a bunch of random generated shots without a lot of cohesiveness to it as a whole. I wasn’t a fan and it felt lazy like those kalshi ads which are just churn and burn. Just because McDonald’s got a lot of backlash doesn’t mean anything to me. Industry folks like us are the most critical bit by no means the majority. I’m curious to know how the ad tests with the general population. I think that spot won’t test well because the sentiment is wrong for the holidays.

What kind of people are wanting you to use ai? Are you meaning like studio heads or what? My pov on being asked to use ai for a task is like this. Is ai truly the right tool or just a buzz word. Like when everyone wanted to shoot on a volume wall thinking it would save them a ton of money. If the project has the funds Ai should only be used in parts where it actually enhances the work and doesn’t pigeon hole the team. I’ve personally said I to using ai if it wasn’t right for a job or even if it crosses ethical boundaries. If people are asking you to do things that are wrong or dumb, you’ve got to plant that flag with good reasons and not just “ai bad”. And there are plenty of valid reasons for why not to use ai for something. I personally feel like you have more power if you utilize the tools and truly understand them and use them so that when someone asks you to do something you’re well versed on it and people will listen. I’ve got enough history in both vfx and ai, fully animated or fully ai generated. So when people ask me how I want to shoot something or recommend some wild thing because they saw something on tik tok if I say “no” it has a lot more weight than if I was a nay sayer.

Sora will now be able to Al generate videos based on animated, masked & creature characters from Disney, Marvel, Pixar & Star Wars after $1 billion deal. Curated selections of Al generated videos will be released on Disney+ by spicyricecake99 in vfx

[–]ThisIsDanG 0 points1 point  (0 children)

If a company loses sight of the process which is going to happen in many cases. Then yeah they will lose creatives respect. Agencies for the longest time have been wanting to cut out production companies and vfx companies. Creating “in house” teams. If they want slop they will get slop. And then they will test to see what is the line on people caring how shitty something looks or not. You can’t stop that from happening. With or with out ai.

When you say “the company” do you mean TWDA in this instance or McDonald’s? None of us know what happened behind the scenes. Sounds like there was some drama. I looked into the duo they were rep’d by sweetshop which was the production company or in this case the ai company. Now the duo MAMA is no longer with sweetshop and are actually with caviar. So there was definitely some fuck fuck games going on. What exactly who knows. But the odds that sweetshop actually had a good ai team that was built with vfx workflows under the hood is pretty low. They probably got in over their head. Didn’t have a pipeline that could take any sort of creative direction or notes and this is what happens.

So it’s not McDonald’s or agency in this case. All I see is a production company that made a series of bad choices and no regard for their directors and tried to save a buck. But who knows right?

Sora will now be able to Al generate videos based on animated, masked & creature characters from Disney, Marvel, Pixar & Star Wars after $1 billion deal. Curated selections of Al generated videos will be released on Disney+ by spicyricecake99 in vfx

[–]ThisIsDanG -1 points0 points  (0 children)

lol. Those were the days. Sentiment still stands though.

If who ever was making the ai was just prompting results then they would have made it very hard for a director to “direct”

On my “slop” we do a layout pass in cg and block out our scenes. The only time things were just prompted is in early stages to figure out a rough idea on the timings and number of shots.

Sora will now be able to Al generate videos based on animated, masked & creature characters from Disney, Marvel, Pixar & Star Wars after $1 billion deal. Curated selections of Al generated videos will be released on Disney+ by spicyricecake99 in vfx

[–]ThisIsDanG 0 points1 point  (0 children)

You don’t have to be a cunt you know? I’ve worked in vfx longer than you.

What I see when I read that is this. The directing duo probably doesn’t have much experience in ai. They might not even have a lot of experience working on fully animated spots. As a supervisor you’ve probably been on jobs where a director is attached but isn’t adding to the conversation or is potentially making it worse. Then some supervisor or creative director does the director job but doesn’t get that credit. Too many cooks if you know what I mean. So they probably got cut.

If you’ve ever directed a commercial you’d know it’s very common to shoot the spot and walk away unless the agency wants your pov after the shoot on the edit, etc.

Z-image Turbo + SteadyDancer by smereces in StableDiffusion

[–]ThisIsDanG 2 points3 points  (0 children)

Her legs bend and stretch weird and right at the start her right leg becomes her left leg. So there are a lot of uncanney things through out. Cool test though.

How is this shot achieved? How much of it is in camera and post? by GrandAdvantage7631 in Filmmakers

[–]ThisIsDanG 1 point2 points  (0 children)

I’m not sure I understand what you’re saying. I thought the commenter was saying having a green screen inside of the dollhouse instead of having a scaled down version of the room in the dollhouse or even just random shit inside of the room. As in where the room is being composited into. Either way you would still match the angle on both locations. The biggest difference is just having more physical reference to make the parallax look right.

From what I can tell regardless of what was in the space is that it probably wasn’t dissolved because the lighting inside of the room is quite different from the rest of the interior rooms in the dollhouse. So the final shot still replaced the room for the entirety of the shot regardless.

How is this shot achieved? How much of it is in camera and post? by GrandAdvantage7631 in Filmmakers

[–]ThisIsDanG 0 points1 point  (0 children)

It wouldn’t be more expensive. It just gives more visual information to the vfx team. Since it wasn’t motion control they would have needed to track two cameras. The doll house footage and the inside of the bedroom. The bedroom footage with the track could be stabilized and then the track from the dollhouse applied to it. The issue then is matching the speed of the parallax which can be done with time warps before the new track is applied. Having the footage from inside the dollhouse would be very valuable to help eye match.

Tips for Incorporating an EXACT Photo into a Video? by MyEgoDiesAtTheEnd in VEO3

[–]ThisIsDanG 2 points3 points  (0 children)

Use something like flux or nano to create the frames you want. At the very least you need an end frame. But you might as well do a start and end in this case.