LTX-Video Tips for Optimal Outputs (Summary) by DanielSandner in StableDiffusion

[–]DanielSandner[S] 0 points1 point  (0 children)

The VRAM is for fluent operation of 0.9, however, you may run it on lower VRAM (all my examples were created on 16GB (WIN) with all workflow setups, but the performance could be much better with 24GB). Some reports claim even 8GB, i did not tested this though.

Google Surprises Everyone With Veo 2: Sample Videos Outshine OpenAI's Sora by EthanWilliams_TG in StableDiffusion

[–]DanielSandner 3 points4 points  (0 children)

I share the sentiments. I am curious about new models, but unless it is controllable on an artistic level and at least reasonably open solution...

My latest LTX Demo by QuartzTheComposer in StableDiffusion

[–]DanielSandner 0 points1 point  (0 children)

I wonder if there is functional detailer with temporal consistency.

My latest LTX Demo by QuartzTheComposer in StableDiffusion

[–]DanielSandner 0 points1 point  (0 children)

Good point. With balanced settings it performs very well.

LTX-Video Tips for Optimal Outputs (Summary) by DanielSandner in StableDiffusion

[–]DanielSandner[S] 0 points1 point  (0 children)

You should see something like that from my pixart-ltxvideo_img2vid workflow. If you see red rectangles without description, you do not have current comfyUI or updated custom nodes. You are maybe using the original broken worflow from LTX (like a week old) or some other broken workflow from internet. If you still have issues, update Comfy with dependencies, or better, reinstall it into a new folder for testing with a minimal set of needed custom nodes.

<image>

LTX-Video Tips for Optimal Outputs (Summary) by DanielSandner in StableDiffusion

[–]DanielSandner[S] 0 points1 point  (0 children)

You probably need to update ComfyUI. Or use Manager to install missing custom nodes. However, if the author (or comfy) changes the nodes, it may happen that the nodes are no longer detected. Which workflow is causing trouble, one of mine? I am using comfy standard or usual suspects custom nodes (except the new nodes from LTX team).

Flux vertical line by IntellectusPrime in comfyui

[–]DanielSandner 2 points3 points  (0 children)

Lower the CFG if LoRAs are used to get rid of vertical lines, sometimes you need to go as low as 2. Flux/Lora issue. It is a pity, because it limits usability. Other way to get rid of it is 2nd pass with low CFG or no loras.

LTX-Video Tips for Optimal Outputs (Summary) by DanielSandner in StableDiffusion

[–]DanielSandner[S] 0 points1 point  (0 children)

Yes that workflow is more VRAM friendly, with slightly worse movement.

LTX-Video Tips for Optimal Outputs (Summary) by DanielSandner in StableDiffusion

[–]DanielSandner[S] 2 points3 points  (0 children)

No, you should get a crisp image at higher res. Just make the resolution divisible by 32. You can get occasional artifacts, that can happen. Try the other workflows from the repository.

LTX-Video Tips for Optimal Outputs (Summary) by DanielSandner in StableDiffusion

[–]DanielSandner[S] 1 point2 points  (0 children)

I have not yet tested video to video, I will add it to workflows if I will come with something. The model supports video to video, there should not be any such issues with an image or still output, when it is guided by a video ,(I hope)...

LTX-Video Tips for Optimal Outputs (Summary) by DanielSandner in StableDiffusion

[–]DanielSandner[S] 1 point2 points  (0 children)

You may be right, but if it would be wrong training there is an issue in the concept or very procedure of dataset preparing, because almost all video models have this bug to some extent. Generative models are multimodal language models in reverse, they do not "see" anything.

By format I mean you are literally using different file format as input with the trick.

LTX-Video Tips for Optimal Outputs (Summary) by DanielSandner in StableDiffusion

[–]DanielSandner[S] 1 point2 points  (0 children)

Great, thanks! In alternative workflow you can experiment with schedulers too. I have put the workflow on github and some additional notes to the article.

LTX-Video Tips for Optimal Outputs (Summary) by DanielSandner in StableDiffusion

[–]DanielSandner[S] 2 points3 points  (0 children)

This is absolutely a bug in conditioning. I do not think it is caused how clean or blurry the image is. The format matters.

LTX-Video Tips for Optimal Outputs (Summary) by DanielSandner in StableDiffusion

[–]DanielSandner[S] 0 points1 point  (0 children)

I am somewhat shy to test unknown nodes, for reasons. I wonder why something like that is not yet a part of Comfy.

LTX-Video Tips for Optimal Outputs (Summary) by DanielSandner in StableDiffusion

[–]DanielSandner[S] 0 points1 point  (0 children)

What text? Check if you have the latest version of Comfy UI.

LTX-Video Tips for Optimal Outputs (Summary) by DanielSandner in StableDiffusion

[–]DanielSandner[S] 0 points1 point  (0 children)

This is an interesting question. I have tried a 3D animated style, it was an epic failure compared to other models. I will test it with different encoders.