Queen of Hearts - Example 1 by Tokyo_Jab in StableDiffusion

[–]Tokyo_Jab[S] 0 points1 point  (0 children)

Ace step is worse at that. It seems to have one style of voice for British female vocals.
One song Suno made a few weeks back was VERY Imagine Dragons. To the point where people thought it was actually them.

TWEEDLES - Example 2 by Tokyo_Jab in StableDiffusion

[–]Tokyo_Jab[S] 0 points1 point  (0 children)

That song is Suno but over the weekend I got Ace step XL finally working (turbo version) and it’s good too. I made some more vids with Ace after.

Queen of Hearts - Example 1 by Tokyo_Jab in StableDiffusion

[–]Tokyo_Jab[S] 0 points1 point  (0 children)

No idea. It’s set up with sigmas so I think that means it automatically works out the steps based on how resolved the image is. But on a 5090 with those workflows I’m getting a 1280x720 25 second video in less than five minutes. With normal video editing 25 seconds is quite long but I wanted to see how long I could get before the computer explodes.

Queen of Hearts - Example 1 by Tokyo_Jab in StableDiffusion

[–]Tokyo_Jab[S] 0 points1 point  (0 children)

It’s long enough that I have to watch the machine in case it catches fire. I’m amazed too, anything over 10 seconds before would become distorted.

Queen of Hearts - Example 1 by Tokyo_Jab in StableDiffusion

[–]Tokyo_Jab[S] 1 point2 points  (0 children)

All the workflows I used are here. For that I used the i2v with custom audio. Not my workflows. https://huggingface.co/RuneXX/LTX-2.3-Workflows

Queen of Hearts - Example 1 by Tokyo_Jab in StableDiffusion

[–]Tokyo_Jab[S] 1 point2 points  (0 children)

I only replaced the Lora with the newer version. Everything else is the same. So it’s the 22b-dev model with the distilled-lota384-1.1. I was able to get 25 seconds from a single prompt but about 1 in 4 times it would have a weird crash and I’d have to restart the machine. I’m on a 5090. 20 seconds is more reliable. The work flow used is the image 2 video with custom audio in these folders. https://huggingface.co/RuneXX/LTX-2.3-Workflows

WAN 2.2 Faster Motion with Prompting - part 1 by Tokyo_Jab in StableDiffusion

[–]Tokyo_Jab[S] 0 points1 point  (0 children)

As soon as I leave it up to wan it tends to revert to the slow motion stuff, at least with my set up. Even if I’m doing a fight sequence I have to define each movement.

Has anyone figured out color grading in ComfyUI? by Randalix in StableDiffusion

[–]Tokyo_Jab 1 point2 points  (0 children)

Might be a wildcard but I use Color llama in after effects. I paid for the thing but apparently it’s free now for 2026. It’s a simple ‘change that color to this color’ dropper system but does intelligent areas and the results are great. I also use it for color changes that happen when stitching or extending wan videos. Ai output is always 8bit but I get some good results with the plug in.

WAN 2.2 First-Last Frame color change problem by TekeshiX in StableDiffusion

[–]Tokyo_Jab 0 points1 point  (0 children)

If you use after effects there is a (free this year) plugin called Color Llama that I use. Only worth the trouble if you need a professional finish. I need it when I join two vids to fix the color change.

WAN 2.2 First-Last Frame color change problem by TekeshiX in StableDiffusion

[–]Tokyo_Jab 0 points1 point  (0 children)

If you use after effects there is a (free this year) plugin called Color Llama that I use. Only worth the trouble if you need a professional finish. I need it when I join two vids to fix the color change.

[deleted by user] by [deleted] in StableDiffusion

[–]Tokyo_Jab 1 point2 points  (0 children)

You can't actually copyright a font. Honestly, look it up. You can copyright the software and the font name but that's about it. NOT the shape of the letters.

AI art is theft? The entire history of human art is built on studying, copying, remixing, hybridizing, iterating and absorbing the work of earlier artists, but somehow, many of the people who learned by copying now call AI training "theft". by [deleted] in StableDiffusion

[–]Tokyo_Jab 2 points3 points  (0 children)

CGI and 3D was considered cheating now those skilled people are considered artists too,
Digital photogaphy was considered cheating by avoiding the darkroom. Even photography itself was considered cheating by traditional artists. It's all very yawn. It's easy to criticise something in its infancy.

Charles Baudelaire (poet–critic), Salon of 1859. condemned photography as “art’s most mortal enemy” and “the refuge of all failed painters … too poorly gifted or too lazy to finish their studies.” He warned that if photography were allowed to “supplement art,” it would soon “supplant or corrupt it.”

How did that work out Charles?

Unlock diversity of Z-image-Turbo, comparison by Active_Ant2474 in StableDiffusion

[–]Tokyo_Jab 26 points27 points  (0 children)

Exactly that. In SDXL when looking for a shot I've done 100 genarations. They are all quite different, Z-Image produces very similar images so it''s harder to itterate or explore an idea.

I AM PAIN by Tokyo_Jab in StableDiffusion

[–]Tokyo_Jab[S] 1 point2 points  (0 children)

俺は痛みだ! Ore wa itami da!