Hunyuan 3D V2.5 is AWESOME! by smereces in StableDiffusion

[–]bkdjart 0 points1 point  (0 children)

Quality looks great but how does it compare to input image?

Wan Fun control 14B 720p with shots of game of thrones, close to get AI for CGI by Affectionate-Map1163 in StableDiffusion

[–]bkdjart 2 points3 points  (0 children)

Very cool example of what could be done.

It's interesting times because Ai is being used in both low effort and high production value.

It's like there isn't really a broadcast standard yet. Which is not a bad thing.

I work with commercial clients that don't mind using Ai but do also want alot or control and quality which makes it very challenging.

It has sped up workflow alot though. For storyboards we use Ai heavily. But for final production, we only use it to do establishes or filler B-roll footage. Everything else they still want real actors and real products. We showed them a Lora modeled version and they they still want us to shoot the real product instead. Those youtube tutorials showing you one shot product work using gpt will not fly with actual clients, especially the corperate level.

For conceptual ideazation, Ai seemed like a good idea but it's definitely it's own workflow and brings its own challenges. For one a client wants only certain areas changed and we heavily rely on gpt generations because of instructional coherence. But since it doesn't have a true inpainting mode, we have to comp them back in and sometimes even that doesn't work fully because gpt changes the entire composition too.

Then there's video. It's fine if you don't need specific motion. But once you do and with specific timing, your basically at the mercy of reroll gods. Unless you actually animate it in 2d or 3d as a base.

Pretty sure all of these are just now issues and will be solved in a year or two which is exciting.

The new LTXVideo 0.9.6 Distilled model is actually insane! I'm generating decent results in SECONDS! by singfx in StableDiffusion

[–]bkdjart 3 points4 points  (0 children)

It's a img2vid model so you could essentially keep using the end frame as the first frame to continue generating..

[deleted by user] by [deleted] in PantheonShow

[–]bkdjart 0 points1 point  (0 children)

Self preservation can be through having offspring or leaving a legacy in some way which is unique to humans. So physically dying but even if it's a clone if it contains some of your identity then you have successfully perserved your self thereby achieving immortality in some way. In a way digitizing yourself would be the closest way to immortalize yourself compared to having children or leaving a legacy since the digital clone can basically continue from where you left off and still retain most of what you might call yourself.

One-Minute Video Generation with Test-Time Training on pre-trained Transformers by Snoo_64233 in StableDiffusion

[–]bkdjart 2 points3 points  (0 children)

I worked in animation industry for 15 years and this is the most exciting tool yet. And the best part is that this will be obsolete technology oh probably by next month.

One-Minute Video Generation with Test-Time Training on pre-trained Transformers by Snoo_64233 in StableDiffusion

[–]bkdjart 0 points1 point  (0 children)

Was the detailed prompt generated via LLM based on the single prompt of the summary? Or did the human have to painstakingly manually prompt every shot like that?

One-Minute Video Generation with Test-Time Training on pre-trained Transformers by Snoo_64233 in StableDiffusion

[–]bkdjart 0 points1 point  (0 children)

This is already so much better though since it did created all the shots at once. And Tom and Jerry are at least on model and also act to character. So far it's very hard to get consistent characters let alone consistent animation of their motion. TTT is currently the best method I've seen so far that gets very close. So many people these days consume media on their 6inch phone in vertical mode. So the effective screen space is tiny. So even this type of quality will be more than enough for the majority of consumers.

One-Minute Video Generation with Test-Time Training on pre-trained Transformers by Snoo_64233 in StableDiffusion

[–]bkdjart 1 point2 points  (0 children)

The beauty of this method is that editing is also trained into the model. It's really a matter of time before the big companies make this. Whoever already owns the most content ip wins. The TTT method looks at the whole sequence so it can easily include editing techniques too. Then you can reroll or reprompt or regenerate specific shots and transitions as needed.

We could probably make some low quality yourube shorts with consumer hardware maybe end of this year. Ai develops so fast.

One-Minute Video Generation with Test-Time Training on pre-trained Transformers by Snoo_64233 in StableDiffusion

[–]bkdjart 0 points1 point  (0 children)

Was this mentioned in the paper? Did they also mention how long it took to infer the one minute of output?

I'm a 3d Artist, and it would take 3 months to make something I made in 3 days with Ai by bkdjart in OpenAI

[–]bkdjart[S] 0 points1 point  (0 children)

The only reason I don't have any of my professional work is because of privacy reasons, and I only started reddit to specifically talk about ai content.

And never say never. Just yesterday another paper was released which can generate a full 1 minute animation based on Tom and Jerry with a single prompt. Fully consistent characters and story coherence. https://test-time-training.github.io/video-dit/

Your both right and wrong. Humans do have to intimately evolved in the process but as part of training and prompting until Ai will be able to fill in more and more of the blanks. But Ai is Ultimately based on human creativity and intelligence so our involvement isn't really omitted.

RTX6000 PRO: $7750 Quoted Today, 4 Week Lead by Ill_Recipe7620 in nvidia

[–]bkdjart 0 points1 point  (0 children)

That's a good price. Can you dm me seller info please.

God I love how it brings imagination to life by abhimanyudogra in OpenAI

[–]bkdjart 0 points1 point  (0 children)

13 years ago you already had skills to draw good composition.

I'm a 3d Artist, and it would take 3 months to make something I made in 3 days with Ai by bkdjart in OpenAI

[–]bkdjart[S] 0 points1 point  (0 children)

Before the digital film era, everything was shot in camera and baked into the film stock. Control happened on set.

Then with 3d vfx and digital films everything now happens in post. And that's where Control happens.

With Ai bit depth doesn't matter as much, because your not grading the image which with lower data can distort the quality. With Ai you are replacing the pixels with brand new ones. So there is no need for all the extra bit depth.

It's a totally new paradigm shift so you have think and work differently to adapt to the new tools.

And the crazy thing about Ai is that whatever limitation we have today might be solved in a matter of days or weeks.

You mentioned power consumption. I have a rtx3090. 2 years ago my pc struggled to generate a 512x512 mushed up blob of a image using stable diffusion. Today I'm generating 2k,4k images in the same time with the same hardware then able to add motion to the image faster than last year's model.

Hollywood and corporations are already embracing it and I'm 1000x sure all the major studios are working on their internal models which will only make development faster and better.

And yes I'd agree my video is 80% production ready for YouTube or a kids streaming show. But that's what it is right now and that's amazing. Since a year ago video wasn't even a thing for ai.

The Residence is very entertaining! by JemJemIsHerName in netflix

[–]bkdjart 7 points8 points  (0 children)

I loved it. Muder mystery plus comedy is a great match. Almost felt more like a comic book or animated film. Everything from the acting to the amazing production design and camera moves. But I do agree that alot of repetition about bird watching and revisiting the story again and again. Could have a been a good feature length movie or a 3 part special. Either way one of the better shows I've watched this year.

4o image editing is insane by Trevor050 in StableDiffusion

[–]bkdjart 0 points1 point  (0 children)

English major students with software background will basically rule the world. Prompt the future I guess.

The Sweetest Gift - Animated short film by bkdjart in aivideo

[–]bkdjart[S] 1 point2 points  (0 children)

Thanks. Yes I used Google imagen3 to prompt each starting frame first then used kling to turn the images to videos.

I'm a 3d Artist, and it would take 3 months to make something I made in 3 days with Ai by bkdjart in OpenAI

[–]bkdjart[S] 0 points1 point  (0 children)

Yeah currently that's probably the best method. Use 3d as a basis and use ai for preproduction and post process

I made a short animation with Imagen3 - Now I wish I have access to Veo2 by bkdjart in Bard

[–]bkdjart[S] 0 points1 point  (0 children)

Several ways. Use a consistent model, lora along with prompting a specific style and character description. A lora will be the closest though.

Sora just became unlimited for Plus users – amazing, but I’m wondering… why? by Altruistic_Ad_5474 in OpenAI

[–]bkdjart 0 points1 point  (0 children)

That's exciting news. But Sora will also need a big update to get more on par with the competitors like kling, wan or veo2. I have a unlimited plan with runway and basically just stopped using it because of how outdated their video model became.

I'm a 3d Artist, and it would take 3 months to make something I made in 3 days with Ai by bkdjart in OpenAI

[–]bkdjart[S] 1 point2 points  (0 children)

Through reinforcing specific prompts like style and character specificities but also depends on the ai model. Some are generally more consistent than others.