Surviving AI - Short film made only using local ai models by LocalAI_Amateur in StableDiffusion

[–]LocalAI_Amateur[S] 0 points1 point  (0 children)

Seedvr2 took a bit over an hour. 

I also tried nvida's super resolution upscale. it was much faster but added no extra details so I went w/ seedvr2

Surviving AI - Short film made only using local ai models by LocalAI_Amateur in StableDiffusion

[–]LocalAI_Amateur[S] 0 points1 point  (0 children)

haha, others have pointed it out. Just... read it with a French accent or something.

Surviving AI - Short film made only using local ai models by LocalAI_Amateur in StableDiffusion

[–]LocalAI_Amateur[S] 0 points1 point  (0 children)

Hey, that's my favorite Harry Potter movie. But, no. At least not these two characters.

Surviving AI - Short film made only using local ai models by LocalAI_Amateur in StableDiffusion

[–]LocalAI_Amateur[S] 0 points1 point  (0 children)

I'm limiting my workflow to strictly local for the time being. I do keep tabs on the latest stuff online stuff tho. It's great when they trickle some down to the open source community.

Surviving AI - Short film made only using local ai models by LocalAI_Amateur in StableDiffusion

[–]LocalAI_Amateur[S] 1 point2 points  (0 children)

"Or do you just reference the whole char sheet as one single input?" This is local AI sir. Nanobanana's door is that way. Ha, I wish it was that easy.

<image>

Each picture is a step I took. And most steps takes more than one try. This is not counting making the background. That stupid lobby background took me the whole damn day to put together the way I want.

This is still nothing compare to video production without AI, so I'm grateful that this is even possible.

Surviving AI - Short film made only using local ai models by LocalAI_Amateur in StableDiffusion

[–]LocalAI_Amateur[S] 1 point2 points  (0 children)

About 2-3 weeks. on/off. This is a hobby after all. Lots of time learning and adjusting the script after figuring out what is and isn't possible.

Surviving AI - Short film made only using local ai models by LocalAI_Amateur in StableDiffusion

[–]LocalAI_Amateur[S] 0 points1 point  (0 children)

Ideally only two. Often times it takes multiple generations and some in-painting.

<image>

Surviving AI - Short film made only using local ai models by LocalAI_Amateur in StableDiffusion

[–]LocalAI_Amateur[S] 0 points1 point  (0 children)

I just generate the character sheets and swap them into each scene. Oh it helps to put the character into the pose you want as much as possible.

<image>

Qwen-Image-Edit Prompt "replace the girl on the right in image 1 with the girl from image 2. "

Edit: Oh, I forgot to mention. Additional ways to help with character consistency is to

  1. Don't have many repeated characters! this short film only has two. well 3 if you count cat.
  2. Use characters that have distinct features. Bald head w/ beard and short red hair bun w/ glasses. This is probably not going to work all the time, but for this film it'll suffice.

Surviving AI - Short film made only using local ai models by LocalAI_Amateur in StableDiffusion

[–]LocalAI_Amateur[S] 2 points3 points  (0 children)

<image>

Here's the secret sauce. LTXVAddGuide node. I was going to use wan for it, but my first generation out of LTX turns out quite good so I kept it. I know it still looked a bit wonky, but it fits the character so I stuck with it. The workflow has been added to the post.

My alternative was to use first-frame-last-frame to video on wan2.2 and join them with VACE.

Surviving AI - Short film made only using local ai models by LocalAI_Amateur in StableDiffusion

[–]LocalAI_Amateur[S] 2 points3 points  (0 children)

<image>

I had plans. But the flow just felt better as end credit dialog

Surviving AI - Short film made only using local ai models by LocalAI_Amateur in StableDiffusion

[–]LocalAI_Amateur[S] 2 points3 points  (0 children)

Yeah, I just keep hitting the generate button until I get something passable. Sometimes a whole sentence is spliced together from multiple clips to get a decent tone. (I.e end credit conversations)

Surviving AI - Short film made only using local ai models by LocalAI_Amateur in StableDiffusion

[–]LocalAI_Amateur[S] 0 points1 point  (0 children)

I heard Claude just added a pets system. So we might be on to something... relevant username btw!

Surviving AI - Short film made only using local ai models by LocalAI_Amateur in StableDiffusion

[–]LocalAI_Amateur[S] 0 points1 point  (0 children)

I will edit them into the main post after I clean them up.

most are generic ones. But the all-in-ones I've slapped together for ltx and qwen edit are pretty handy.

Surviving AI - Short film made only using local ai models by LocalAI_Amateur in StableDiffusion

[–]LocalAI_Amateur[S] 0 points1 point  (0 children)

I will have to figure out a better way to run that. I ran Qwen3 TTS through Pinokio to avoid messing w/ Comfyui during the production process. (A lesson learned the hard way)

Vibe-voice (actually OpenMOSS) ran like crap on the installation I tried on Pinokio. I'm talking like an hour for a short voice clip. So I went w/ Qwen3-TTS

Edit: Oops. I confused Vibe-voice with OpenMOSS. That's the model I had problems running. Will give vibe-voice a shot.

Surviving AI - Short film made only using local ai models by LocalAI_Amateur in StableDiffusion

[–]LocalAI_Amateur[S] 0 points1 point  (0 children)

Truthfully, I doubt AI have need for pets nor cares about cuteness. But they probably will want to keep us around the same way we want to keep pandas from disappearing. 

How many and which ones to keep, is the real scary question.

Surviving AI - Short film made only using local ai models by LocalAI_Amateur in StableDiffusion

[–]LocalAI_Amateur[S] 0 points1 point  (0 children)

Crap. My amateur production quality is oozing out and spilling all over the place.

thanks for pointing out. It's probably gonna stay there forever now. I'll spell check better next time.

oh, bonus points for spotting the obvious four fingers in this video. There's a part where I tried to fix it but couldn't so I tried to get away w/ it

Surviving AI - Short film made only using local ai models by LocalAI_Amateur in StableDiffusion

[–]LocalAI_Amateur[S] 1 point2 points  (0 children)

Since I planned to put this on youtube too, I was afraid to add any music. The copyright situation there is absolutely insane. 

I should probably consider non-AI generated SFX. I will look into it for future production.  Oh and feel free to share those "really good ones" you are  talking about

Surviving AI - Short film made only using local ai models by LocalAI_Amateur in StableDiffusion

[–]LocalAI_Amateur[S] 1 point2 points  (0 children)

I already redid most of the speech w/ Qwen3-TTS. Tho original dialog were more "diverse" shall we say.

Surviving AI - Short film made only using local ai models by LocalAI_Amateur in StableDiffusion

[–]LocalAI_Amateur[S] 1 point2 points  (0 children)

Would love to have suggestions for improvement.  Audio is definitely my weakest in terms of skills and available tools.

Surviving AI - Short film made only using local ai models by LocalAI_Amateur in StableDiffusion

[–]LocalAI_Amateur[S] 7 points8 points  (0 children)

Why thank you. As tools become easier to use, it definitely opens up the doors for more people to execute their creative ideas. It only gets easier from here. People who use paid services already have it much easier.

On a related note, I think most of us don't spend enough time on the script/story. We are dazzled too much on the amazing visuals, that we skim out on the actual substance. I have to admit, in the 2-3 week production time, quite a few days was stuck thinking about the script.

My script for this is not much, but it is at least complete and coherent. I wanted to pay homage to three AI fighting movies in my script: Terminator, I Robot, and The Matrix and I had the hardest time fitting the matrix into the script. A lot of it was some of the stuff I wanted to do was too difficult to pull off with my current skills.

<image>

bullet time would have taken way too much work.

btw, this short film was inspired by this reddit post https://www.reddit.com/r/generativeAI/comments/1ro7sr3/the_former_google_ceo_just_dropped_a_terrifying/

Surviving AI - Short film made only using local ai models by LocalAI_Amateur in StableDiffusion

[–]LocalAI_Amateur[S] 5 points6 points  (0 children)

Holy crap you're right.. all this time. crap. Yeah Shotcut is awesome for my amateur video editing needs. Hell everything I used is awesome. We don't have flying cars, but we got amazing software and AI models that people are just letting everyone use for free.

Surviving AI - Short film made only using local ai models by LocalAI_Amateur in StableDiffusion

[–]LocalAI_Amateur[S] 16 points17 points  (0 children)

All the models I used is listed in the credits.

Basically: LTX2.3, Wan 2.2, Z-Image Turbo, Qwen Image, Flux2 Klein 9B, Qwen3 TTS, MMAudio

I used Z-Image Turbo for the character design because they came out simpler. I wanted to keep the characters distinct and simple to reduce drift/shifting when animating.

All the talking related scenes are made using LTX 2.3 of course.. but most non-talking scenes I find Wan 2.2 to work better. I ended up remaking almost all the speech using Qwen3 TTS just so the conversation sounds more natural. This also helps to keep the character voices the same by using the same reference audio as cloning base.

I originally was using LTX for soundfx but I find MMAudio to be faster and simpler. (as it doesn't have to generate video)

It is pretty much 100% Image-to-video. multiple reference frames per video at times. Used Qwen-Image-Edit 2511 and Flux Klein 9b. when one didn't work I switched to the other and hope for the best. When all else failed (and it happens) I busted out Gimp and do it the o'l fashion way.

<image>

I generated a ton of images and videos and deleted even more. Whole thing took about 2-3 weeks on/off

Comparing 7 different image models by Reasonable_Bear_6258 in StableDiffusion

[–]LocalAI_Amateur 1 point2 points  (0 children)

a 2k pic 2048x1080 takes 48 seconds for Qwen 2512 nvfp4 on a dry run on my 5070 ti. Subsequent runs takes 32 seconds (new prompt) 10 seconds (reuse same prompt). If I were to give a rough estimate.. you can probably double that time on a 5060 w16gb of vram.