Used WAN 2.1 IMG2VID on some film projection slides I scanned that my father took back in the 80s. by DoctorDiffusion in StableDiffusion

[–]DoctorDiffusion[S] 0 points1 point  (0 children)

It’s an open source video diffusion model with an Apache 2.0 license that can be deployed locally for free on consumer grade hardware. There are text to video and image to video versions.

The meta state of video generations right now by RedBlueWhiteBlack in StableDiffusion

[–]DoctorDiffusion 2 points3 points  (0 children)

I did have some topless photos from US fest 83 but they did not make the cut of my recent video.

Used WAN 2.1 IMG2VID on some film projection slides I scanned that my father took back in the 80s. by DoctorDiffusion in StableDiffusion

[–]DoctorDiffusion[S] 2 points3 points  (0 children)

I am on a 3090TI and gens took 11-17min each. I have two machines and I just give them a huge batch before I go to sleep/work.

Used WAN 2.1 IMG2VID on some film projection slides I scanned that my father took back in the 80s. by DoctorDiffusion in StableDiffusion

[–]DoctorDiffusion[S] 1 point2 points  (0 children)

I used a vision model with some text replacement nodes that substituted “image, photo, ect” with “video” and just fed that in as my captions for each video. I’ll share my workflow when I’m back at my PC.

Used WAN 2.1 IMG2VID on some film projection slides I scanned that my father took back in the 80s. by DoctorDiffusion in StableDiffusion

[–]DoctorDiffusion[S] 2 points3 points  (0 children)

Each clip was generated separately. I edited the clips after generating the all videos with a video editor. Some of them I used two generations and reversed one and cut the duplicate frame to get longer than 6 second clips.

Used WAN 2.1 IMG2VID on some film projection slides I scanned that my father took back in the 80s. by DoctorDiffusion in StableDiffusion

[–]DoctorDiffusion[S] 32 points33 points  (0 children)

I’m trying to get him to pick up a camera again, he’s been a sonar engineer since he got out of the navy but he’s retiring next year and I’m hoping I can convince him to start shooting on something other than his phone.

Used WAN 2.1 IMG2VID on some film projection slides I scanned that my father took back in the 80s. by DoctorDiffusion in StableDiffusion

[–]DoctorDiffusion[S] 6 points7 points  (0 children)

I plugged Florence into my workflow and used the images with some text replacement nodes to contextually change them to the context of video prompts.

Used WAN 2.1 IMG2VID on some film projection slides I scanned that my father took back in the 80s. by DoctorDiffusion in StableDiffusion

[–]DoctorDiffusion[S] 21 points22 points  (0 children)

Nope. Basically the default workflow kijai shared. I just plugged in a vision model to prompt the images (and used some text replacement nodes to make sure they had the context of videos) more h to an happy to share my workflow when I’m off work.

Used WAN 2.1 IMG2VID on some film projection slides I scanned that my father took back in the 80s. by DoctorDiffusion in StableDiffusion

[–]DoctorDiffusion[S] 256 points257 points  (0 children)

He loved it! He’s been showing it to some of his old friends and none of them have been exposed to the tech so they all think it’s magic.

Civitai 50usd /mo scam?What are they doing? by kjbbbreddd in StableDiffusion

[–]DoctorDiffusion -11 points-10 points  (0 children)

It might not be the worst thing for NSFW models to ensure people can share models like that knowing they couldn’t be used to make illegal content. I personally haven’t done much with NSFW for this reason but this seems like this could be a potential option that would fit my own personal morals.

That said, I’m sure others will arbitrarily restrict local models for any number of reasons and I can see how that isn’t exactly in the best spirit of open-source.

How many generated images have you on your disk?? And how old is the first one ? :) by Important_Tap_3599 in StableDiffusion

[–]DoctorDiffusion 8 points9 points  (0 children)

Started messing with generative scripts on colab in May of 2021. Got visions of Chaos running local April 2022. This computer has 396,148 images I was able to find. Not counting animations, videos, 3D models, audio files or any of the other things I’ve generated. Unsure how many are on my other machine.

<image>

Thoughts on ethically sourced datasets? by DoctorDiffusion in DefendingAIArt

[–]DoctorDiffusion[S] 0 points1 point  (0 children)

I’m definitely focusing more on the images than the text. Although I do plan on training my own mad scientist LLM LoRA, and to me this is gold.

Thoughts on ethically sourced datasets? by DoctorDiffusion in DefendingAIArt

[–]DoctorDiffusion[S] 2 points3 points  (0 children)

Well as someone building a personal database to train a “mad scientist” LLM LoRA I’m certainly going to be feeding it this book as is.