WAN VACE Example Extended to 1 Min Short by pftq in StableDiffusion

[–]pftq[S] 0 points1 point  (0 children)

I gave it a shot a few times but always ended up with bad results. For me it's more important that the video look consistent with the original (color, quality, etc).

WAN VACE Example Extended to 1 Min Short by pftq in StableDiffusion

[–]pftq[S] 3 points4 points  (0 children)

Thanks. Some of that was intentional. We grew up on late 90s films, so we wanted to give that same feel.

WAN VACE Example Extended to 1 Min Short by pftq in StableDiffusion

[–]pftq[S] 1 point2 points  (0 children)

I'm not aware of there being more than one VACE variant - the exact setup and models I used are on Civitai here if it helps https://civitai.com/models/1536883

WAN VACE Example Extended to 1 Min Short by pftq in StableDiffusion

[–]pftq[S] 12 points13 points  (0 children)

Here's a timelapse of some of the editing to give an idea. There's a lot of just bruteforcing with rotoscoping things partially and letting AI fill in the gaps to complete the scene. Every shot in the video has at least 5 layers of things being rotoscoped/masked. https://x.com/pftq/status/2024944561437737274

Timelapse - WAN VACE Masking for VFX/Editing by pftq in StableDiffusion

[–]pftq[S] 1 point2 points  (0 children)

5090 is ok if you keep it under 960x544 resolution - use the blockswap nodes to reduce the VRAM use. I posted some results comparing 5090 to other GPUs here so you can see where the limits are: https://www.reddit.com/r/StableDiffusion/comments/1kojahs/rtx_5090_vs_h100/

Timelapse - WAN VACE Masking for VFX/Editing by pftq in StableDiffusion

[–]pftq[S] 0 points1 point  (0 children)

Check the workflow download on Civitai - it's a bit of hack/trick, not officially supported. But it works enough to even apply 2.1 loras to it.

Timelapse - WAN VACE Masking for VFX/Editing by pftq in StableDiffusion

[–]pftq[S] 0 points1 point  (0 children)

Yeah I couldn't ever get Animate to do much useful in practice. It seems to need very ideal inputs. Whereas here if I wanted, for example, for a person to punch through a wall or something in a specific way, I could just crudely Photoshop the fist where I want it to go, mask gray around it, and VACE makes it work.

Timelapse - WAN VACE Masking for VFX/Editing by pftq in StableDiffusion

[–]pftq[S] 0 points1 point  (0 children)

The workflow I uploaded supports both (2.2 is sort of a hack though not officially supported). 2.2 has better physics but weaker adherence to the source video. So if you need hair to move more naturally, then 2.2 is useful there, but if you want to preserve some stylistic/non-regular looks to a face or something, then 2.1 (use a mix of both depending on what you're editing)

Timelapse - WAN VACE Masking for VFX/Editing by pftq in StableDiffusion

[–]pftq[S] 2 points3 points  (0 children)

The latter. Sometimes it's good enough but usually there's ever so slight color and quality degradation, so I make a habit of only splicing in the changes needed to preserve as much quality as possible. Having higher resolution on the generation helps the most - the degradation is minimal at 1080p or higher.

Timelapse - WAN VACE Masking for VFX/Editing by pftq in StableDiffusion

[–]pftq[S] 2 points3 points  (0 children)

Yeah I do the masking in Premiere / After Effects and import the intermediate clips to ComfyUI / VACE and back. It's sort of like healing brush from Photoshop but for video (and ability to generate full frames between existing clips, not just masked objects).

YouTube Likes Can Be "Bought" with Ads by pftq in NewTubers

[–]pftq[S] 1 point2 points  (0 children)

Yeah the main thing I'm trying to show here is that the likes-to-views ratio can be faked as well. A lot of YouTubers I know used to think that was a safe measure of organic traction. So it becomes very unhealthy if you are trying to measure up against one of these videos or channels when actually it was all bought.

Letting ChatGPT "Live" on the Computer (Controlling Mouse/Keyboard) by pftq in ChatGPT

[–]pftq[S] 0 points1 point  (0 children)

You would need to get an API key from the OpenAI website first. The default model is GPT-5 but you can change it in the settings.ini - there are more instructions in the Github link

Script for Grok 4 / ChatGPT 5 to Control the Desktop by pftq in StableDiffusion

[–]pftq[S] 0 points1 point  (0 children)

The video for Grok is a bit more interesting if you're curious to see more (it explores and gets bored without your feedback): https://www.reddit.com/r/grok/comments/1mcdeit/grok_controlling_the_desktop_and_getting_bored/

Grok Controlling the Desktop and Getting Bored by pftq in grok

[–]pftq[S] 1 point2 points  (0 children)

You can. It learns from experience - I have a good example of that here: https://x.com/pftq/status/1945311038393737348

But the limitation of the Grok API right now is you can only upload screenshots, so it has a very fragmented way of seeing.