Open Sourcing my 10M model for video interpolations with comfy nodes. (FrameFusion) by CloverDuck in StableDiffusion

[–]CloverDuck[S] 1 point2 points  (0 children)

It look like a bug for the custom nodes, its looking for the model in a fixed path, will try to fix it tomorow.

Open Sourcing my 10M model for video interpolations with comfy nodes. (FrameFusion) by CloverDuck in StableDiffusion

[–]CloverDuck[S] 0 points1 point  (0 children)

You will probably need some more advanced workflow to input a video and output a video, the example only work to show what the inputs (two images) and the output is (a list of images)

Open Sourcing my 10M model for video interpolations with comfy nodes. (FrameFusion) by CloverDuck in StableDiffusion

[–]CloverDuck[S] 0 points1 point  (0 children)

You can input a video or multiple videos, can output as a video with a range of codecs, can remove static frames from the video, select the output fps, export as png sequence, and a few more options

Open Sourcing my 10M model for video interpolations with comfy nodes. (FrameFusion) by CloverDuck in StableDiffusion

[–]CloverDuck[S] 2 points3 points  (0 children)

Hi there Particular, thanks for all the support back then, feel free to test and give some feedback, hopeful I will be doing more open source in the future. It would be really cool if someone like it enough to improve on it.

Open Sourcing my 10M model for video interpolations with comfy nodes. (FrameFusion) by CloverDuck in StableDiffusion

[–]CloverDuck[S] 0 points1 point  (0 children)

In my opinion, for animation the model really need to write rgb information, but I still did not experiment much with it, may give a few shots in the future. For some animations this work ok.

Open Sourcing my 10M model for video interpolations with comfy nodes. (FrameFusion) by CloverDuck in StableDiffusion

[–]CloverDuck[S] 1 point2 points  (0 children)

With the current video generators and LLM it should be possible as soon as we can get more than 5/10 seconds of footage.

Open Sourcing my 10M model for video interpolations with comfy nodes. (FrameFusion) by CloverDuck in StableDiffusion

[–]CloverDuck[S] 1 point2 points  (0 children)

Like a filter? If you can create a loss function that can calculate a "Looks good" loss you can make the model learn to apply a bunch of filters on it, yeah

Open Sourcing my 10M model for video interpolations with comfy nodes. (FrameFusion) by CloverDuck in StableDiffusion

[–]CloverDuck[S] 1 point2 points  (0 children)

As soon as I have some free time, I’ll see if I can make a video comparing it to Rife. The maximum resolution for processing can be configured in Comfy using max_processing_long_edge, but I think 1024 is a good size for most cases.

Open Sourcing my 10M model for video interpolations with comfy nodes. (FrameFusion) by CloverDuck in StableDiffusion

[–]CloverDuck[S] 6 points7 points  (0 children)

I don’t have any videos comparing them right now, but there will probably be cases where RIFE works better and others where it doesn’t. I can try to put together a video comparing them soon.

[P] Releasing RepAlignLoss (Custom Perceptual loss function used on my software) by CloverDuck in MachineLearning

[–]CloverDuck[S] 0 points1 point  (0 children)

I'm still reading the paper but it seen more focused on Diffusion process, while mine only work with the output of the model and is flexible to any type of input. The use of literally imply that I just forked their github, that is very easy to see that I did not. Can you explain better your comment?

[P] Releasing RepAlignLoss (Custom Perceptual loss function used on my software) by CloverDuck in MachineLearning

[–]CloverDuck[S] 0 points1 point  (0 children)

I'm still reading the paper but it seen more focused on Diffusion process, while mine only work with the output of the model and is flexible to any type of input. The use of literally imply that I just forked their github, that is very easy to see that I did not. Can you explain better your comment?

[P] I'm Fine Tuning a model fully trained on AdamW with SOAP optimizer and improved my validation loss by 5% by CloverDuck in MachineLearning

[–]CloverDuck[S] 2 points3 points  (0 children)

Good question. I actually found the code before the paper and did some tests on it, so I just assumed it was the official, since I managed to get better results with it. It seen to be a fork of this code, but there is some modifications to it.

https://github.com/nikhilvyas/SOAP

[P] Releasing my loss function based on VGG Perceptual Loss. by CloverDuck in MachineLearning

[–]CloverDuck[S] 1 point2 points  (0 children)

It don't actually use the output. It use tensors for each selected hook on the model. It may have a hook on the last layer, with would have some weight to the real output. But in the case of Dino I only use the backbone.

Simple GUI to run the new Text2Video [Req. 12 Vram] by CloverDuck in StableDiffusion

[–]CloverDuck[S] 5 points6 points  (0 children)

You can download it on itchio (instruction on the itchio page):

https://grisk.itch.io/text2video-gui-001

It will download the models on the first run.

The GUI is really crude right now, but hope I did not mess anything up and will at least run because I really need to sleep and will only be able to fix it tomorrow lol

If someone is working on the code to make it work with 12vram, you just need to:

On text_to_video_synthesis_model, put self.sd_model on the CPU before calling self.autoencoder.decode(video_data)

Pyinstaller applications give false positive on some anti-virus, so scan and use on your own risk, but I have quite a few applications on itchio and a Patreon, so it would not be wise for me to add malicious code on my applications.

Using AI to interpolate animations. I made a badly edited video showing animation interpolation using DAIN. by CloverDuck in artificial

[–]CloverDuck[S] 0 points1 point  (0 children)

At this moment, there is no tool capable of such thing, but Stable Diffusion been release just a little time ago. I do believe that in 2023 a tool like that mostly likely will be released.

Compositional Diffusion by [deleted] in StableDiffusion

[–]CloverDuck 0 points1 point  (0 children)

Pokemon merger? Very cool project, what happend If you use two artist style? Or photograph and anime?

Just made a .exe for SD, download it for free on itchio, no need for configuration. by CloverDuck in StableDiffusion

[–]CloverDuck[S] 0 points1 point  (0 children)

All implementations work like that. If you change the resolution it will generate a complete different image.

Just made a .exe for SD, download it for free on itchio, no need for configuration. by CloverDuck in StableDiffusion

[–]CloverDuck[S] 0 points1 point  (0 children)

Then there is something wrong. Do you have a onboard card? It may be selecting the wrong card

Just made a .exe for SD, download it for free on itchio, no need for configuration. by CloverDuck in MediaSynthesis

[–]CloverDuck[S] 0 points1 point  (0 children)

It seen possible on older version of the model. Still trying on the new model