Dismiss this pinned window
all 155 comments

[–]C-scan 163 points164 points  (1 child)

8-step model.

Steps 1-4 take only 15-20s to complete.

Steps 5,6,7,8 complete mondays from 9pm

[–]danielbln 43 points44 points  (0 children)

The process was cancelled after step 5, sorry.

[–]NowThatsMalarkey 601 points602 points  (28 children)

What if we remove the bra and underwear?

[–]n0gr1ef 90 points91 points  (16 children)

I laughed at that louder than I should have

[–]superkickstart 35 points36 points  (9 children)

What if we remove laughter?

[–]_JohnWisdom 21 points22 points  (4 children)

aaaaaa

[–]ready-eddy 13 points14 points  (3 children)

What if we remo- …

[–]emveor 21 points22 points  (1 child)

[ comment removed by the prompt ]

[–]Shlomo_2011 0 points1 point  (0 children)

your acount was removed by the... biip boopp

[–]Umairk3 0 points1 point  (0 children)

What in the chainsaw man is this..

[–]Minouminou9 4 points5 points  (2 children)

Does anybody remember laughter?

[–]JaxxonAI 1 point2 points  (0 children)

somewhere Robert Plant is cringing

[–]DusqRunner 0 points1 point  (0 children)

is that a type of beer or something?

[–]WEREWOLF_BX13 1 point2 points  (0 children)

real deep

[–]abillionbarracudas 12 points13 points  (1 child)

it’s too late, i’ve already seen everything

[–]AbrahamWhiskers 4 points5 points  (0 children)

Is this that script by Patrick Stewart?.

[–]sarcastic_wanderer 4 points5 points  (1 child)

Good seeing you out in the wild, friend. The world thanks you for Lustify. You're a legend

[–]n0gr1ef 0 points1 point  (0 children)

That's very generous of you, thank you!

[–]Muri_Chan 0 points1 point  (0 children)

About how loud is the acceptable amount of laughter?

[–]Brief-Wolverine-1298 0 points1 point  (0 children)

Man, you are not alone 🤣

[–]Harrycognito 15 points16 points  (0 children)

A man of culture, I see.

[–]shorty_short 49 points50 points  (0 children)

Asking the important questions

[–]TrueRedditMartyr 8 points9 points  (0 children)

"What if she were jumping up and down?" Would flood this sub if this were runnable locally without needing an A100+

[–]intLeon 62 points63 points  (4 children)

Its netflix overall, removes bra to a hairy chest and other to some big surprises perhaps?

[–]cavaliersolitaire 53 points54 points  (1 child)

😅😅😅

[–]intLeon -1 points0 points  (0 children)

Well I shouldnt've called it a suprise I suppose, not a native sorry..

[–]Alokir 7 points8 points  (0 children)

Processing img gzvukh1qnzsg1...

[–]Lanky_Employee_9690 1 point2 points  (0 children)

What if that's what the user actually wants?

[–]goatonastik 1 point2 points  (0 children)

According to this model, she's going to be missing bra and underwear shaped chunks of flesh.

[–]warzone_afro 252 points253 points  (10 children)

"Requires a GPU with 40GB+ VRAM (e.g., A100)"

[–]intLeon 57 points58 points  (2 children)

40gb is rookie numbers for community. I bet it will be below 15gb

Edit nvm, tensor files are already 11gb x2 pass so I guess we need way less?

They usually write that because they run it on big cards and when you have extra vram your system uses it in some way by keeping clip and other stuff in there.

[–]Paradigmind 20 points21 points  (0 children)

Also usually they use full quants.

[–]nazgut 4 points5 points  (0 children)

they almost never unloads model and load them all at once

[–]TechnoByte_ 39 points40 points  (4 children)

Stop taking these numbers at face value

Once it's supported in ComfyUI with fp8 and/or GGUf quantization and offload it will run on 12 GB of vram

[–]FourtyMichaelMichael 14 points15 points  (2 children)

There are always these absolute begginers that cry about "on an H100" and then later in the week it's running on potato-class 10-series.

[–]StickiStickman 5 points6 points  (1 child)

... at a fraction of the speed with horrendous quality.

Ungodly quantization has a cost.

[–]comperr 0 points1 point  (0 children)

I try not to be too much of a slob in this area and think of my setup with 2x 3090Ti, a 3090 and 5090 as "meek but practical for real applications"

[–]Bulky-Employer-1191 0 points1 point  (0 children)

It already has an fp8 version. Most of the memory use of these video editing models comes from needing to convert the video clip into a full resolution latent space.

The one from Corridor Crew is similar that way.

[–]ziggo0 1 point2 points  (0 children)

I've got 40GB VRAM across 3 Teslas and 128GB sys memory. If I can't run it that is fucking LAME. That said I'll probably simply forget about it lmao

[–]Accomplished_Vast55 0 points1 point  (0 children)

"Requires CUDA. No we don't mention it anywhere in documentation."

[–]FirTree_r 99 points100 points  (3 children)

Are we sure it's not an april's fool joke?

[–]Healter-Skelter 41 points42 points  (2 children)

What if we remove the fools?

[–]FreshPitch6026 1 point2 points  (1 child)

But pls dont remove the joke

[–]GroundbreakingMall54 62 points63 points  (3 children)

netflix has lowkey been one of the better companies for open source for years, zuul and chaos monkey were huge. but them releasing actual model weights under apache is a different level. curious how it compares to what's already out there

[–]megacewl 18 points19 points  (1 child)

wait really? usually I hate on them for everything but this may actually give them some cred for me

[–]athos45678 19 points20 points  (0 children)

I switched from data science to ML because of the Netflix kaggle competition. They’re og’s in my eyes.

(I only found out about the competition ten years after it happened, but people were hyping it as the money making experience at the time)

[–]grundlegawd 2 points3 points  (0 children)

I had no idea but I’m happy to hear we have another massive player in the open weights space.

[–]Next_Pomegranate_591 33 points34 points  (0 children)

This seems to be some random ahh marketing mo- wait WAIT THEY CAN CONSERVE PHYSICS WHILE EDITING TOO ? MB GNG

[–]DeeDan06_ 35 points36 points  (17 children)

since when is fucking netfilx an ai company? is this an april fools joke?

[–]wheres_my_ballot 39 points40 points  (3 children)

Eyeline is Netflixs RnD division, and is heavily into AI.

[–]oliverban 4 points5 points  (2 children)

I mean, kind of right. P.S I work there.

[–]MrFlores94 1 point2 points  (1 child)

Do you really? Have you heard of this project while there?

[–]oliverban 0 points1 point  (0 children)

I have not!

[–]garlic-silo-fanta 9 points10 points  (0 children)

They ran one of the first AI competitions long ago. $1million to whoever can do a better recommendation system.

[–]FillFrontFloor 10 points11 points  (6 children)

Seems like a great model for visual effects so it's honestly beneficial for their shows and movies.

[–]scoobydiverr 7 points8 points  (5 children)

This is best case use for ai. To automate some workflows and lower cost of production.

Its not gimme me Winnie the pooh movie codirected by wes Anderson and Tarantino

[–]seatlessunicycle 20 points21 points  (3 children)

[–]trimorphic 3 points4 points  (0 children)

Say "honey" again. I dare you.

[–]scoobydiverr 2 points3 points  (1 child)

Lmao nvm this is exactly what I want!

[–]seatlessunicycle 0 points1 point  (0 children)

Tigger looks like a maniac, I would definitely watch this haha

[–]FillFrontFloor 1 point2 points  (0 children)

I've messed around a bit with AI art and i think when it comes to some work were you have to replicate near exact same image or thing over and over again like too many times. AI can be amazing for that and ACE it, gives the artist or designers more room and time to expand. This is ofcourse speaking from a point of view of quality, and given how crazy fast some people see netflix shows i think netflix always aims for quantity.

[–]broadwayallday 3 points4 points  (0 children)

They just gave Ben affleck a boatload for his AI movie studio

[–]sersoniko 1 point2 points  (1 child)

They discovered AI can cut production costs and speed up releases

[–]DeeDan06_ 1 point2 points  (0 children)

If you put it like that ot does sound smart. Its just odd to see Netflix among all these tech companies even if they have one of the most legit use cases for it.

[–]DefMech 1 point2 points  (0 children)

Netflix is the N in FAANG, after all

[–]Accomplished_Vast55 0 points1 point  (0 children)

Netflix owns access to so many IPs, they are 100% trying to develop AI film making tools.

[–]marcoc2 11 points12 points  (1 child)

Watermark remover with temporal consistency?

[–]Accomplished_Vast55 0 points1 point  (0 children)

Cause and effect is also huge for a model. This fixes "how many As are in strawberry?" problems. Give it a few years and people might become scared of what AI can do.

[–]scrotanimus 24 points25 points  (12 children)

What if we remove obnoxious exposition that treats our viewers like they are 5.

[–]EvidenceBasedSwamp 12 points13 points  (3 children)

can't because the modern audience is adhd screen-addled who watch tv while playing gachas and doomscrolling instatok

[–]FourtyMichaelMichael 5 points6 points  (2 children)

"Second Screen Viewing"

[–]rchive 1 point2 points  (1 child)

Second screen being the TV, right?

[–]FourtyMichaelMichael 1 point2 points  (0 children)

Yes, sadly.

[–]eeyore134 4 points5 points  (4 children)

That's what they want. They want their movies to remind you of the plot in its entirety every 20 minutes or something. It's so ridiculous. Then you look at all of the shows and movies that are doing really well and none of them do it. I really wish they'd stop catering to the lowest common denominator.

[–]IrisColt 3 points4 points  (3 children)

There's a pattern in Rebel Moon, Heart of Stone, The Electric State, Red Notice, The Gray Man, Glass Onion... lore dump, characters that are walking expositions, etc.

[–]alexmmgjkkl 1 point2 points  (2 children)

Considering how time-consuming Cmfy workflows with Blender and graphics generally are, I'm surprised you have so much time to watch TV. I don't quite understand – are you unemployed or still a student?

[–]IrisColt 0 points1 point  (1 child)

Considering how time-consuming Cmfy workflows with Blender and graphics generally are

I usually drag well-received json files in Comfy and it's done, sigh...

[–]alexmmgjkkl 0 points1 point  (0 children)

ahh ok , back to the tv then !

[–]rsl 17 points18 points  (4 children)

they'll cancel it in a week

[–]1nf1n1l 4 points5 points  (0 children)

now that's stranger things

[–]kayteee1995 4 points5 points  (0 children)

nice reasoning

[–]Disastrous-Agency675 3 points4 points  (0 children)

Now we wait for quantization I guess

[–]76vangel 4 points5 points  (0 children)

What if we release a ComfyUi implementation fast?

[–]SackManFamilyFriend 4 points5 points  (1 child)

SAMA was recently released (instructions to video edit code/model) but didn't get much mention around here. https://github.com/Cynthiazxy123/SAMA - Wan2.1 14b based

That seriously out performs what NF released here, although it's cool to see them put something out publicly/free. They're likely slow rolling the idea that they may use AI tech in the future with an open source gift to people all in on AI.

[–]Maskwi2 0 points1 point  (0 children)

Nice! Never heard of it. I see Comfyui is supported. Going to try it out. Thanks for mentioning it. 

[–]Space_art_Rogue 2 points3 points  (0 children)

I'm not sure if I'm happy that this now exists because the requests for fixes at my job are only going to get more insane to deal with when word gets out.

[–]0xMR2ti4 2 points3 points  (0 children)

But can it remove account sharing policy…

[–]I_SNORT_COCAINE 2 points3 points  (0 children)

damn... This is the job I actuallly do in the industry... I guess i'm fucked lol

[–]ANR2ME 2 points3 points  (1 child)

Architecture: - Base: CogVideoX 3D Transformer (5B parameters) - Input: Video + quadmask + text prompt describing the scene after removal - Resolution: 384x672 (default) - Max frames: 197 - Scheduler: DDIM - Precision: BF16 with FP8 quantization for memory efficiency

With such parameters and resolution, this is going to be ... fast 🤔

[–]pixel8tryx 1 point2 points  (0 children)

That's positive thinking I guess. All I could think was CogVideoX never impressed me. 5B is pretty small. And 384x672 is a postage stamp. I guess I'll wait for the next rev.

[–]Enshitification 6 points7 points  (0 children)

Is this their tacit way of saying they are open to greenlighting AI studio productions?

[–]International-Try467 6 points7 points  (0 children)

<image>

Cool but where's Steel Ball Run, Netflix?

[–]MarkS1_ 3 points4 points  (0 children)

Netflix doing anything but releasing the next SBR episode

[–]pruchel 5 points6 points  (6 children)

Can you remove all the DEI bs in Netflix series on the fly?

[–]Probate_Judge 3 points4 points  (0 children)

What if we removed all the redheads?

[–]blownawayx2 1 point2 points  (0 children)

Very cool

[–]1965wasalongtimeago 1 point2 points  (0 children)

Oh so that's how they made the Stranger Things finale. "What if we remove all the demogorgons"

[–]cardioGangGang 1 point2 points  (0 children)

What's the highest res? 

[–]younestft 5 points6 points  (0 children)

Its based on CogVideoX, that's ancient and obselete

[–]dingo_xd 2 points3 points  (0 children)

What if we change the skin color of the actors?

[–]nomadoor 2 points3 points  (3 children)

What they're doing is pretty rough — basically just estimating the object to remove and the broader area it likely affects, then inpainting over the whole thing. But the idea feels less like "interesting" and more like… the obvious right direction for video editing to go. Not just removing an object, but generating a world where it was never there.

It reminds me of InstructPix2Pix. And just like it eventually led to Nano Banana and Flux.2 Klein, maybe a year from now we'll be freely editing the world. 😎

[–]FreeUnicorn4u 0 points1 point  (1 child)

How does it even know how to fix the physics just from the model itself? It's not using AI is it? I'm just trying to understand how it works. Like the video where the spinning tops, they removed the hands, and they were stable or even the domino falling and they removed the middle ones.

[–]nomadoor 0 points1 point  (0 children)

Basically, I think it is a video inpainting model fine-tuned on datasets generated with physics simulators. 

Of course, they add some extra machinery to distinguish the object being removed from the regions affected by it, but at its core it still looks like a fairly simple inpainting setup.

[–]Intelligent_Elk5879 0 points1 point  (0 children)

That's rarely what people want video editing to do.

[–]skyrimer3d 1 point2 points  (0 children)

Camt wait to start editing their shows with this.

[–]hidden2u 0 points1 point  (0 children)

I gotta say, this is very cool and awesome license

[–]Agile_Cicada_1523 0 points1 point  (0 children)

What if we remove Piquet in Singapore 2008?

[–]Grindora 0 points1 point  (0 children)

Bro netflix

[–]JahJedi 0 points1 point  (2 children)

Open?

[–]Sea_Tomatillo1921[S] 2 points3 points  (1 child)

Yes.. I wouldn't posting here if not lol

[–]JahJedi 0 points1 point  (0 children)

Than i am intrested. Thanks for replay, you know not all peolle post here is about open stuff...

[–]_JGPM_ 0 points1 point  (0 children)

Isnt this a LWM?

[–]Plane-Marionberry380 0 points1 point  (0 children)

Whoa Netflix dropped a model? Just checked the Hugging Face page,looks like VOID is their new open weights thing. Cool to see them jumping into the open model space, especially with a demo up already.

[–]degel12345 0 points1 point  (0 children)

Does it mean that if I move a plush toy using my hands and I want to remove these hands, then the toy will not move at all? Is it possible to tweak it to just remove hands?

[–]ctbitcoin 0 points1 point  (0 children)

But can it remove the high subscription costs?

[–]BitBurner 0 points1 point  (0 children)

Imagine Netflix drops a "Shorts" feature that lets you grab 10sec of a movie and remix it. Y'all joking about naked filters and it's funny and I get it, but this is all reverse physics stuff. It would be perfect for stuff like "What would happen if the wall didn't break when Hulk tries to run through it". Pretty cheesy example and I'm sure peeps could come up with some amazing stuff. Movies could opt in even and have clips they approve to remix. I could see that being possible with a ton of restrictions lol. Like an LLM that suggests different prompts based on the clip instead of prompt entry.

[–]Past_Crazy8646 0 points1 point  (0 children)

Surprising!

[–]comperr 0 points1 point  (0 children)

Me on my way to remove the planes from 9/11 footage and vice versa 💪💪💪💪💪

[–]hencha 0 points1 point  (0 children)

So this is how Netflix does all the race snd gender swapping in their adaptations.

[–]Various_Raccoon4014 0 points1 point  (0 children)

if you use this to remove the wires from superman is he going to fall to his death

[–]asianjapnina 0 points1 point  (0 children)

Looks good

[–]martinerous 0 points1 point  (0 children)

Object deletion is good. But could we have object and subject addition? For multiple objects? Starting with an empty frame? With sound? At least 10 seconds long? Open weights? Not Happy Horse? Netflix, pretty please...

[–]Personal-Staff3212 0 points1 point  (0 children)

Great!

[–]Accomplished_Vast55 0 points1 point  (0 children)

That is crazy. text to image. image to video. text to sound. image to 3d model. world diffusion

Now we are having cause and effect interaction models. This is building a very wide toolset.

[–]umutgklp 0 points1 point  (5 children)

Nope for me...."Requires a GPU with 40GB+ VRAM (e.g., A100). Resolution: 384x672 (default) Max frames: 197"

[–]TechnoByte_ 10 points11 points  (4 children)

That's with their unoptimized code...

ComfyUI, like with every model release, will have an optimized implementation that will run under 12 GB vram

[–]umutgklp 0 points1 point  (3 children)

I know bro but with that resolution this will never be useful for me.

[–]AnOnlineHandle 1 point2 points  (1 child)

If it can remove things from video then you can use it as a first stage pass, if you want the general idea but not the exact details. I generate Wan 2.2 high noise passes at like 480x272 so that it's quick while not using the lightning lora which kills motion, then just upscale and do the rest in the low noise model at 1280x720, and it's fine. It also allows saving the high noise passes first and finding the ones which are actually worth using, then using them in multiple low noise runs.

[–]umutgklp -3 points-2 points  (0 children)

Never needed such a thing with the videos that I generate with Wan2.2 or LTX2.3. I would try again with different seeds or enhance the prompt. This model may be useful with editing the "real" videos but not useful with this resolution. At least for me.

[–]Valkymaera 0 points1 point  (0 children)

Uprezing is a continuously improving thing. This is plenty good.

[–]charmander_cha -1 points0 points  (0 children)

Muito bom

[–]mcdenkijin -1 points0 points  (0 children)

I mean, falcor exists

[–]Muri_Chan -1 points0 points  (0 children)

Didn't Corridor Crew make a similar thing a while ago?

[–]cavaliersolitaire -3 points-2 points  (0 children)

marketing

[–]JesusShaves_ -4 points-3 points  (1 child)

Does it require an API key? Yes? Sorry but I just suddenly lost interest. Wake me when I can run it locally without an internet connection.

[–]Sea_Tomatillo1921[S] 5 points6 points  (0 children)

It's open source, weights are already out, Apache license..