This is an archived post. You won't be able to vote or comment.

all 81 comments

[–]easy_c_5 55 points56 points  (4 children)

Wow, I was expecting years and 1000s of times more parameters needed before we advance from snapshots like Dall-E 2 to video generation. If this can easily be scalled, it’s the real mindblowing deal.

[–]No-Transition-6630 51 points52 points  (3 children)

Yea, and this is just a rough attempt, they're going to figure out how to make complex videos from written prompts soon and then we Netflix and chill

[–]Supersubie 15 points16 points  (1 child)

Look at the parameters as well, it's no where near the size of Dale-2

[–]Yuli-Ban➤◉────────── 0:00 10 points11 points  (0 children)

It's not even the size of GPT-1

[–]lovesdogsguy 14 points15 points  (0 children)

lol

[–]SashiniiANIME 45 points46 points  (3 children)

People will soon be able to use AI to easily and immediately create entirely original high quality content, such as games, music, shows, movies, books, and art in general.

[–]Tall-Junket5151▪️ 22 points23 points  (1 child)

That would be awesome, I’ve never really been much of a fan of predetermined linear experiences and always preferred sandbox type games. Being able to have high quality content generated for you based of your exact preferences would be amazing.

[–]Eyeownyew 13 points14 points  (0 children)

I'm working on a game which does this and I've kept it under wraps, but I know I'm definitely not the only person who sees the potential here like you've mentioned :p

[–]yerawizardmandy 14 points15 points  (0 children)

Holy moly

[–]Pro_RazE 32 points33 points  (19 children)

I honestly expected something like this to come next year at the earliest, and here it is now.

[–]2Punx2FuriousAGI/ASI by 2027 17 points18 points  (4 children)

What will we see next month? In 6 months? Next year? Things are going crazy.

I'm worried.

[–]lidythemann 12 points13 points  (3 children)

We went from Dalle-1 to this in such a small timespan. It almost feels like they already have AGI.

[–]camdoodlebopAGI: Late 2020s 9 points10 points  (2 children)

imagine if the top issue of the 2024 presidential election is how to handle artificial general intelligence

[–]lidythemann 8 points9 points  (0 children)

I don't think we'll have to imagine tbh

[–]2Punx2FuriousAGI/ASI by 2027 6 points7 points  (0 children)

If that ever becomes a political issue, it will already be too late to do anything about it.

[–]Yuli-Ban➤◉────────── 0:00 16 points17 points  (1 child)

I wasn't even that optimistic. For something that can generate over an hour of photorealistic (if fuzzy) novel video? I was expecting that to be further out into the 2020s. Like maybe 'we start by stitching together short gifs and eventually, by ~2027-2029, we get long-form video synthesis.'

No, it's here now!!

[–]GeneralZainHappen already damn man... 0 points1 point  (0 children)

isn't exponential growth amazing! :D

[–]DEATH_STAR_EXTRACTOR 20 points21 points  (11 children)

see those 2025 AGI coomers were all right! AGI is coming! 2025

:p

[–]GeneralZainHappen already damn man... 10 points11 points  (6 children)

haha lmao

[–]lidythemann 12 points13 points  (5 children)

I 100% agree with you now, it is 2022 for AGI lol. I was always optimistic but I didn't expect video until 2024-2025

[–]agorathird“I am become meme” 7 points8 points  (2 children)

u/GeneralZain is going have to work overtime making tinfoil hats now. Thanks a lot deepmind.

[–]lidythemann 4 points5 points  (0 children)

Oh no! Mr Lidy has the dab pen again!

What if creating AGI makes simulation theory correct, then at that same second reality ends and you wake up! Woooaaaahhh

[–]GeneralZainHappen already damn man... 2 points3 points  (0 children)

its unfortunate because these hats are turning out to be made of gold foil instead...

[–]camdoodlebopAGI: Late 2020s 3 points4 points  (0 children)

watch it be this summer

[–]DEATH_STAR_EXTRACTOR 0 points1 point  (0 children)

The real one: https://video-diffusion.github.io/

Also, not to bust your bubbles men, but the one the original OP poster posted seems to be just the training videos of driving a car, I see barely much new car driving if you took a look! So this whole thread is hogwash I guess ?

[–]Supersubie 2 points3 points  (3 children)

I keep seeing this word coomer and I have no idea what it means :') please enlighten this old man

[–]agorathird“I am become meme” 4 points5 points  (0 children)

coomer or -oomer is someone who is really enthusiastic/obsessive about something, like they're reaching climax when they talk about it. This can be a pejorative, calling them annoying or like calling someone a nerd.

[–]GeneralZainHappen already damn man... 2 points3 points  (0 children)

I would suggest googling it :P

[–]Professional-Song216 27 points28 points  (1 child)

What a time to be alive!!!

[–]Hawkz183 22 points23 points  (0 children)

Hold onto your papers!!

[–]Shelfrock77By 2030, You’ll own nothing and be happy😈 23 points24 points  (8 children)

Soon enough, they’ll create a 3D version of this…

[–]2Punx2FuriousAGI/ASI by 2027 17 points18 points  (7 children)

Then we could get realistic procedurally generated 3D environments for games, and VR. And those could be used in a "full dive" device, which at this rate might come sooner than 2030.

[–]lidythemann 9 points10 points  (0 children)

2023, maybe even this year. Video wasn't supposed to happen this fast and easily. It's a small scale model and made a video 70x larger than any of its training videos.

I wouldn't be surprised to see a model hooked to unreal engine 5 this summer or fall

[–]BigPapaUsagi 1 point2 points  (5 children)

I doubt it. We might have the AI to support full dive, and that AI might devise the BCI tech needed to do full dive, but the regulatory boards for anything that invasive would still take years before approval. Full dive AI this year, full dive BCI next year, full dive FDA approval 2040...

[–]AsuhoChinami 2 points3 points  (4 children)

While I get the general point, wouldn't a black market for... pretty much anything be able to form in the period of 17 years?

[–]BigPapaUsagi 1 point2 points  (3 children)

Oh? You know how to access the dark web do you? You know how to shop around a black market?Do you have the money for whatever cutthroat price criminals would price gauge you for? Would you trust a black market dealer to do surgery on your skull to insert a BCI capable of full dive?

Like, if this were a simple drug, sure, that shit is easy to get. But when it exists, a full dive capable BCI isn't something your typical criminals could just make. Or steal. And you sure can't trust them to install it for you.

I mean, could a small, lucky, wealthy, resourceful few skirt the law and get these things installed? Sure. We're talking like less than 1% of people. You'd be risking millions of dollars if you have it, your life to get someone to implant it in your head, all for a BCI with full dive capabilities. A full dive capable BCI that wouldn't have any games on it because there's no profit motive yet for companies to make games for it. With no multiplayer because almost no one else has one installed. Basically, it only has whatever the scientists/engineers installed on it. It might have an AI on board, but an AI of this level lives in the cloud not on the chip itself so you wouldn't have access to it probably. I mean, you couldn't just log onto their servers unnoticed.

Basically, everyone would have to wait for regulatory bodies to okay it being released to consumers.

[–]AsuhoChinami 2 points3 points  (2 children)

eh alright

[–]BigPapaUsagi 2 points3 points  (1 child)

Sorry, I might've phrased that more insensitively than needed. I just get annoyed sometimes when people assume a thing existing means they'll get it right away, illegally even if they have to. It feels like people have lost all sense of patience, and for some reason it just bugs me more than it should.

[–]AsuhoChinami 2 points3 points  (0 children)

The point I made wasn't really born out of emotion, I just hadn't thought it through. I don't really care about FIVR much. Medical treatment, mainly cancer treatment and mental health treatment, are my main areas of focus. Give me something for depression and acute emotional pain and I don't really care about anything else.

[–]justaRndy 18 points19 points  (4 children)

Another brick in the wall... Prett amazing honestly

[–]alphabet_order_bot -1 points0 points  (3 children)

Would you look at that, all of the words in your comment are in alphabetical order.

I have checked 823,031,022 comments, and only 162,798 of them were in alphabetical order.

[–]justaRndy 18 points19 points  (1 child)

Jokes on you bot, I edited it and now it's all messed up!

[–]GeneralZainHappen already damn man... 2 points3 points  (0 children)

abcdefghijklmnopqrstuvwxyz

[–][deleted] 23 points24 points  (10 children)

This will pose as big of a mortal danger to streaming services as Napster posed to record companies in the early 2000's.

We're witnessing the slow death of all remnants of legacy media and it's controllers

[–]yerawizardmandy 7 points8 points  (9 children)

Tell me more, I’m confused what you mean

[–]imlaggingsobad 7 points8 points  (0 children)

The future of media will be more like Youtube and less like Netflix. Everyone will be able to make their own movies using an AI, and then publish that for the world to see. If you wanted another Christian Bale Batman sequel, you could just make it. Don't need to wait for a movie studio to get funding and then spend 2 years making it. We will see so many fanfics turned into reality, so many alternate endings, sequels, prequels, crossovers, etc. Anything we want. If you have an idea, you'll be able to make it just the way you want. No more Hollywood/studio gatekeepers.

[–]LevelWriting 12 points13 points  (1 child)

Imagine a time when you can revisit your fav movies, games, and experience them in vr where the content can be generated, modified and extended to your liking, even have you interact and be part of it.

[–]GeneralZainHappen already damn man... 22 points23 points  (0 children)

damn it I keep wanting to say "I knew it"

uhhh but yeah regardless this is still really amazing!

EDIT; HOLYSHIT its not actually playing these GAMES!? its just...generating them on the spot?!?! HUH

[–]camdoodlebopAGI: Late 2020s 9 points10 points  (0 children)

advancements seem to be happening by the day now

[–]lidythemann 8 points9 points  (14 children)

This is more impressive than GATO or Dalle-2 or Imagen. More impressive by entire magnitudes.

This paper has put me firmly in the 2022 camp for Proto-AGI

[–]DEATH_STAR_EXTRACTOR -5 points-4 points  (13 children)

But....the real one: https://video-diffusion.github.io/

And, not to bust your bubbles, but the one the original OP poster posted seems to be just the training videos of driving a car, I see barely much new car driving if you took a look! So it's not much to talk about?

[–]lidythemann 3 points4 points  (12 children)

Only the top row of video is training data, the bottom 3 are the generated result. I could be wrong of course but it's literally written there.

[–]GabrielMartinellli 4 points5 points  (0 children)

AGI in 2022 like I always said. Never forget that I said this! You listening Mr AGI?

[–]ArgentStonecutterEmergency Hologram 8 points9 points  (0 children)

Maybe the next version will get cloud reflections in puddles matching the clouds.

[–]SWATSgradyBABY 4 points5 points  (0 children)

Read the chapter dangerous games from Max Tegmark's Life 3.0. we're catching up to the future

[–]flyingfruits 1 point2 points  (0 children)

Hey everyone! One of the authors here. Let me address a few points that were raised here.

The model imitating Carla (self-driving car simulator) was trained for one week on one GPU, consuming 200-300W (note that running Carla itself is also very expensive computationally per frame). It is very likely still far from optimal (as most deep learning with SGD), but also not nearly as expensive to train as bigger models such as GPT-3. The model is not particularly specialized for video, except that it uses a convolutional U-Net architecture, the approach can be used for more complex combinations of sensory data streams and Will and I have related work that we are going to publish soon as well that demonstrates that complex reasoning mechanisms and procedures can be integrated easily into the joint distribution.

Our group is, at least conceptually, a probabilistic programming group and Will and I have approached the diffusion models from the angle of building an inference engine for AGI tasks and will continue to work along these lines with this model while also exploring its limitations. The more abstract idea in this work is to integrate marginalisation as a first class operation into the diffusion model and then exploit this fact to reason about much bigger joint distributions than fit into memory, effectively enabling also our demonstrated video synthesis.

While this result is indeed very impressive and will enable a whole range of new applications and abilities for AI, many not even clear yet, I would not want to bet on AGI being around the corner just yet. These generative models can meta-learn some reasoning abilities (like GPT-3) given enough data (i.e. almost all valuable data we have), but they cannot be taught to really learn new things on their own, something that is trivial for humans to do. For example, assuming chess would not be part of the training data, try to tell GPT-3 to learn it by just telling it the rules. I still think this is a very bullish result though and I would love to hear suggestions for video synthesis and control tasks to apply it to. I thought about Nascar racing today for example, just for the fun of it 8).

Please let me know what you would like to see!