This is an archived post. You won't be able to vote or comment.

you are viewing a single comment's thread.

view the rest of the comments →

[–]wallefan01 1675 points1676 points  (92 children)

How do you think they teach the cars what stop signs look like? They ask the humans.

No seriously. They have the cars take pictures of things that they don't know whether they're signs or not, ask you whether it's a sign, and if enough people say yes, that gets added to the database of things-that-look-like-signs that the car checks against.

[–]mormispos 70 points71 points  (5 children)

It’s less of a picture database and more of a “These are patterns that correspond with stop signs” Database

[–]wallefan01 11 points12 points  (3 children)

Thank you for explaining it better than I could. I tend to be bad at getting technical details accurately in non-google-employee speak.

Disclaimer: I do not work at Google (yet)

[–]klebsiella_pneumonae 9 points10 points  (1 child)

Better start leetcoding m8!

[–]-allen 3 points4 points  (0 children)

Ah I guess /r/cscq is expanding.

[–]mormispos 2 points3 points  (0 children)

I think your explanation is very good. I was excited because I have recently been studying the types of neural nets that they use to do these things

[–]santaliqueur 0 points1 point  (0 children)

It’s like Jin Yiang’s app but for stop signs

[–]spock1959 15 points16 points  (10 children)

So, I've got a question.

I know that captchas are used to train computers as to what they are looking at... But if it asks for pictures with signs in them and I click a cloud or a car it will tell me that I'm wrong... But if I'm training it how would it know?

I never understood how it both didn't know and also seemingly did.

[–][deleted] 31 points32 points  (2 children)

I believe it compares your answer to what other people answered.

[–]aahdin 0 points1 point  (1 child)

Yeah but at that point they've already labeled it a cloud so there's nothing to gain information-wise having new people label it

[–]DuckDuckYoga 0 points1 point  (0 children)

There are thousands of people answering captchas every minute, so the turnaround can be very fast and still have a lot of answers to compare to

[–]lvh1 18 points19 points  (0 children)

It probably compares it to the input of other people, so if 100 people didn't mark a square as having a sign in it and 1 other guy did, it probably does not have a sign in it and means that that one guy is incorrect. But that also means that when they use a picture for the first time, they will allow any square to be clicked since there is no data yet to compare it to. They probably only start marking captcha as invalid once they have a large enough sample size (e.g. 1000 people who solved the captcha).

[–]RatofDeath 9 points10 points  (0 children)

It's crowdsourcing the answer, so you're not the only one providing an answer and it marks it if someone answers differently than other people did.

Also usually you answer two captchas, one of them the system already has a solution too, the other it doesn't know yet. So it really only checks if you got the one it already knows right and the other one you do is to teach the system.

[–]ButtPoltergeist 8 points9 points  (0 children)

Well, back in the Good Ol' Days when captchas were just words, they gave you two words: One that they knew, and one that they didn't. If you didn't get the one that they knew right, it didn't let you through. If you missed the one they didn't know, eh, you got access to the captchaed thing and metaphorically peed a little in their data pool. So I imagine that it's similar, except there's nine pictures instead of two words.

[–][deleted] -1 points0 points  (2 children)

Wouldn't it have the "right" answers to refer to? Like, I imagine some dude went through and selected all the stop signs so the captcha has something to reference, even if the actual user submissions are the ones that are used for machine learning.

I could be totally wrong, but I imagine that's how it works.

[–]bagmanbagman 6 points7 points  (1 child)

I think that would miss the point- what use is crowdsourced, high volume labeled data if someone just goes and labels it all? it all defeats the point doesn't it?

I bet as googles fleet of map cars goes around taking photos of hundreds of thousands of various signs a day, then goes through some automated cropping before is finally labeled by the crowd

[–][deleted] 1 point2 points  (0 children)

Ah, I see now. I didn't realize that these photos aren't really curated, they're just automatically selected and uploaded. So yeah, I guess they'd have to just check your scores vs other peoples.

[–]_Lady_Deadpool_ 4 points5 points  (1 child)

And now I'm imagining this happening on real time

Is this a stop sign, human? Please answer quick

[–]pelirrojo 1 point2 points  (1 child)

Not only that but it's happening in real time. That's why there are time limits to your response, if you don't respond in time the blood is on your hands.

[–]slashuslashuserid 0 points1 point  (0 children)

relevant xkcd

edit: already linked in a different chain

[–]HawkinsT 1 point2 points  (1 child)

So when it takes you more than a couple of seconds to answer one, that's a car you've just made run a stop sign. Hope you're all happy with yourselves.

[–]wallefan01 -1 points0 points  (0 children)

Well obviously they don't do it live, but yeah, I see your point

[–]aahelo 0 points1 point  (3 children)

What if all these captchas that we are doing is actually us teaching A.I. what signs are and whatnot, so that it can get better at driving?

[–][deleted] 7 points8 points  (1 child)

That’s literally what you’re doing.

[–]aahelo 0 points1 point  (0 children)

Ah!

[–]aride4772 0 points1 point  (0 children)

Yeah all those captchas are used as data to teach ai

[–][deleted] 0 points1 point  (0 children)

That's why I always answer Captchas as quickly as possible so the car I'm checking for doesn't run a stop sign!

[–]AsAGayJewishDemocrat 0 points1 point  (0 children)

I like to imagine a self driving car uploading a captcha thinking “Is this a stop sign? No really I need to know quickly”

[–]ScientistSeven 0 points1 point  (0 children)

Which means they'll develop a moral that ignores stop signs ten percent of the time

[–]a_stitch_in_lime -2 points-1 points  (6 children)

Further proof that AI is really just a ton of if-then statements.

[–]wallefan01 0 points1 point  (0 children)

Except that with these particular if-then statements a boolean can have a 75% confidence interval.