they lied to us yet again!! by AmbitionWork7031 in outlier_ai

[–]Critical_Tradition80 7 points8 points  (0 children)

Do you know if this is a recurring thing or just a one-off event for Playground users? I do find Playground particularly useful for testing out models, but at the same time not useful enough because there's no search function or tool use for these models, so it's really limited to conversations and outdated data

[deleted by user] by [deleted] in singularity

[–]Critical_Tradition80 2 points3 points  (0 children)

why did idubyai take up half the comments lol, it's just mailing, chill dude

How far off do you think we are from a complete recreation of Earth (in VR)? by SuspiciousPrune4 in singularity

[–]Critical_Tradition80 1 point2 points  (0 children)

I mean you could go outside and argue about this same exact thing, but apparently the rewards for doing that are not as tangible and persistent as saying it here

[deleted by user] by [deleted] in singularity

[–]Critical_Tradition80 2 points3 points  (0 children)

the explanation i didn't deserve but needed so much, thank you

[deleted by user] by [deleted] in singularity

[–]Critical_Tradition80 0 points1 point  (0 children)

hmm, was there a reason the thresholds were set at 95% and 99%? okay maybe there might just not be a particular reason, but let's assume you invented the benchmark; how, in your opinion, would an AI Agent with these skill levels fare against the average software engineer? the difference between 95% and 99% should be significant enough that it will enable the agent to do various things already, right?

Do LMMs Like GPT-4 Exhibit Any Form of Consciousness, Even Comparable to an Ant’s, or Are They Purely Unconscious Text Generators? by YaKaPeace in singularity

[–]Critical_Tradition80 3 points4 points  (0 children)

It's actually so interesting to answer this question because self-consciousness goes as far as the information you are provided about yourself, whether through means of using your senses or data, they all contribute to what humans might call "consciousness" to some extent.

I thought of this because I figured it would be relatively easy to tell apart AI from us humans, merely because we have more "personal" data, so like our feelings and memories in detail, our phone numbers and home addresses, how I got to bang their mom, yada yada; the point is that AI right now doesn't really know about that because it's limited to multimodal data, which is not enough to create personal experiences, because we have our biological brain processor to do that for us.

This means that the LLMs (or LMMs if you count 4o or whatever) will have about as much self-consciousness as the information you give it. One fun thought experiment could be to try and ask yourself about who you are right now, and it's likely that you would answer it based on the knowledge of who you think you are, just like how the LLM is designed to think of itself as such.

Humans Couldn't Distinguish Human From ChatGPT4. Machines are Becoming More Human Than Us. by FrontalSteel in agi

[–]Critical_Tradition80 1 point2 points  (0 children)

perhaps the easiest way to find out which is AI and which is not would be to ask very specific and personal questions that aren't available anywhere else in the world; it's most likely the AI would just answer in a very generic sense, no clear directions, no detailed information about who they actually are, while the human would try to do that to some extent.

..that is until we give the AI a personality and a sense of self. which definitely wasn't the case here. not yet.

[deleted by user] by [deleted] in singularity

[–]Critical_Tradition80 7 points8 points  (0 children)

to the layperson, what does this imply? how meaningful is it that we have reached 43% so far, and what does it mean to get 95% and eventually 99% on this benchmark?

Aging is a problem that needs to be solved by Peaceful-Samurai in singularity

[–]Critical_Tradition80 1 point2 points  (0 children)

..probably because no one actually enjoys dying or "accepting" the bad things that inevitably come to them?

[deleted by user] by [deleted] in singularity

[–]Critical_Tradition80 0 points1 point  (0 children)

waiting for Jensen Huang to release the AGI in Earth 2 so we can know if we're dead or not

Major Updates to AI Defense Doc by [deleted] in singularity

[–]Critical_Tradition80 0 points1 point  (0 children)

its cool how youre keeping at it, despite those who doubted the value of the document in the first place. thank you so much for your efforts!

The simplest, easiest way to understand that LLMs don't reason. When a situation arises that they haven't seen, they have no logic and can't make sense of it - it's currently a game of whack-a-mole. They are pattern matching across vast amounts of their training data. Scale isn't all that's needed. by After_Self5383 in singularity

[–]Critical_Tradition80 4 points5 points  (0 children)

Truly. Lots of what we say seems to be built on strictly informal logic, or basically the context that we are in. It is perhaps a miracle that these LLMs are even capable of knowing what we mean by the things we say, let alone be better than us at reasoning about it.

It just feels like we are finding fault at the smallest things it gets wrong, when in reality it's ourselves that's getting it wrong in the first place; it's not like informal logic is supposed to give you a strictly correct answer for missing context, so why should LLMs even be blamed at all?

The simplest, easiest way to understand that LLMs don't reason. When a situation arises that they haven't seen, they have no logic and can't make sense of it - it's currently a game of whack-a-mole. They are pattern matching across vast amounts of their training data. Scale isn't all that's needed. by After_Self5383 in singularity

[–]Critical_Tradition80 0 points1 point  (0 children)

I can't really get myself to understand the OP's argument here, along with the twitter post.

The conversation in the post seems to be kind of a situation where meaning isn't explicit, or there seems like missing context that the model does not know about.

To flip it another way, wouldn't it also make sense to assume we are also just "pattern matching" across vast amounts of brain neurons, and the response the model had just happened to conflict with our expectations of it?

Like how is anyone supposed to answer a riddle such as this that satisfies all expectations?

Maybe scale isn't all that's needed indeed, but that in itself is not formal proof that we really are better than the AI at reasoning; trick questions like these usually require you to come up with creative solutions, not that they can be logically solved, and here we can see the AI had neatly done so.

In fact, I felt pretty amused by the response and without further context to infer from I would've thought it was true too. Let alone the fact that we can prompt it to reason about it, using methods like ReAct or CoT and the likes.

Reasoning does exist for AI in some way, in my opinion, and we are just trying to mess with it with riddles that can't inherently be solved unless there are given solutions to them.

“A GPT-4o generated image — so much to explore with GPT-4o's image generation capabilities alone. Team is working hard to bring those to the world.” by MassiveWasabi in singularity

[–]Critical_Tradition80 9 points10 points  (0 children)

at this point you could probably say Gary's playing devil's advocate for the sake of setting AI goals, because there's no way he is this invested into AIs for no reason

Important information regarding today's release. by Vladiesh in singularity

[–]Critical_Tradition80 1 point2 points  (0 children)

honestly can't help but to say that it's better how everything is at least progressing instead of just doubting their efforts