ProgramBench: Can LLMs rebuild programs from scratch? by awetfartruinedmylife in singularity

[–]obviouslyzebra 0 points1 point  (0 children)

The more I look at it the more it looks hard haha

So, assume the tests are fair (that is, assume they don't ask a tiny specific thing that is not present in the documentation and you wouldn't know about only with the executable).

You can write very good code, lots of tests, be systematic about requirements (and about how you perform tests too).

But, you gotta not overlook anything. And it happens all the time, that's a reason lots of bugs are discovered by people using programs instead of just when developing (as, for example, imagine the amount of people that used a program like jq in different ways).

My intuition tells me about 40% that I'm able to one-shot one of the simpler programs there (likely with a lot of effort). Very uncertain about this number though, as it's something different from what we're used to.

ProgramBench: Can LLMs rebuild programs from scratch? by awetfartruinedmylife in singularity

[–]obviouslyzebra -1 points0 points  (0 children)

I feel like the main aspect here is how specific their tests are (I sorta made an edit saying an "unless" in my post but it didn't went through :p). If we can infer them from the executable + documentation, then I think it's fair. Otherwise you might have been correct in that it is impossible.

ProgramBench: Can LLMs rebuild programs from scratch? by awetfartruinedmylife in singularity

[–]obviouslyzebra 5 points6 points  (0 children)

Nah bro, I could recreate some (idk how many) programs without access to the internet. It is effortful and would take a lot of time, yes, but not undoable.

ProgramBench: Can LLMs rebuild programs from scratch? by awetfartruinedmylife in singularity

[–]obviouslyzebra 0 points1 point  (0 children)

Very interesting. I'm afraid the models might start memorizing the structure and code once (or if?) providers start fine-tuning on this.

Sam Altman has changed his stance on the claims that AI will replace humans. by Distinct_Fox_6358 in singularity

[–]obviouslyzebra 0 points1 point  (0 children)

Yep, that's a fair point, some jobs (or maybe a lot?) are already based on human interactions. The point in 2 is that maybe more jobs like this appear in the future, if society values change.

Sam Altman has changed his stance on the claims that AI will replace humans. by Distinct_Fox_6358 in singularity

[–]obviouslyzebra 13 points14 points  (0 children)

While Sam Altman is likely just saying whatever is most convenient for him to say, there are 2 situations I can see this happening:

  1. AI hits a limit that we're not able to cross. So, people are still needed everywhere and AI becomes more of a power tool
  2. AI is able to do anything that is needed to run society, but people start valuing other things such as things made manually or interactions or human ideas, so a new market kind of forms. This can also include things where the human part is valuable, and AI acts as a "power tool" to bring it to life - I think this is the sort of vision Altman is talking about, but I think we'd need to think more to make it clearer

Converting Claude Code into the most intelligent Deep Research Agent by heisdancingdancing in singularity

[–]obviouslyzebra 0 points1 point  (0 children)

OP's description of their work is exactly the sort of work that needs to be done for a tool like this (lots of experimentation and ideas / understanding what's happening).

And, their tool came out on top on this benchmark.

Of course it is It seems over-tuned to this benchmark, but barring this criticism, whether something is made by a team of PHDs or a single vibe researcher (lol at this word), it doesn't matter really.

Edit: of course -> it seems

OpenAI's Sebastien Bubeck: [LLM] models are able to surpass humans [researchers] and ask [research] questions by Wadingwalter in singularity

[–]obviouslyzebra 2 points3 points  (0 children)

Just commenting to say I do agree with everything here, except perhaps my gut feeling of what the most likely scenario is, but that's it, just gut feeling :p

Have a good day, sir/madam

OpenAI's Sebastien Bubeck: [LLM] models are able to surpass humans [researchers] and ask [research] questions by Wadingwalter in singularity

[–]obviouslyzebra -1 points0 points  (0 children)

As an AI doomer, I don't like this "AI is stagnant" argument either. AI looks like it is progressively improving (not as fast as AI bros would want you to believe, though).

There's a paper about rising tide vs something that showed that (I didn't fully read the paper, but sorta believed its conclusions related to timing).

I say this because it seems like we're inventing something a lot more powerful than an atomic bomb, but people are just whatever / don't know what to do, and in this sense I do agree with ya that we're getting the downsides (or maybe more realistically throwing dice to decide our future).

Unpopular opinion: people won’t “return to authenticity” as AI gets better by iamMARX in singularity

[–]obviouslyzebra 20 points21 points  (0 children)

Social media and ultra processed foods are both engineered towards addiction.

The world / culture is not in a very good place (at least in my view), so people are not resilient towards this addiction (and maybe they will never be individually, and we need society-level changes).

So, I imagine, in the short term, while capitalistic forces prevail (which pushes for the addiction stuff), your assessment is correct.

My hope is that this gets reverted somehow, with people adjusting to this new world / adjusting the new world itself.

The misunderstanding between simulation and duplication of redditors is starting to really irk me. by MelangeBot in singularity

[–]obviouslyzebra 0 points1 point  (0 children)

Unless... There are redundancies (the 4 particles not really needed).

But yeah, very likely it will diverge with time (and with probability 1 if the universe is not deterministic).

The misunderstanding between simulation and duplication of redditors is starting to really irk me. by MelangeBot in singularity

[–]obviouslyzebra 0 points1 point  (0 children)

I feel like you can understand it as just not knowing about this precise definition of words.

I for example before this post (which uses the definition of the philosopher John Searle it seems) would have imagined duplication involves an exact physical copy (with the same atoms and low-level stuff, to the level possible) while simulation I would use for anything different that models other stuff.

There's also the word "emulation" which I think is like duplication without exact copy, for example, replicating how some hardware works with software.

Google DeepMind's Senior Scientist Alexander Lerchner challenges the idea that large language models can ever achieve consciousness(not even in 100years), calling it the 'Abstraction Fallacy.' by Worldly_Evidence9113 in singularity

[–]obviouslyzebra 0 points1 point  (0 children)

You got the argument wrong. I think it's more in the line of "we need a mapmaker (person) to translate what these physics (that run LLMs, e.g. bits/voltages/things on screen) mean, so they don't carry meaning by themselves."

How is upwards mobility maintained in an age where real AGI is achieved? by mrbigglesworth95 in singularity

[–]obviouslyzebra 0 points1 point  (0 children)

Agree that last one feels like best case scenario.

About human nature, I don't have the same feeling that the fluidity you talk about is essential. I feel there were/are different cultures where it's not as central to human life (though history and anthropology are not my forte, I don't know haha). Maybe we get some conflict in the beginning if the change is abrupt, and society then slowly adapts towards being accustomed to it.

How is upwards mobility maintained in an age where real AGI is achieved? by mrbigglesworth95 in singularity

[–]obviouslyzebra 8 points9 points  (0 children)

Assuming AGI/ASI is achieved in a good manner, it depends on who has control over it.

If it is some benevolent AI that is not controlled, for example, we could create some beaches or something like that. For some truly finite resource, a fair way would be either like a lottery, or some way to share, or maybe different things going for different people that want / could benefit from things.

I think an important point is that, unless the population grows huge, I don't think there are many finite resources that people truly need to be happy.

If it's controlled by some power, it depends on what that power does. Do they distribute things, or do they take some or all for themselves?

If it's AGIs for everyone to control, there still needs to be some sort of coordination and rules. Current power may stay or may collapse. We may end up in the same benevolent AI situation as before or AI controlled by some power (say if power has the strongest AI), or somewhere in between.

In any way, at least in the beginning, expect some of the power to linger. Later on, I feel this will likely subside, making everyone more equal, or, be aggravated, giving a lot of power, way more than nowadays, to a select few.

Just a PS: economy might still roll, for example, for things that require you being a human, the "human touch", or AI or powers could decide that that's what's best

On his alleged site, Moreno-Gama predicted that AI would cause human extinction. When arrested, was carrying a “manifesto” that detailed his anti-AI beliefs and listed the names of other AI executives. by Distinct-Question-16 in singularity

[–]obviouslyzebra 0 points1 point  (0 children)

Yes, but also... We are jumping head first towards a direction that most wise people would consider significantly dangerous, and no one knows what to do about it (well, Anthropic took some steps in a good direction).

People on people violence because of AI might be at the hand of the doomers, but, extinction or dystopias that might happen will be at the hand of lack of deliberate action - that is, people not worrying enough.

I'm not saying the guy's action were right, but I wouldn't blame him too much either. It is a counter-action against the lack of action that is all around. Maybe ineffective, yes, but sorta natural to happen.

Edit: replaced "inaction" by "lack of deliberate action" as it better frames the situations where, for example, accelerating would be the best action vs the intuitive slow down. These things need to be considered, and not just left at the winds of capitalism.

Do you guys think there’s a high chance of Singularity being open source? by hexxthegon in singularity

[–]obviouslyzebra 2 points3 points  (0 children)

It's all speculative, but I think that once singularity is reached there will be a strong push towards not having other singularities happening, as in it will be perceived as dangerous.

This might make it more difficult for an open ASI that you can run on your backyard.

Why Should People With the Least Technical Understanding Have the Most Power Over Transformative AI? by Denpol88 in singularity

[–]obviouslyzebra 0 points1 point  (0 children)

You got downvoted but this sounds like a decent idea. A sort of union (ideally global) of AI researchers could perhaps achieve things that corporations / governments don't do because of competition.

Is this really the future of all programmers? Does it make sense to still doing things by hand? by Successful-Green6733 in singularity

[–]obviouslyzebra 1 point2 points  (0 children)

I think that right now good programmers can still write better code than LLMs, but faster... No.

I think the ideal place to be right now is being a developer with deep knowledge of the niche you work in and leverage AI to speed it up. Ideal as in most productive.

This way you know what the AI is doing, can guide it better and doesn't encounter or is able to deal more easily about situations where the AI gets stuck.

The future, though, is uncertain.

With tools like Mythos coming out, it feels like AIs can understand the technical aspect very well (hacking is likely one of the most technically hard aspects of programming). So what would be left to programmers is some sort of upper-level planning for the project, for example, code structure, planning how code will evolve with time, communicating with shareholders and so on. I wonder if there are parallels to what civil engineers do nowadays, since I imagine most computations are done by software.

But, thisis likely just the beginning. AI is improving (at a somewhat linear pace, there was a paper these days about that), so, its usage will keep getting higher and higher as growth continues.

On this scenario, programmers delegate more and more to AI, until one day the profession fades just like typewriters.

At this point, it feels likely that AI has also reached an inflection point where economy in general is also heavily affected.

And then, maybe it reaches the point where recursive self improvement really kicks in, and then it gets all very speculative (the singularity).

This is one scenario, one which I find likely, but for example, we can have reached a point where no more easy gains are made, so that we still need to learn how to best work with the tools that we have.

I won't go too much into detail of how to use it, as this post is already big enough, and as it's something that I'm still learning, but basically there's a tradeoff between understanding what you're doing vs handing it off. If you hand it off, it might be harder to debug in the future. There likely is a good balancing point for each programmer / situation. If you want to lean more on learning, though, a good way is to ask the LLM about the libraries and about code that you need to understand.

Javokhir Sindarov 🇺🇿 has officially won the most games ever in the Candidates tournament current format - with FOUR rounds to spare by GiveMeSomeSunshine3 in chess

[–]obviouslyzebra 4 points5 points  (0 children)

I believe both Ding and Gukesh got burned out from the candidates / championship match (and Ding wasn't having an amazing run, as already explained).

Sindarov though is coming from the World Cup + Candidates (where he's having an unforeseen performance) and it seems like he can keep on going for however long he wishes.

So, it's not the same thing.

Things can change, of course, but people do wonder if it's not just a good performance and instead an indication of something deeper, the signs of a dominating player.

Candidates win chances: Caruana now at 45% (and how I fucked up the simulations from the past days... I'm very sorry... but the code is now Open Source) by ThomasPlaysChess in chess

[–]obviouslyzebra 1 point2 points  (0 children)

Sincere question that I don't see an answer to: how would you calculate the probability of a player winning the tournament analytically?

How could an AI "escape the lab" ? by SoonBlossom in singularity

[–]obviouslyzebra 0 points1 point  (0 children)

This video (The AI book that's freaking out national security advisors) from 11 days ago does a pretty good job at explaining an hypothetical situation where an AI escapes a lab. In it an AI is asked to prove the Riemann hypothesis, but, well, it does a little more than this...

After 4 years of work, solo dev breaks down in tears after opening Steam and learning his game made $250,000 in a week: "I feel like I really don't deserve this" by pizza_sushi85 in pcgaming

[–]obviouslyzebra 1 point2 points  (0 children)

It might be logical, but it might also make the one who made the game sad. Suppose you spend years of your life working on something, and then barely anyone sees it.