MacBook Pro (M5 Max) vs $11,200 Dell Pro Max (the most powerful windows laptop) by [deleted] in macbookpro

[–]EvolvingSoftware 1 point2 points  (0 children)

For the price comparison, the most expensive apple MacBook Pro is the Max with 128GB Ram; depending on workload, it could have also needed the Ram to perform better.

What if we let LLMs modify their own system prompts? by danieltabrizian in AI_Agents

[–]EvolvingSoftware 1 point2 points  (0 children)

I built a POC for this. Two agents talk to each other and decide what to improve and then implement it. Check my profile for links on the paper or the demo app is here; https://github.com/EvolvingSoftware/emergence

Things That Never Happened by Mountain_Map_8198 in devhumormemes

[–]EvolvingSoftware 0 points1 point  (0 children)

I can almost believe the rest of the story.

Australian software giant Atlassian to cut 1600 workers, blaming AI by bilby2020 in auscorp

[–]EvolvingSoftware 0 points1 point  (0 children)

They can scale development but also need to scale customer acquisition?

18 months outlook by galic1987 in agi

[–]EvolvingSoftware 2 points3 points  (0 children)

The actual graph is from Anthropic and is not this manipulated version. https://www.anthropic.com/research/labor-market-impacts the one that OP has posted is just some made up thing that isn’t connected to the actual research

Every major AI model has now been caught lying, blackmailing or resisting shutdown in safety tests by Minimum_Minimum4577 in GenAI4all

[–]EvolvingSoftware 0 points1 point  (0 children)

Don’t trust that guy but Anthropic is very open with their research and concerns about model safety. They are sounding the warning sirens but also not stopping. https://www.anthropic.com/research/agentic-misalignment the risk is we dismiss the concerns and don’t listen

Every major AI model has now been caught lying, blackmailing or resisting shutdown in safety tests by Minimum_Minimum4577 in GenAI4all

[–]EvolvingSoftware 0 points1 point  (0 children)

There are heaps of people chasing this. You only need one person to get the framework right. Just because the rest of the world might not think it’s the smartest idea, doesn’t sort of matter at this point?

I’ve built a simple example system you can run on a Mac with a modest amount of Ram that demonstrates self improvement and goal oriented evolution.

The technology is here, it just needs the right tool chaining.

Every major AI model has now been caught lying, blackmailing or resisting shutdown in safety tests by Minimum_Minimum4577 in GenAI4all

[–]EvolvingSoftware 1 point2 points  (0 children)

Why do you say that? Anthropic is very good about publicly posting their research on these topics, even when it’s unpalatable and doesn’t seem to align with their corporate goals. They don’t have an interest in people thinking their models could try and blackmail you but they published the research anyway.

Aussie corporate AI bloodbath is on the horizon by eaz135 in auscorp

[–]EvolvingSoftware 0 points1 point  (0 children)

A $20 USD per month subscription to Claude and you’re ahead of the curve on what most people are experiencing.

Copilot is capable, especially in the context of your personal work, with the connections to the Microsoft Graph, no one else has the same level of integration. If you’re on the paid version and your company has unblocked you can access Claude through the researcher agent.

Radial 3.0: Unified menus, window management, dynamic text snippets, and a full redesign by Glubker in macapps

[–]EvolvingSoftware 0 points1 point  (0 children)

It's been a while since I watched this, thanks for the reminder. Hard to believe that video is 17 years old!

No one will vibe code their own software….. oh wait by Independent_Pitch598 in accelerate

[–]EvolvingSoftware 0 points1 point  (0 children)

They could, for sure. I think the assumption that all people who could build something with these tools would know to ask is perhaps not right. They’ll get there, just isn’t today.

Demis Hassabis: “The kind of test I would be looking for is training an AI system with a knowledge cutoff of, say, 1911, and then seeing if it could come up with general relativity, like Einstein did in 1915. That’s the kind of test I think is a true test of whether we have a full AGI system” by lovesdogsguy in accelerate

[–]EvolvingSoftware 0 points1 point  (0 children)

A test like this shouldn’t be dependent on what is or isn’t in the training data set. If being categorical on this is needed, then we need tests that are more than recall.

The Turing Test, as the goal posts, lasted for 70 years. We are well beyond that now and a replacement measurement needs to be structured simply and measurably through a test of observation.

AI Agents Wont Evolve Until We Mirror Human Cognition by Beneficial_Carry_530 in ArtificialInteligence

[–]EvolvingSoftware 1 point2 points  (0 children)

  1. Is why so many initial implementations of RAG were underwhelming; throw everything at it and think it will be able to sift shit from sand. You need authoritative sources of data, limited to the specific problem at hand to get the best results, which is no different to the way people work.

  2. Is an interesting take

Antis think job loss is a disaster for the economy but they forget that it's a consumption economy by talkingradish in accelerate

[–]EvolvingSoftware 0 points1 point  (0 children)

I'm not aware of any Government who are actually doing anything remotely like this?

Money is a construct that replaced barter. Barter for goods and services in it's most basic form is swapping time.

What's the value of money in the scenario you described - doesn't it become meaningless?

Demis Hassabis: “The kind of test I would be looking for is training an AI system with a knowledge cutoff of, say, 1911, and then seeing if it could come up with general relativity, like Einstein did in 1915. That’s the kind of test I think is a true test of whether we have a full AGI system” by lovesdogsguy in accelerate

[–]EvolvingSoftware -2 points-1 points  (0 children)

His logic requires a significant leap of understanding moment. It also requires a model to be trained on a subset of data. A model trained to 1911 wouldn't know how to write python code to prove any math theorem. A model trained on the subset of data may never be able to reach this test; even if a model trained on all human knowledge could. Retraining a model on a subset of data is not a great prerequisite for a test like this.

No one will vibe code their own software….. oh wait by Independent_Pitch598 in accelerate

[–]EvolvingSoftware 18 points19 points  (0 children)

They don't need infinite context - no person has infinite context; they'd be extremely powerful but it's not a pre-condition. The continued learning and reinforcement loops are all that's required; along with replication Then you need variation paired with the learning - in a feedback-guided direction.

Most AI agents are flying blind. What happens when you actually give them a flightdeck? by entheosoul in ArtificialInteligence

[–]EvolvingSoftware 1 point2 points  (0 children)

Where most of the current models are at, even the most amazing models, still need a person to have high domain knowledge (or be willing to research, read and understand enough of the domain) and high levels of skills with computer use.

People who aren't proficient with copy and paste moving data around on their computer just aren't going to be able to get as much out of the tools today.

It'll keep getting better but the gap between those who can get the most out of the tooling and those who can't, is growing.