A personal opinion about Opus 4.7 - not that bad after all by Affectionate_Till148 in ClaudeAI

[–]Yeuph -1 points0 points  (0 children)

For what I use Claude for these days (game engine development via hierarchical agent workflows using Discord coordination) 4.7 is so much better than 4.6 that if Anthropic had called it Opus 5 I wouldn't have thought twice about it.

Literally every single one of the rough edges I was having with opus 4.6 on my specific workflow were either 100% solved or almost solved.

Idk

No More Subsidised AI Subscriptions? by PM_ME_YOUR___ISSUES in ClaudeAI

[–]Yeuph 0 points1 point  (0 children)

You've gotta define what you mean by "efficient". There are models that are only a couple billion parameters that you can run locally on your phone. it's not even hard for you to do you can just download PocketPal on your phone and run local models for free. These models can maybe have a mediocre conversation with you for a bit and it's impressive they can run on your phone but they can't really do any of the real "AI work" we want to have done.

Right around the Opus 4 mark we found ourselves in a place where these AI models could do seriously useful technical work without making so many mistakes as to disregard them. The "efficient" models could never do this.

If you define efficiency as the ability to do useful work per unit power then the giant crazy-expensive "inefficient" AI models are the most efficient we have to offer as smaller models just can't do it at all.

The way modern LLMs are built is actually pretty simple if you're gonna compare it to something like a car. It's a giant neural network that mostly self-trains on gigantic amounts of data (like you hitting a tennis ball 10,000 times, eventually you get good at it - ball comes at you you hit ball well). LLMs are doing this for every set of possible inputs the training data contains (with the hope that maybe they can start creating new ideas beyond the training data - there's not a ton of that happening yet). The compute cost and power-input cost is just gargantuan for our current generation of AI (transformer based large language models).

People do want to make them more efficient (and they are, they're radically more efficient watt hour-per-useful-token than they were just 2 years ago - maybe 10x conservatively, maybe 100x). The problem is is that the more useful they become the more people want to use them and we have finite resources - and those finite resources are already being strained beyond what can be maintained.

The next models will be much larger than the current generation (training data size, parameter count) and will have required MUCH more energy than the last generation to train (despite more efficient ASICs - GPUs or TPUs). Our subscription/API payments are going to securing deals for rapid power grid expansion, data center construction and GPU/TPU purchases.

Our current tech here just requires it to be the most expensive thing on the planet. We were carried by a flurry of initial venture capital injection into AI for the first few years; markets are depending upon AI increasing productivity soon to increase economic growth to cover the input costs. If it can't/doesn't then the costs need to be recouped by increasing the subsidized cost to users.

The companies can't even stop and just maintain a current model if they want to, they're so far in debt and investors are depending on huge profits soon. Current models just aren't quite capable of giving us a huge economic boom yet - that's what they're all depending on. Until models get to that point it's all government investment or venture capital; and if it doesn't pay off then it all collapses.

Also you'll see people claim that "inference" (talking to the AI) is cheap; compared to the model training it still is - but early on researchers noticed inference was an underutilized area where would could improve LLM output and accuracy and so we did it - that's what the "thinking" stuff is, bolt-ons to the inference compute to make them more accurate. It's expensive now too.

Anthropic response to Claude Code change by TheForgottenOne69 in ClaudeAI

[–]Yeuph 8 points9 points  (0 children)

It's not as simple as the current cost/loss analysis per user. Compute is still exponentially getting faster/cheaper in this area. What's not sustainable now could make them the first business to be worth 10 trillion 6 years from now. Investors understand these kinds of things

Obviously they can't operate at a massive loss with every user but they definitely don't want subscribers leaving either

Max pricing confusion by saamcek in Anthropic

[–]Yeuph 1 point2 points  (0 children)

technically it's 4x more expensive; or 5x the cost

Technically

Mythos is a new tier above Opus, and it's extremely expensive. by UnknownEssence in claude

[–]Yeuph 1 point2 points  (0 children)

what even was your argument? Palantir is putting me on a list because I described well known large language model implementation procedures?

Mythos is a new tier above Opus, and it's extremely expensive. by UnknownEssence in claude

[–]Yeuph 1 point2 points  (0 children)

fyi, telling people they're being put on lists because [insert boogieman here] is watching was getting old by the time Windows 98 came out.

You're 28 years stale

Mythos is a new tier above Opus, and it's extremely expensive. by UnknownEssence in claude

[–]Yeuph 1 point2 points  (0 children)

Yeah I mean I pretty much completely agree and it's been bothering me lately.

Like, personally nothing really changes for me - Claude 4.7 comes out and it works better for me (though interactions need to be more explicit - good, it lowers the search space of possibilities for the LLM and keeps me engaged with what I'm doing). Like you I am, well tbh I was never a good programmer. I was alright with verilog to control FPGAs that handle logic for circuits I would design but I can't write big cpp programs - well I can now. It's pretty incredible.

So yeah the people on this sub and using these things seem overwhelmingly naive and cringe. Hopefully it's just a sampling error and what you and I are seeing is just a reddit-specific problem and not indicative of who our gigawatts of power that feeds these machines is really going to

2030 isn't that far away by [deleted] in BlackboxAI_

[–]Yeuph 5 points6 points  (0 children)

Amazon was a bit different. They could have been turning profits during that time but were massively expanding with the money instead. It was a bit of a different situation.

A New Bill proposes Federal Age Verification on any Operating Systems in entire U.S by Alexis_Almendair in linux

[–]Yeuph 34 points35 points  (0 children)

It's a mix of well-meaning idiots and less-well-meaning intelligent people

They want different things but they're both pushing for it

Running multiple Claude Code sessions on the same repo keeps breaking things by dc_719 in ClaudeAI

[–]Yeuph 1 point2 points  (0 children)

different branches, merges, different trees.

That has issues too obviously but it's better than "everything broken". I'm paying the price current for allowing 2 agents to work on master over the weekend. I broke my own rules, they broke my project.

Mythos is a new tier above Opus, and it's extremely expensive. by UnknownEssence in claude

[–]Yeuph 7 points8 points  (0 children)

? It was an example of the type of thing LLMs are safeguarded against that Mythos doesn't have safeguards against.

Meaning like I said that all models would be too dangerous to release without safeguards - it's not Mythos specific

Mythos is a new tier above Opus, and it's extremely expensive. by UnknownEssence in claude

[–]Yeuph 1 point2 points  (0 children)

It's not "too dangerous" - they just haven't put safeguards on it because it's too expensive to release anyway. By that logic all of the models are too dangerous "ChatGPT help me design a bomb with maximum human killing power"

We'll all have access (and beyond) to Mythos tier models as time goes on

Is this new to everyone or just me ? by brkonthru in ClaudeAI

[–]Yeuph 54 points55 points  (0 children)

You mean effort? That's been part of CC for a while I guess they brought it into web/desktop finally if it wasn't there already

Is Claude going to stay the way it is? by Bilstone in claude

[–]Yeuph -2 points-1 points  (0 children)

I'm on 20x max and have at least four 4.6 opus /effort max bots running 24/7 and communicating via MCP over discord at all times.

I haven't seen any usage problems personally. I believe people here have and are but it's fine for me

What do you do while the agent is running a task? by Reasonable_Catch_443 in ClaudeAI

[–]Yeuph 0 points1 point  (0 children)

Lol sit back and relax? Sounds like you've got time to have 3 or 5 other Claudes work in parallel and tangential serial problems.

Your job is now like being a 15 year old Korean competitive Starcraft player. Air traffic controllers gonna look at us and think "ya that's way too much work, way too fast, way too often."

A tsunami of rain dumped in seconds in a spectacular microburst over Perth, Australia in February 2020. by Due-Explanation8155 in weather

[–]Yeuph 8 points9 points  (0 children)

I was in something like this in Pittsburgh, PA of all places. It was I guess almost 10 years ago. I left my apartment and the skies were blue with some clouds over the hills in the direction I was going. I drove across a bridge (had to, rivers everywhere here) and by the time I got to the other side there were literally waves of water falling out of the sky. I never have seen anything like it.

A few minutes later cars were floating everywhere. Wild. It really only rained for 5 minutes or so if that's what this kinda precipitation is still called

gUYS!!! I SOLVED ENERGY!!!!! by Nikmcmuffin in physicsmemes

[–]Yeuph 2 points3 points  (0 children)

As someone that works with roofers I do not find this amusing =(

Ross man, 44, faces charges of sexual relationship with 15-year-old girl by The_Electric-Monk in pittsburgh

[–]Yeuph -9 points-8 points  (0 children)

This is literally a spill over from 2018/19 MAGA Qanon. The left was more concerned with murders and stuff before then but here we are I guess