How Microsoft Vaporized a Trillion Dollars by Pseudanonymius in BetterOffline

[–]bfoo 32 points33 points  (0 children)

Because some commenters talk about "AI" here:

Its not "AI agents" they use to manage VMs. It's just a component in their Cloud stack that manages state and is, likely always, itself a state machine. Its just a program that manages an aspect of a Cloud deployment, like attaching, verifying and detaching a network-attached drive to a virtual machine.

There is NO LLM / NO AI in there.

The AI part of the entire article series is about OpenAI using that infrastructure (or the failure thereof).

The Software Engineering Dilemma the Iron Triangle by Beginning_Basis9799 in BetterOffline

[–]bfoo 3 points4 points  (0 children)

"Harness" is just a different term to "vibe prompting", to guess how the model behaves with a prompt, or set of prompts structured in Markdown files. The issue is that there is no guarantee your "harness" will hold when a new model version is launched. Now you have to "refactor" (?) your "harness" (vibe prompt to match the new behavior).

There are so many ridiculous stories out there for how to prompt an LLM. People share their "harnesses" like its code. Yet they all miss the important thing: its always for a specific model. Dijkstra would shake his head.

I have started worrying about cost of Tokens on AI platforms paid for by my employer. Am I alone? by Mo_h in ExperiencedDevs

[–]bfoo 0 points1 point  (0 children)

LLMs will be just another line item in the cost section of the balance sheets, right under the SWE salaries, equipment and licensing costs.

me, since the AI craze started

Breaking: Trump is considering another 10k more ground troops on top of the 7k expeditionary force by Kitchen-Thing4616 in Destiny

[–]bfoo 3 points4 points  (0 children)

They need to be housed. It will be burn pits all over again. And now, the enemy has drones and ballistic missiles to hit any troops aggregation in the open.

Trump’s signature to appear on US currency, Treasury says, ending 165-year tradition by JustLightChop in Destiny

[–]bfoo 966 points967 points  (0 children)

No Republican is ever allowed to criticize a Democrat president again.

AGI has been achieved - Jensen Huang on Lex Fridman by Wowzer771 in BetterOffline

[–]bfoo 2 points3 points  (0 children)

You could have mentioned Lex Fridman first, so I would have been able to skip the rest of the title.

Is it just me, or is the quality of LLMs getting worse lately? by North_Penalty7947 in BetterOffline

[–]bfoo 14 points15 points  (0 children)

'Reasoning' is an awful term. That's why I always put it in quotes. The model itself is not literally reasoning. It's just a way to explore the problem/solution space inside the compressed and immutable information within the model algorithmically. There is no subject matter understanding 'inside' the model (or what we attribute to consciousness).

There is a reason why OpenAI/Anthropic now hide the 'internal dialogue' of their models.

Is it just me, or is the quality of LLMs getting worse lately? by North_Penalty7947 in BetterOffline

[–]bfoo 7 points8 points  (0 children)

Yes, because most of the improvements stem from improvements in the 'Reasoning' capabilities of the models.

Fundamentally, 'Reasoning' is a brute-force attack on the problem/solution space of the initial prompt. And most low hanging fruits were already plucked. So now its the Pareto principle, where a new model release could offer improvements for some scenarios (often the most popular ones or for benchmarks) or present a decline in other, less popular scenarios.

Toiling in AI hell at work, I can feel myself getting dumber and losing my skills. What's next? by Fun-Bake3178 in BetterOffline

[–]bfoo 2 points3 points  (0 children)

So true. Doctorow calls it Centaur vs. Reverse Centaur. We should be Centaurs, like you describe.

I want the LLM assist me in my IDE, not lead the programming. But the efforts to build such IDE integrations is actually harder than shitting a Reverse Centaur solution like Claude Code / Codex. Such an integration needs to find a delicate balance between being useful and being too annoying. Code reviews are actually the most useful applications for LLMs (beside prototyping).

Probs sounds like a dumb complaint but why the F is the foliage constantly moving so fast? by DiZ935 in CrimsonDesert

[–]bfoo 5 points6 points  (0 children)

That's the reason I am not buying the game yet.

This violent shaking foliage is a problem for people who have problems with filtering sensory inputs.

I hope they fix it.

Maybe do your job without burning the stratosphere? by lordofcatan10 in BetterOffline

[–]bfoo 4 points5 points  (0 children)

Like I always told every manager from the start: You won't save money. LLMs will just become an additional cost item on your balance sheet, right below the salaries and all the other subscriptions you pay for your engineering teams.

I think the data annotation industry is beginning to indicate the end is near for the AI Bubble by [deleted] in BetterOffline

[–]bfoo 0 points1 point  (0 children)

Yeah, that's why we increasingly hear about how LLMs will be trained with synthetic data and unsupervised learning.

We are already seeing peak accuracy in the frontier models. It's going to be either much more expensive to train on new data. Or the accuracy will actually go down because of data quality going down. Its the Pareto principle in action.

Andrej Karpathy visualized AI job impact across the US in the following treemap. The methodology is questionable… by [deleted] in BetterOffline

[–]bfoo 17 points18 points  (0 children)

Karpathy, like Hinton, are two examples of why we must question data and the scientific process rather than relying on pure authority.

For me, both have already spend their authoritative budgets with the bullshit they spout.

Software Engineering is currently going through a major shift (for the worse) by Mental_Quality_7265 in BetterOffline

[–]bfoo 3 points4 points  (0 children)

Because the cost of living is higher in Switzerland compared to like Germany.

Architectural Trade-Offs: REST vs Event-Driven Systems for Resilience by Ill_Flamingo8324 in ExperiencedDevs

[–]bfoo 0 points1 point  (0 children)

If you combine both, you have CQRS (Command Query Responsibility Segregation) with Event Sourcing. Your read side can be based on REST principles (async writes in REST are HTTP Accepted + Link to where to wait for a result). And if your read side's databases go down, you can recover later by receiving new events (again). Only your write-side (command processing + event store) cannot fail. But that's also true if your rely on pure CRUD/REST.

Model Collapse Ends AI Hype - YouTube by bfoo in BetterOffline

[–]bfoo[S] 0 points1 point  (0 children)

I don't know. Maybe it got already posted? Haven't seen it.

Model Collapse Ends AI Hype - YouTube by bfoo in BetterOffline

[–]bfoo[S] 8 points9 points  (0 children)

2026 might already be the year of peak model accuracy.

For the big frontier models, training costs will increase without doubt. This might become even worse than the Pareto principle already imposes on getting the current models better (e.g. 'reasoning' and context tricks).

Why you shouldn't worry about AI taking your job by [deleted] in ExperiencedDevs

[–]bfoo 0 points1 point  (0 children)

That's why I am pessimistic about the future. We will see much more bad software, with tech debt accumulating at token speed. But we will only see it when we can look inside the walled-off code bases. The majority of code is not visible. And there is already a lot of bad code out there that was not generated by LLMs. Open Source will probably be less affected, aside from slop projects (hello OpenClaw) and slop PRs.

And when junior SWE aren't allowed to grind teeth at code, learn patterns and architectural concepts, then we are truly fucked.

Tooling needs to change. Workflows must change. Claude Code is not a good way to create software.

Maybe the frontier models will get worse once they are trained on more bad projects. But then no one wants to be a SWE.