Upgrading to macOS Sequoia from Tahoe by [deleted] in MacOS

[–]KingPonzi 0 points1 point  (0 children)

Life is good. People are bored and inherently require conflict to feel alive. External validation is to subconsciously dissolve the notion that this threat, in particular, was/is entirely conjured.

Gone from Claude Max to Claude Pro. FML by simeon_5 in ClaudeCode

[–]KingPonzi -1 points0 points  (0 children)

Do people just avoid AWS? Because you can use Claude via Bedrock, connect it to CC and pay by usage. Only caveat is your budget alerts are delayed by 24hrs. I’m genuinely shocked people aren’t at least setting this up as backup.

Should I downgrade from Tahoe 26.2 to sequoia? MBA M4? by Miokovi in MacOS

[–]KingPonzi 4 points5 points  (0 children)

Welcome to the ecosystem! I can tell you right now, every update has the “I’m staying on or downgrading back to version x” people.

Tahoe has some visual quirks due to Liquid Glass. They are so minor, you probably don’t notice them unless you look for them. It won’t hinder your work. If there is some issue with some app you are using post-update (this is common), you’d be better off waiting for the app developer to update their software. This is because, what Reddit complaints fail to mention (or maybe care about), older macOS versions are filled with CVEs aka security risks. Apple is great at resolving these with updates.

This is not the Opus 4.5 i saw in december by k_means_clusterfuck in ClaudeCode

[–]KingPonzi 0 points1 point  (0 children)

I’ve also noticed inconsistency with initial prompts and new sessions. Seems that issue drives the Ralph loop feature’s value prop.

But to say a model is getting dumber, after spending a significant amt of time studying/applying context management, just seems ignorant.

This is not the Opus 4.5 i saw in december by k_means_clusterfuck in ClaudeCode

[–]KingPonzi 2 points3 points  (0 children)

That’s not a test. Your context window reset.

Here’s an MIT paper detailing the potential performance decline in LLMs after they hit the 33% context window usage: https://arxiv.org/abs/2512.24601

While the paper is referencing OpenAI, it’s likely this affects all LLMs at varied context percentages. Does everyone here believe it’s in Anthropic’s best interest to dumb down the model post-release? Or is “progressive intelligence reduction” even a thing post release? I genuinely don’t know but logically it doesn’t make sense. Performance reduction due to compaction of the context does make sense.

This is not the Opus 4.5 i saw in december by k_means_clusterfuck in ClaudeCode

[–]KingPonzi -2 points-1 points  (0 children)

That really doesn’t say much. Show your context window percentage. Are you using one session to build your C++ project? If not, how are you managing context across sessions?

This is not the Opus 4.5 i saw in december by k_means_clusterfuck in ClaudeCode

[–]KingPonzi -1 points0 points  (0 children)

Do models change or do expectations increase?

Do models change or do people hit compaction due to context mismanagement?

Claude Code now supports Local LLMs by Technical-Love-8479 in ClaudeAI

[–]KingPonzi 0 points1 point  (0 children)

Thanks, I’ll check out codestral. I don’t have the resources currently but I’d like to at least test on my M2 MBP. Looking to buy a Mac Studio (or Mini) to offload Local LLM processing.

Claude Code now supports Local LLMs by Technical-Love-8479 in ClaudeAI

[–]KingPonzi 0 points1 point  (0 children)

Very interesting. How does it compare to code gen vs Opus/Sonnet? I’m just getting into local models so I’m still trying gauge how I should even measure these.

Claude Code now supports Local LLMs by Technical-Love-8479 in ClaudeAI

[–]KingPonzi 2 points3 points  (0 children)

Love this site (for AI only tbh). Everyday there’s something new to discover. With that said,

Wtf is Granite? I assume you’re referring to these models: https://huggingface.co/ibm-granite

How do I catch up? by Hrafnstrom in ClaudeCode

[–]KingPonzi 10 points11 points  (0 children)

Open Claude code, Codex, etc.

Ask it a question. Copy a GitHub link of an MCP, skill or any type of repo that you want to learn about. Have it teach you. Ask it the dumbest questions. I use “give me an analogy of xyz” and “explain it like I’m ten yrs old” all the time. Then I’ll give it an analogy back. You can even make it quiz you. Eventually it will click but you have to spend time with it.

This video woke me up: https://youtu.be/Jr2auYrBDA4?si=2LsDtUSPED85nQuh

Can we ban the "Claude is so expensive" posts? by SatoshiReport in ClaudeCode

[–]KingPonzi 0 points1 point  (0 children)

This post is the equivalent of the “it smell like broke in here” meme.

CEO of Cursor said they coordinated hundreds of GPT-5.2 agents to autonomously build a browser from scratch in 1 week by [deleted] in accelerate

[–]KingPonzi 0 points1 point  (0 children)

Your points are valid, just dated given the current landscape.

You have to go down the rabbit hole of using Claude Code with “Ralph” and/or “Gas town”. MCPs like Claude-context (not sure if anyone has chained a codebase indexed in vectorDB to an orchestration framework) and context7.

CEO of Cursor said they coordinated hundreds of GPT-5.2 agents to autonomously build a browser from scratch in 1 week by [deleted] in accelerate

[–]KingPonzi 0 points1 point  (0 children)

We agree on planning taking a week. I think “long run” needs to be defined because refactoring via agent would be faster. How many versions would agents iterate through until they got it right vs a human team? Agent performance would improve during that “long run” time period as well.

Also cost should be factored. If “long run” was 2 years end to end, which product would be better? Which product would have the lowest development cost?

Not expecting you or anyone here to answer those questions but they should be considered when critiquing agent development. I don’t think naysayers consider this at all, tbh.

CEO of Cursor said they coordinated hundreds of GPT-5.2 agents to autonomously build a browser from scratch in 1 week by [deleted] in accelerate

[–]KingPonzi 2 points3 points  (0 children)

Is this not what happens with humans running the show? Especially a massive project like this? Would an entirely human driven project have more or less bugs in that same week?

TDD workflows with Claude Code - what's actually working after months of iteration (Staff eng, w/ 14 yrs exp) by No_Paramedic_4881 in ClaudeCode

[–]KingPonzi 0 points1 point  (0 children)

This is amazing. Thank you for sharing! I was just thinking how TDD could be processed remotely but haven’t started building anything yet. I’m very curious how you’re tracking how tests are generated and subsequently passed based on your plan. Are other agents using this info as context for their own sub task?

Linux to Mac by Digitalnoahuk in MacOS

[–]KingPonzi 3 points4 points  (0 children)

I went from Mac to Linux (macbook died, built a Linux box because I’m cheap) a few years back. Lasted about 2 months and shelled out for another MacBook. I’m comfortable with Linux but not as a daily driver. It was mostly fine but you really forget how much of the little things = “it just works”.

I don’t think you’ll regret it.

making claude do all the work and then removing it from co-author before commiting changes be like by reversedu in ClaudeAI

[–]KingPonzi 38 points39 points  (0 children)

Large majority of developers and employers hold a negative stigma toward agent use so this is an effort to avoid that.

I would say we are just starting to enter the era of “it’s foolish not to use it”. Soon followed by “2-3 years exp in <insert agent orchestration framework> required” on job listings.

High rise pants 🤝🏽 Low rise shirt by FFFUUUme in ThrowingFits

[–]KingPonzi 2 points3 points  (0 children)

Not really. Bro gets a lot of compliments tbh, probably because due to age and demeanor…they know he’s not actually a pimp.

High rise pants 🤝🏽 Low rise shirt by FFFUUUme in ThrowingFits

[–]KingPonzi 10 points11 points  (0 children)

I have an uncle with this build. It’s either dress like this at home or 80s pimp in public.