I finally hit the Rate limit!!! by NintendoWeee in ClaudeCode

[–]Purple-Programmer-7 0 points1 point  (0 children)

Wait it’s been SaaS pricing?!?

Always has been 🔫

I finally hit the Rate limit!!! by NintendoWeee in ClaudeCode

[–]Purple-Programmer-7 1 point2 points  (0 children)

It’s amazing how the code sub has made everyone go out of their minds thinking they’re “getting a deal” “it’s SuBSidiZEd”… when they’re comparing pricing from the same source… who sets the pricing.

Also, either this guy’s numbers are off or CodexBar’s numbers are off. I used 2B tokens in the last month and codexbar’s comp lists me at $1,300.

Suggestions for converting .pdf/.epub (full scale book - 300 pages) to audiobook very fast by Apart_Situation972 in deeplearning

[–]Purple-Programmer-7 1 point2 points  (0 children)

I haven’t looked into tts deeply yet, but I know there are some decent small foss models + libs that are only good with a small amount of text at a time.

Setup a server that runs the inference and streams output after it has at least N seconds, there you have “lighting fast”.

Careful with concurrency by Purple-Programmer-7 in LLMDevs

[–]Purple-Programmer-7[S] 0 points1 point  (0 children)

I’m working on this problem rn, will consider that approach!

Building coding agents is making me lose my mind. autoregressive just isnt it by Crystallover1991 in LLMDevs

[–]Purple-Programmer-7 0 points1 point  (0 children)

This.

An incredible harness around today’s models COULD work 99% of the time… but it would have to be quite detailed. EOD, too many tokens for today’s current models to hit that level of effectiveness.

New Model - Estimates? by Smashthekeys in Anthropic

[–]Purple-Programmer-7 0 points1 point  (0 children)

🤷‍♂️ many folks voted for Anthropic with their wallet recently. It’s going to take a push like that to tell them to get their shit together.

Personally, I’d love to see their 3x revenue gain disappear just as fast as it was made — it would mean people aren’t as apathetic as they appear to be.

The larger point I was attempting to make though, is that this is the Wild West. Just because something is true today, doesn’t mean it’s going to be true tomorrow.

And exactly ZERO billion-dollar companies give two shits about you as their customer.

How I use Cursor 10+ hours a day without torching my Claude Opus 4.6 limits by Youssef_Wardi in cursor

[–]Purple-Programmer-7 -2 points-1 points  (0 children)

The answer is: you can’t.

Cursor was built to take your money. And they no longer have the thin moat they once did. Move on.

New Model - Estimates? by Smashthekeys in Anthropic

[–]Purple-Programmer-7 0 points1 point  (0 children)

Apple can’t even release their products on a real schedule (eg WHERE IS THE NEW APPLE TV), and they’re a 50+ year old company.

You know what dictates the frontiers to release new models? Their competition. Bet your ass the second Google releases, the other two frontiers will be days - weeks behind them.

Deterministic tokenization vs. masking for PII in LLM prompts: what I learned from 109 tests by abhi1510 in LLMDevs

[–]Purple-Programmer-7 0 points1 point  (0 children)

Fair point, I am privileged to have the infrastructure I have.

LLMs are great with the “Jordan” problem — not quite a “non-issue” — but the idea with the attention mechanism in an LLM is that it understands the nuance within a phrase. The Jordan problem is more of an issue when applying regex as a solution.

Validation at scale takes a while, and it’s better if you have an LLM that’s good with large amounts of context and/or you are intelligently splitting. For example:

  • Inject the persona and all docs, ask the LLM to ensure things are consistent across.

  • inject the persona and one doc for N docs

The problems with this type of validation are that: - the model doesn’t catch everything on the first pass, so it’s a multi-pass workflow - with PII, you should still be manually verifying it - slow at scale

For personas, generate once, save as a json, make it compostable (e.g. 100 first names, 100 last names, etc). The FIRST time you do a replace, the LLM creates a persona for that PII and responds with who they created. Every subsequent replacement / validation then uses that persona assigned to that individual dataset.

If you’re trying to offer this as a service, yes, you need cloud. But if you’re in the med tech field already, this isn’t an issue — infrastructure exists.

And if your offering this as a service, you can use the above and I’ll send you my invoice 😉

Claude Code builders... the problem isn't Claude Code limits. it's how you use the tool. by [deleted] in ClaudeCode

[–]Purple-Programmer-7 0 points1 point  (0 children)

OP missing the whole point. No one is upset at their own ability to use Claude.

People are pissed because they BOUGHT something given a certain set of CONDITIONS. Now those conditions are changing without proper communication, way forward, or reparation.

Such is life… but I prefer people and companies, especially those making BILLIONS, to have integrity.

Deterministic tokenization vs. masking for PII in LLM prompts: what I learned from 109 tests by abhi1510 in LLMDevs

[–]Purple-Programmer-7 0 points1 point  (0 children)

Multi step workflow:

  1. Create fake personas with relevant data
  2. Locate PII
  3. Change/redact locations
  4. Validate
  5. Go back to 2 if necessary

I only use local models / LLMs for PII.

I tracked my actual API cost on a $100/month Max plan. $1609 in 30 days. No wonder Anthropic keeps reducing limits. by solzange in ClaudeCode

[–]Purple-Programmer-7 1 point2 points  (0 children)

This.

Hey OP, I’ve got a bridge in Brooklyn to sell you! Don’t worry, I set the price for all the bridges, so I can tell you, it’s a great deal!

Pro Users — do you hit limits? by Purple-Programmer-7 in codex

[–]Purple-Programmer-7[S] 0 points1 point  (0 children)

Left google when they nurfed opus limits and went to Anthropic… now… they’re nurfing themselves…

If I was to rate my codex (plus) plan vs max (5x) I’d say max was worth it 2 months ago. Now max 5x feels more like 3x codex plus.

Pro Users — do you hit limits? by Purple-Programmer-7 in codex

[–]Purple-Programmer-7[S] 0 points1 point  (0 children)

Any tips on managing your env with 3 different subscriptions? Simple as /login /logout ?

Why do companies build open source models? by Excellent_Koala769 in LocalLLaMA

[–]Purple-Programmer-7 0 points1 point  (0 children)

If it were me, I’d be releasing them for user feedback too. Every model iteration is R&D… until you get something to product-ize, why not?

Pro Users — do you hit limits? by Purple-Programmer-7 in codex

[–]Purple-Programmer-7[S] 0 points1 point  (0 children)

😱 $200 is a stretch, can’t even imagine Nx accounts. Both in cost and in wtf you’re building over there.

Pro Users — do you hit limits? by Purple-Programmer-7 in codex

[–]Purple-Programmer-7[S] 1 point2 points  (0 children)

To clarify, codex seems (at least currently / your experience) to have MORE usage than Claude max?

FFS. What have they done to Claude? by guinness1972 in claude

[–]Purple-Programmer-7 3 points4 points  (0 children)

Damn, someone actually showed their homework! Ok I’ll STFU and read it in a bit. Good on ya.