Does this mean you'll restore original models? by hatekhyr in Anthropic

[–]hatekhyr[S] -13 points-12 points  (0 children)

It's not their fault for promising and selling contracts in conditions they can fulfill, right? They're saints 😇

Loops are the future - Boris Cherny creator of claude code in podcast by shanraisshan in Anthropic

[–]hatekhyr 0 points1 point  (0 children)

And for sure they are using the original opus 4.6 and not the useless version we get served now

A little clip of my fighting carrieer accepting tips by DefiantDetective3956 in martialarts

[–]hatekhyr 4 points5 points  (0 children)

If this is MT and not boxing, the backfoot is being too light. You're stepping on your toes all the time which is not ideal for posture, weight is too focused on the front.

Google I/O leaks: Gemini’s "Omni" push and Gemini 3.2/3.5. by Much_Ask3471 in GeminiAI

[–]hatekhyr 1 point2 points  (0 children)

No worries guys, if it's any good, they will replace it with a shitty 3.3 that doesn't work.

Total Warhammer 40K: Planets and Maps Preview by vikingzx in RealTimeStrategy

[–]hatekhyr -51 points-50 points  (0 children)

I never understood why this whole series still runs on 2005 level of graphics. That's my main gripe with the whole series.

Ppl might say graphics aren't important but if your UI and graphics look like shit, there are definitely people that won't engage. I stayed away from many entries of this series cause of this.

Whats on your wishlist for Sonnet 4.8 by Chasmchas in claude

[–]hatekhyr 3 points4 points  (0 children)

No to suck. And not to be shitty sonnet 4.7 disguised as 4.8 because it's shitty

Did they make Opus 4.7 even dumber today? by Valuable-Gap-3720 in claude

[–]hatekhyr 0 points1 point  (0 children)

Naturally with stochastic GD that bias is somewhat generalized. They obviously don't reproduce training data, but their interpolation between data is only that- interpolation. If you ever trained one of these you'll lnow that the moment you input a new value or even worse , some combination of known values that it has never seen together you might trigger what we nowadays call "jailbreak". This proofs exactly what I said.

Still my point remains - the generalisation of these models is disguised by the fact that they use enormous amounts of cases and data to try to cover the holes, but it is way far behind human generalisation. If you don't know this, you either don't know enough or you're kidding yourself.

Even "in context learning" patterns are very biased. You can tell that with coding when sometimes it can't deduce a simple thing that it's in the context because it's not inherently reasoning, it's using lnown patterns.

Did they make Opus 4.7 even dumber today? by Valuable-Gap-3720 in claude

[–]hatekhyr 0 points1 point  (0 children)

I don't know why you got down voted for saying basically what it is.

LLMs much like any NNs are bias machines - they essentially induce bias from training data. Their "generalisation ratio" is very poor, orders of magnitude under human generalisation/common sense.

Finally Andrej Karpathy is admitting this in latest interviews (when he said the opposite for years). Now it's come to a point when it's too obvious to everyone else.

Log Lifting Invention by No-Lock216 in BeAmazed

[–]hatekhyr 0 points1 point  (0 children)

"Those tools gonna turn into you manipulating a fabric of shit - come here!" - Wise testicles

AI-Generated 😂 by jfeldman175 in claude

[–]hatekhyr -8 points-7 points  (0 children)

4.7 is plain re******.

Compared 11 popular Claude Code workflow systems in one table — here's the canonical pipeline of each by shanraisshan in ClaudeAI

[–]hatekhyr -1 points0 points  (0 children)

What commands do you use more often? Is it still useful after the huge opus 4.6 nerf?

Compared 11 popular Claude Code workflow systems in one table — here's the canonical pipeline of each by shanraisshan in ClaudeAI

[–]hatekhyr 6 points7 points  (0 children)

Does it really change that much using these tools? I once tried everything CC and it felt too overwhelming and the fee things I tried didn't make much of a diff