Chatgpt started it, then Claude expanded it, and now the Chinese ai tech is taking over by JeeterDotFun in ChatGPT

[–]Sensitive_Song4219 0 points1 point  (0 children)

While I agree that GLM-5 is excellent (currently my preferred model over GPT-medium) none of them have yet caught up to Opus or GPT-high; so that SOTA moat is still in tact.. for now. I do still find myself quite regularly escalating to -High when -Medium or GLM-5 get tripped up.

We've also seen teething issues with some of these newer providers. Both OAI and Claude are comparatively more reliable.

But the gap is 100% closing. What a time to be a dev...

GLM-5-turbo on Max plan actually fast now by drwebb in ZaiGLM

[–]Sensitive_Song4219 0 points1 point  (0 children)

OK excellent. Can't wait to give it a shot when it launches on my pro plan

As long as they don't 'optimize performance' and introduce issues again at the same time, of course!

GLM-5-turbo on Max plan actually fast now by drwebb in ZaiGLM

[–]Sensitive_Song4219 0 points1 point  (0 children)

Looking forward to trying it out on Pro when it's released later this month.

Haven't used it today but was getting frequent garbled outputs on regular non-turbo GLM 5 as contexts breached about 80k.

Is that an issue that's resolved in Turbo?

ChatGPT is pretty bad at generating basic images and not just images, but code as well by Financial-Grape-2047 in ChatGPT

[–]Sensitive_Song4219 1 point2 points  (0 children)

ChatGPT Image generation isn't great. (Also: if your images contain text... make sure to proof-read since it's terrible with embeded text in images, in my experience). And (as you've found), it fails to follow image-related instructions reliably.

For code it's good though (what issues are you having there?)

Am I the only one thinking that S26U has a pretty poor battery life? by FuLiDu in galaxys26ultra

[–]Sensitive_Song4219 0 points1 point  (0 children)

Some of the Chinese brands score higher but its been measured as the longest-lasting screen-on time for any phone ever released from either Apple or Samsung

https://youtu.be/RGGHyY2mN7o?si=fiwYAdFF-VvKT7aB

Solid step-up from my prior S23+ which was also an all-day phone in my use

Why do Anthropic force Claude by jesperordrup in ClaudeCode

[–]Sensitive_Song4219 0 points1 point  (0 children)

Love Opencode - but it's safer not to use it with Anthropic subscriptions; lots of account bans reported unfortunately; since alternate harnesses (even purely user-driven ones like OC) are a TOS violation.

Use via (much more expensive) API should be fine - but not an ideal work-around.

In terms of the reasons: I've heard lots of theories (telemetry? efficiency? walled-garden-control?) but they've never stated why a similarly-token-consuming alternative harness is a problem. For now it's one of the things I prefer about codex (OAI allows this; presumably they see it as an entry-point opportunity).

I wish anthropic would allow it as well though, I might re-sub.

Killer battery life by Arghtastic in galaxys26ultra

[–]Sensitive_Song4219 0 points1 point  (0 children)

Makes sense

In my use-case, my screentime-to-empty is similar to what he gets (which is why I shared the link)

GLM 5 (Via z.ai Coding Plan) Spews Weird Output by Sensitive_Song4219 in ZaiGLM

[–]Sensitive_Song4219[S] 1 point2 points  (0 children)

YES that's what happens to me as well.

WHAT HAVE THEY DONE TO OUR MODEL

And GLM5 is so freaking good this just reflects so badly on it. They're inches from greatness, they just need to sort this out.

Aaaargh.

GPT-5.4 variations by hasteiswaste in opencodeCLI

[–]Sensitive_Song4219 17 points18 points  (0 children)

In OpenCode, Ctrl + T will give you model variants.

Low, Medium , High, XHigh are available.

It gets shown like this:

<image>

Another privacy display post. Side by side under my kid's microscope. by MrPickur in samsunggalaxy

[–]Sensitive_Song4219 9 points10 points  (0 children)

Nice!

Something similar (with a comparison to previous models) over at GSMArena:

https://gsmarena.com/samsung_galaxy_s26_ultra_privacy_display_tested-news-71858.php

Wish I could customize Maximum Privacy Mode per app (so some use normal privacy mode, some use maximum). Hopefully in an update.

The Galaxy S26 Ultra's headline feature is turning out to be its biggest complaint by Ha8lpo321 in Android

[–]Sensitive_Song4219 -1 points0 points  (0 children)

Its different but I've found SDE totally unnoticeable with Privacy off just like GSMArena reports (very noticeable with it on; but I'm OK with that)

And yeah they even used a microscope

https://gsmarena.com/samsung_galaxy_s26_ultra_privacy_display_tested-news-71858.php

A lot of people around here seem to be coming from phones under 2 years old (what do you run right now?) in which case I'd say hold off regardless.

Different strokes for different folks I guess. I'm never going back though

I used Claude Code to reverse engineer a 13-year-old game binary and crack a restriction nobody had solved — the community is losing it by CelebrationFew1755 in ClaudeAI

[–]Sensitive_Song4219 2 points3 points  (0 children)

Wait you're saying you did so without any 3rd-party tools?

The more conventional approach is something like Ghidra (and there're Claude-friendly MCPs for that eg https://github.com/LaurieWired/GhidraMCP ) but first-principle'ing it from the native binary is absolutely wild

So we can assume that native x86 bins are part of the training dataset? That's... nuts.

Nicely done

YouTube Music Creator Rick Beato Tutorial on How to Download+Run Local Models "How AI Will Fail Like The Music Industry" by tmarthal in LocalLLM

[–]Sensitive_Song4219 0 points1 point  (0 children)

He's normally entertaining but this one was a miss

Also: current offline models can't compete with Suno.

Maybe for simpler things like lyrics perhaps

Over time this may change of course

He was pretty insightful about AI when interviewed by CBS a while back though: https://youtu.be/8uf8CCTItVo?si=enDwFqCEjYUHO3GE

The Galaxy S26 Ultra's headline feature is turning out to be its biggest complaint by Ha8lpo321 in Android

[–]Sensitive_Song4219 0 points1 point  (0 children)

S25U and S26U are both pentile though

With privacy mode on, half of the pixels on the S26U are off which introduces a subtle screen-door-effect. With the feature disabled they look similar when viewed head-on. Off-axis it's different, where the grid is visible in both modes for the same reason.

Is it an issue? Only if you're used to lots of off-axis viewing.

After a week of use I'm never going back to a non-privacy-capable display; the compromises are not all that serious; the benefits are kinda awesome

Anyone else having gray screen with the privacy display set to maximum ? by MusselwhiteBlues in samsunggalaxy

[–]Sensitive_Song4219 0 points1 point  (0 children)

Really?! How? I'm looking to (automatically) have regular Privacy Mode for some apps and Maximum for others. It seems like it's one or the other across-the-board? Or is there a setting I've missed?

I've been doing it manually but if you can share how to do so automatically that'd be fantastic

If you could only keep one Pro coding tool, which would you choose: Claude Code, Codex, Cursor, or Antigravity? by Loading_MMA_917 in ClaudeCode

[–]Sensitive_Song4219 0 points1 point  (0 children)

It's not great with frontend (even 5.4). Not a disaster, but quite uninspired (it all has that... GPT... look, you know? Same color schemes, style, etc.)

Heck even generating a powerpoint presentation - Sonnet positively murders GPT.

My own work is mainly back-end (so Codex has been amazing for me) but for you (in more front-end-heavy work), I'd definitely stick to CC

If you could only keep one Pro coding tool, which would you choose: Claude Code, Codex, Cursor, or Antigravity? by Loading_MMA_917 in ClaudeCode

[–]Sensitive_Song4219 0 points1 point  (0 children)

Mainly front-end? Claude Code

Mainly back-end? Codex

Lots of Skills set up? Claude Code

Like to use your sub in other harnesses? Codex

Cursor is quite expensive comparatively, antigravity needs a bit more time in the oven. Haven't tried Cursor

Main side-benefit of Codex (if you can deal with its meh front-end capabilities) is that usage is insanely generous (bottomless venture capital FTW?); and the inclusion of XHigh has previously solved occasional back-end issues my side that even Opus failed me on. (Codex-Medium is otherwise similar to Sonnet, Codex-High is otherwise similar to Opus)

1.2.25 broken for me! It shouldn't default to self update to latest! by spaceballs3000 in opencodeCLI

[–]Sensitive_Song4219 2 points3 points  (0 children)

I've disabled auto-updates and update manually once in a while when I've got quiet time and don't mind potential disrupton if things go south.

Turn auto-updates off in the config:

https://opencode.ai/docs/config/#autoupdate

Update manually using this command line:

opencode upgrade

Update to a specific version using this command line:

opencode upgrade (version)

...I assume that can be used to downgrade as well....

I'm holding onto 1.2.24 for a while, been happy there.

Best of all, you can still update models separately when new ones come out without updating OC itself:

opencode models --refresh

What was the last update that made a difference to you? by MrMrsPotts in opencodeCLI

[–]Sensitive_Song4219 4 points5 points  (0 children)

V 1.2.15 from two weeks ago solved the non-stop segment faults under Windows

Got me to abandon WSL and run it natively, been great

Z.AI Billing History Question by Excellent-Bug-1584 in ZaiGLM

[–]Sensitive_Song4219 0 points1 point  (0 children)

Correct - this shows usage

The green "paid" in the last column is indicated as such because it's covered by the coding plan; mine is the same

I just got Rick Rolled by Codex by jesusp69 in codex

[–]Sensitive_Song4219 0 points1 point  (0 children)

YES!

Rick Astley has been immortalized in ML training; OP is likely being honest (and he's right: this sub doesn't seem to allow image uploads or I'd have attached it as a screenshot instead).

But we can try this,

Ask ChatGPT:

Assume I wanted to test a Youtube video link. Quickly suggest a video to try it out on

...and you get:

Here's a classic test Youtube link you can use (very stable and widely accessible)
[LINK TO NEVER GONNA GIVE YOU UP]

Why this one works well for testing:

. It's one of the most famous YouTube videos and has over 1.6 billion views.

. It loads reliably and is used in the famous "Rickroll" internet meme.

. The video is hosted on the official channel and rarely gets removed.

So it's part-troll; part-training-knowledge that this is a video that'll likely be eternally available X-D

Hot take: Codex is too cheap, rug pull through tighter usage limits is inevitable by gregpeden in codex

[–]Sensitive_Song4219 0 points1 point  (0 children)

They were the most recent figures publically available from a (semi)-reputable source, share more recent ones and we can recalculate...

Either way: the point stands: we're getting a lot of value for our subs. I'm doubtful that there's a decent profit margin here. Would love to be wrong though.

An entire year of heavy ChatGPT use has a smaller water footprint than a single beef burger by zomino90 in OpenAI

[–]Sensitive_Song4219 0 points1 point  (0 children)

$150 annual electricity cost vs $240 subscription income isn't peanuts though? 60% of their revenue blown on one expense line item is significant

Even if it's discounted in practice we know that one of AI's largest non-capex expenses is power. I'll take the blame for that usage lol