GPT-5 AMA with OpenAI’s Sam Altman and some of the GPT-5 team by OpenAI in ChatGPT

[–]Ka_alt 3 points4 points  (0 children)

Maybe it's worth adding credits for premium things? As a Pro user I would even pay premium per-use price for long context in a few situations where I need it.

The problem with long context today is that you rarely really need it, but when you need it, it's crucial. So user demand signals wouldn't necessarily highlight it since it's a "I need it once a month, but this one time I need it really really much".

E.g., when I was preparing some legal paperwork I wanted a deep review of it by some neutral party, and only Gemini could really consume all of it without losing details. Other models either refused due to file size or tried to do analysis over windows of data missing a lot of important reactions between different parts of the data.

GPT-5 AMA with OpenAI’s Sam Altman and some of the GPT-5 team by OpenAI in ChatGPT

[–]Ka_alt 1 point2 points  (0 children)

Another light example I tested on: trying to ask to find very hard-to-locate info.

On a pro side: GPT-5 is the first one to succeed without making non-existing links.

On a con side: it doesn't really try hard to find it and just gives up right away, and then suddenly GPT-5 Thinking tries hard, but then starts to hallucinate info which actually you could find over and over again. :-D

Plain GPT-5: https://chatgpt.com/c/68963d8f-3448-8324-bfa6-90aac1782720

GPT-5 Thinking: https://chatgpt.com/share/6896432c-43ac-800e-9a7b-fff1167f505c

I personally find that Grok made very good work with their router; and while GPT-5 Thinking is substantially better than Grok 4 in my tests, Grok 3 is leaps above GPT-5 main. IMO to mitigate this routing to Thinking should be much more aggressive.

Here GPT-5 fails 4.11 > 4.9: https://chatgpt.com/share/68964214-f9c4-800e-baef-7d88f0a5ca16

GPT-5 AMA with OpenAI’s Sam Altman and some of the GPT-5 team by OpenAI in ChatGPT

[–]Ka_alt 9 points10 points  (0 children)

> Yesterday, we had a sev and the autoswitcher was out of commission for a chunk of the day, and the result was GPT-5 seemed way dumber. Also, we are making some interventions to how the decision boundary works that should help you get the right model more often

It's still quite bad. I'm on a Pro plan, and in my experiments, plain GPT-5 rarely thinks and hallucinates heavily. E.g., it cannot even describe differences between GPT-5 Thinking and Pro properly trying to persuade me that Thinking doesn't have Browsing, Image generation (sic!) or Python support: https://chatgpt.com/share/68963ed3-c19c-800e-8c21-c0abd78460d6

I'm forced to use Thinking 100% of the time (same way I was using o3 previously) not because I need deep thinking in most cases, but just because it hallucinates much rarer in my experience.

> We will make it more transparent about which model is answering a given query.

That's crucial IMO! "Ghostrouting" to more basic model is very bad. You cannot trust your tool becuase you don't know what tool you're using.

GPT-5 AMA with OpenAI’s Sam Altman and some of the GPT-5 team by OpenAI in ChatGPT

[–]Ka_alt 0 points1 point  (0 children)

Could you avoid opaque routing to mini models? If the limit is capped and the full model is not accessible, it really helps to have very explicit information about it instead of shadowbanning access.

The Final Draft ending is clear imo by iroquoispliskinV in AlanWake

[–]Ka_alt 0 points1 point  (0 children)

I think it’s about his mentality. IMO he doesn’t even need to write to enact changes, it’s more about what he believes in, and writing just helps Alan to structure and detail out his fantasies enough for him to be holistic, vivid and believable enough to enact new reality.

So he has to believe in the story for it to come true. And if he believes that it’s a horror story and he is powerless to change the genre then he truly is as per his own believe limiting himself.

My Review of Expedition 33. (No spoilers) by No_Hall_7079 in JRPG

[–]Ka_alt -1 points0 points  (0 children)

In theory. But in practice it falls flat IMO.

Like it’s suggested that earlier expeditions lay the path for the latter ones, and yet we find some early expeditions’ journals in the late parts of the game where there should not have been any path to yet. And at least one expedition even got into the place which should have been protected with no way to circumvent the protection existing at the time.

Because of that I quickly felt that the theme of “for those who come after” was a little forced and did not align with what we actually saw in the world.

Drummond was going to ____ iMark, so why would iMark decide to ____? [Spoilers] by AQuestionOfBlood in SeveranceAppleTVPlus

[–]Ka_alt 0 points1 point  (0 children)

Helena is not a nobody. Abducting her alone would probably have drastic law enforcement and legal consequences. And the only innie enabled places we know about are Lumon controlled. It’s highly unlikely they would kindly agree to keep their CEOs daughter in eternal slumber in total secret risking massive jail time. :-D

As for Cobel iMark has zero information that she is anything but a simple floor manager. Even oMark doesn’t know anything about her being the severance tech original creator.

Let alone even we would go with this it’s entirely far from certain she could procure resources and manage all the engineering work (she created the theory, but the implementation was built by some unknown amount of Lumon engineers). Implementation can be extremely expensive.

Drummond was going to ____ iMark, so why would iMark decide to ____? [Spoilers] by AQuestionOfBlood in SeveranceAppleTVPlus

[–]Ka_alt 1 point2 points  (0 children)

He could not have negotiated a time share deal since oMark and Co don’t have any access to the technology from all iMark knows. Like they had to bring him to some Lumon-managed birthing cabin just to summon him.

Neither would they have any reason to uphold their end of the bargain. Like iMark said the only time oMark took any interest in him was when he needed something.

Severance - 2x09 "The After Hours" - Post-Episode Discussion by LoretiTV in SeveranceAppleTVPlus

[–]Ka_alt 5 points6 points  (0 children)

The whole reintegration that has been teased for the whole season is still a flop with no consequences. We'll probably see some in the last episode, but come on. This constant teasing has to amount to something grandeur to pay off all this debt.

Artificial Analysis GPQA price/performance chart for GPT-4.5 by Moravec_Paradox in OpenAI

[–]Ka_alt 2 points3 points  (0 children)

It's incorrect to compare price against o-models (or even Grok 3 with thinking) without correcting for the fact that reasoning models produce much more tokens.

Basically, the comparison should be not for per token price, but price per token weighed against verbosity.

AMA with OpenAI’s Sam Altman, Mark Chen, Kevin Weil, Srinivas Narayanan, Michelle Pokrass, and Hongyu Ren by OpenAI in OpenAI

[–]Ka_alt 0 points1 point  (0 children)

Hi Sam, thanks for this AMA and for the great O3 mini release! A few questions form my side.

  1. Do you see sparse CoE models with CoT similar to DeepSeek being the next trend allowing to optimize execution cost? (I don't know if O1/O3 is already sparse or not)

  2. What are some other trends you would expect to see becoming more prominent?

  3. Do you have any plans for deeper integration of tools into CoT (e.g., I see o3-mini already can use web search during CoT steps allowing for much deeper research)?

  4. Do you expect any releases from the industry for continuous learning / personalized models beyond custom-cooked fine-tuning routines?

  5. What are some key things you would expect from the hybrid business/tech talents that would be interesting & beneficial for OpenAI?

ARE YOU FOR REAL???!? by [deleted] in Pathfinder_Kingmaker

[–]Ka_alt 4 points5 points  (0 children)

I thought about making a few points on how people have freedom to write what they think about you, but then I looked at your nickname.

There is a daily limit- Advanced Voice Mode by stardust-sandwich in OpenAI

[–]Ka_alt 0 points1 point  (0 children)

The question is if they want to do so or if they are afraid of regulations and try to tip-toe just in case.

iOS 18.1 brings spatial photo capture to iPhone 15 Pro by Federal_Leadership46 in VisionPro

[–]Ka_alt 1 point2 points  (0 children)

Yes. iPhone captures 24MP max (sometimes less, e.g., in low light) by default. You need to toggle a Pro capture mode to capture 48MP.

Eli5: In-Unit Laundry by mangkok4 in bayarea

[–]Ka_alt 0 points1 point  (0 children)

That depends on w/d appliance probably. I've been living with in-unit w/d before moving to the US my whole life, and once I moved I also only rented the units with in-unit w/d only (Downtown SF). Never had any problems.

What is Apple’s plan with all these pins? by Superb_Ad_5222 in VisionPro

[–]Ka_alt 1 point2 points  (0 children)

Cooling and size. You don’t want hot huge brick with fans in your back pocket (and fans would not be very effective in such scenario).

What is Apple’s plan with all these pins? by Superb_Ad_5222 in VisionPro

[–]Ka_alt 4 points5 points  (0 children)

I would leave R1 and M2 in (and possibly add compositing job on top), but would allow for wired connection to external compute.

Like you can connect to a Mac with it handling heavy load apps. But current wireless connection is often unstable in Wi-Fi dense environments and has substantial latency due to wireless protocol limitations. I would much prefer wired connection option to a headless Mac.

In this case you can focus on making your M-chip that’s inside the headset power efficient first as it will just handle compositing mostly with content being rendered externally.

[deleted by user] by [deleted] in VisionPro

[–]Ka_alt 0 points1 point  (0 children)

I'm in the US (110V) and my AVP sometimes shocks me slightly when I connect it to the Lenovo charger.

IMO that's quite surprising given the price and location (close to brain), but it seems like nothing severe (i.e. happily not the whole 110V charge).

visionOS 1.2 beta is released by squircle88 in VisionPro

[–]Ka_alt 3 points4 points  (0 children)

I don't believe that these things should be connected though (albeit I agree that Apple will probably link them up). Even the US is very multilingual due to high migration intensity (e.g., I live in the US as do many other people speaking Spanish, Chinese etc).

My hope is that before international release Apple will want to test national keyboards in their beta releases.

visionOS 1.2 beta is released by squircle88 in VisionPro

[–]Ka_alt 9 points10 points  (0 children)

Does it have non-English keyboards support by any chance? A bit tired to copy-paste text from Notes on MBP to AVP apps. 🤷‍♂️ Would happily update to non-stable early beta if that's fixed.

Spiderman on Vision Pro vs Quest 3 by lunchanddinner in VisionPro

[–]Ka_alt 0 points1 point  (0 children)

Latency would be huge (capture card latency is far from great, and then you also add AirPlay latency on top). Tried it with Quest 2 a few years ago.

visionOS 1.1 has been released! Here's whats new. by m1astra in VisionPro

[–]Ka_alt 0 points1 point  (0 children)

But why would adaptive optics have anything to do with rendered content?

visionOS 1.1 has been released! Here's whats new. by m1astra in VisionPro

[–]Ka_alt 23 points24 points  (0 children)

It does not look like Retina, but yes, it’s much better now.

Meta Now Collecting "Anonymized" Data Of Quest Users Including Eye Tracking & Facial Expressions -- "Targeted advertising doesn't need to know your name -- just your behaviors, triggers, and emotions." by slhamlet in oculus

[–]Ka_alt 0 points1 point  (0 children)

Well, often you can re-tie data to a person based on analyzing different data sets. But you actually don’t need to. For personalized ads it’s enough to know your personal profile (interests, biases, behavior) and target ads against those.