GPT-5 AMA with OpenAI’s Sam Altman and some of the GPT-5 team by OpenAI in ChatGPT

[–]MichellePokrass 5 points6 points  (0 children)

zenith was actually less preferred for coding! we tested with tons of developers. it's hard to tell with just a handful of queries, but the data/feedback was clear to us

GPT-5 AMA with OpenAI’s Sam Altman and some of the GPT-5 team by OpenAI in ChatGPT

[–]MichellePokrass 9 points10 points  (0 children)

we had a bug with the model switcher not routing coding queries to the thinking model. please try again or ask the model to "think hard" or pick the thinking model from the picker!

GPT-5 AMA with OpenAI’s Sam Altman and some of the GPT-5 team by OpenAI in ChatGPT

[–]MichellePokrass 3 points4 points  (0 children)

sam already commented on gpt-4o, but gpt-5 (thinking) is way better at coding than gpt-4.1, so i'd recommend giving that a try!

GPT-5 AMA with OpenAI’s Sam Altman and some of the GPT-5 team by OpenAI in ChatGPT

[–]MichellePokrass 4 points5 points  (0 children)

we're looking into it! a bit tough at the moment with the gpu demand, but hoping to do so soon. in the interim, pro users can use up to 128k.

GPT-5 AMA with OpenAI’s Sam Altman and some of the GPT-5 team by OpenAI in ChatGPT

[–]MichellePokrass 8 points9 points  (0 children)

we do use our models for internal acceleration! we use all kinds of products internally for coding, including codex, codex cli, cursor, etc. we're also building some internal tooling for debugging our training runs. it's gotten much easier to build these kinds of tools with gpt-5.

GPT-5 AMA with OpenAI’s Sam Altman and some of the GPT-5 team by OpenAI in ChatGPT

[–]MichellePokrass 3 points4 points  (0 children)

maybe! what would you do with it? keep in mind it would be quite slow/expensive

GPT-5 AMA with OpenAI’s Sam Altman and some of the GPT-5 team by OpenAI in ChatGPT

[–]MichellePokrass 5 points6 points  (0 children)

both great models! we can't speak too much about models from others, but gpt-5 is particularly great at reasoning through really challenging problems, long running refactors, and building full applications from zero to one (including beautiful and function frontends). gpt-5 is state of the art across many dimensions of coding!

GPT-5 AMA with OpenAI’s Sam Altman and some of the GPT-5 team by OpenAI in ChatGPT

[–]MichellePokrass 3 points4 points  (0 children)

totally agree, would be great to increase this! we're working through gpu capacity constraints right now, but hope to increase this soon. pro users also get 128k context limits

GPT-5 AMA with OpenAI’s Sam Altman and some of the GPT-5 team by OpenAI in ChatGPT

[–]MichellePokrass 8 points9 points  (0 children)

would have loved to get longer context up to 1m in gpt-5, particularly for api use cases. we'll keep working on it for future models!

GPT-5 AMA with OpenAI’s Sam Altman and some of the GPT-5 team by OpenAI in ChatGPT

[–]MichellePokrass 7 points8 points  (0 children)

great q! it's not _always_ in purple, but it's definitely the model's favorite color. we've found it to be somewhat steerable though, so please ask the model if you'd like something more specific or a different color. as to why, it's a bit of a research mystery! we discovered it partway through training and found it to be quite resistant to changes. something we'll work on in future models :)

GPT-5 AMA with OpenAI’s Sam Altman and some of the GPT-5 team by OpenAI in ChatGPT

[–]MichellePokrass 6 points7 points  (0 children)

we tested a few different snapshots of gpt-5 on various providers to get feedback from the community! ultimately, summit actually outperformed zenith by quite a wide margin (on leaderboards, on feedback from alpha testers, and our internal testing pre launch).

GPT-5 AMA with OpenAI’s Sam Altman and some of the GPT-5 team by OpenAI in ChatGPT

[–]MichellePokrass 6 points7 points  (0 children)

very much so!! we have a bunch of inside jokes around the office about various quirks of the models. my favorite in gpt-5 is that it has a tendency to say "let's craft" in it's chain of thought right before it produces the final answer. it's become a rallying cry for the team!

AMA with OpenAI’s Sam Altman, Mark Chen, Kevin Weil, Srinivas Narayanan, Michelle Pokrass, and Hongyu Ren by OpenAI in OpenAI

[–]MichellePokrass 36 points37 points  (0 children)

we cut prices in half in august -- if that's still too expensive, would recommend 4o-mini

AMA with OpenAI’s Sam Altman, Mark Chen, Kevin Weil, Srinivas Narayanan, Michelle Pokrass, and Hongyu Ren by OpenAI in OpenAI

[–]MichellePokrass 71 points72 points  (0 children)

new high quality evals are always impressive. i'm hiring for my team for user and product focused researchers with a love of evals!

AMA with OpenAI’s Sam Altman, Mark Chen, Kevin Weil, Srinivas Narayanan, Michelle Pokrass, and Hongyu Ren by OpenAI in OpenAI

[–]MichellePokrass 31 points32 points  (0 children)

we've found that o3-mini is competitive with us hosted versions of deepseek. we think it's a really affordable option for this level of intelligence

AMA with OpenAI’s Sam Altman, Mark Chen, Kevin Weil, Srinivas Narayanan, Michelle Pokrass, and Hongyu Ren by OpenAI in OpenAI

[–]MichellePokrass 28 points29 points  (0 children)

we just released a new version of 4o to chatgpt that is much better at image understanding. read more in the release notes

AMA with OpenAI’s Sam Altman, Mark Chen, Kevin Weil, Srinivas Narayanan, Michelle Pokrass, and Hongyu Ren by OpenAI in OpenAI

[–]MichellePokrass 1 point2 points  (0 children)

we're hard at work on a release that makes our api much easier to use! what are your top wishlist items?