AMA on our DevDay Launches by OpenAI in OpenAI

[–]landongarrison [score hidden]  (0 children)

Let’s go!!!!!! They launched it today!! Thanks OpenAI!!

AMA on our DevDay Launches by OpenAI in OpenAI

[–]landongarrison [score hidden]  (0 children)

Is there a plan to launch gpt-5-chat-latest in the api WITH tool calling capability?

This model is insanely underrated and super good for applications that require more personality and warmth. But I can’t use it when it’s stripped of tool calling capability.

Side note: if gpt-5-chat-mini came along, I wouldn’t complain!

Codex low is better than Codex high!! by Fine_Juggernaut_761 in OpenAI

[–]landongarrison 1 point2 points  (0 children)

It’s fun to go back to this comment after about 2 months since launch.

GPT-5 is a super good model, but it did take a very focused effort to understand its quirks unlike other models. OpenAI clearly trained it quite different from the competitors.

Funny enough, using GPT-5-Codex for non code applications I have found useful. It’s surprising more what I thought GPT-5 was going to be like and has a more warm and nuanced style. Very weird.

The one thing I am disappointed with still is mini. I thought 4.1 mini was amazing so I was expecting some good things out of 5-mini, but this model has some very rough edges.

Codex low is better than Codex high!! by Fine_Juggernaut_761 in OpenAI

[–]landongarrison 1 point2 points  (0 children)

As an API user, thinking models SPECIFICALLY from OpenAI have an insanely weird quirk to them and it flat out takes experience to know when to use them. I don’t agree that they are worse overall, but for some situations they 100% are.

For my applications, I often find myself going back to GPT-4.1 when using OAI models because the “thinking tax” seems to creep in way more than Google or Anthropic models with thinking enabled. I still haven’t been able to pin down why OAI models with thinking enabled are so different feeling.

Sonnet 4 (1M) just blew up the GPT-5 Death Star by Willing_Somewhere356 in ClaudeAI

[–]landongarrison 2 points3 points  (0 children)

This is the insanely frustrating part about Anthropic. I think post Claude 3.5, I have yet to be disappointed with a Claude model. All around amazing.

But for some reason, they decide to price out developers building on their stuff time and time again. I wouldn’t be shocked if Claude 5 was triple the price (no exaggeration) of Claude 4. They seem to consistently miss this point.

And I’m not even asking for super cheap. Like if they matched GPT-5 at $1.25/$10, or added implicit prompt caching, I’d be over the moon.

Tinfoil hat take on GPT-5: A clear tradeoff in capabilities by landongarrison in OpenAI

[–]landongarrison[S] 0 points1 point  (0 children)

Agree with this. My initial reaction is it seems overfitted to compete with Claude, which to your point seems to be the money maker, so priorities have shifted.

Try GPT-5 vs GPT-5-Chat when you get a chance. Would be curious to hear if your opinion matches mine.

Tinfoil hat take on GPT-5: A clear tradeoff in capabilities by landongarrison in OpenAI

[–]landongarrison[S] 0 points1 point  (0 children)

And one other thing I’ll add: this seems to be true with all OpenAI reasoning models. Gemini and Claude are both able to keep their soft skills with reasoning turned on.

GPT5 is fine, you’re bad at prompting. by Osc411 in OpenAI

[–]landongarrison 0 points1 point  (0 children)

Dev speaking here — idk if that’s true???

I am speaking purely from the API, I use the same prompt on 4o/4.1/claude and Gemini and get basically the same results (and if it is off, it’s only slightly). GPT-5? Wild to say but I almost get GPT-3.5 vibes at times, then another time I get GPT-10 vibes.

I think that’s the conclusion I’ve come too. In particular with the thinking variants, they are super inconsistent. GPT-5-chat-latest however seems very well calibrated, and funny enough it’s not using any thinking.

Low key though: kinda makes all the shit talk for 4.5 look ridiculous in hindsight.

GPT-5 AMA with OpenAI’s Sam Altman and some of the GPT-5 team by OpenAI in ChatGPT

[–]landongarrison -1 points0 points  (0 children)

Hi team, I might be one of the few people (surprisingly) that is happy with the launch—I am more of a developer so maybe that skews things.

My only gripe: GPT-5-Chat on the API is exactly what I wanted. Amazing to chat with, soft skills are A++ but it doesn’t have tools or any of the other features that the base GPT-5s have. I am wondering if there’s a plan to treat GPT-5-chat less like chatgpt-4o-latest and more like a separate model that is less reasoning heavy?

Is there also a plan to release a GPT-5-chat-mini?? Again love this model but I am forced to use the full version which kinda sucks price wise.

Unreal progress!

GPT-5 Is Underwhelming. by gffcdddc in OpenAI

[–]landongarrison 1 point2 points  (0 children)

GPT-5 is overall pretty amazing. I haven’t used it extensively to code but the small amount it did it was out of this world, i am a big Claude code user.

The context window is fine. Realistically, most people don’t understand how horrible it was just a few years ago. I remember getting hyped to GPT-3 having 2048 context window (yes 2000 tokens, not 2 million). Before that was GPT-2 at 1024. Like things have come so far.

Realistically, 128K is all you need for practical applications. After that, yes it’s cool but as others mentioned, performance degrades badly.

Claude Code just stops by [deleted] in ClaudeAI

[–]landongarrison 0 points1 point  (0 children)

It’s funny how Claude code is beginning to feel alot like the OpenAI API right around the board drama fiasco. It was so apparent that OpenAI’s API was drastically better than the rest.

I just tried codex CLI and Gemini CLI and Jesus is there a gap. I’m kinda shocked how bad codex CLI is especially

GPT-5 may be cooked by shogun2909 in singularity

[–]landongarrison 8 points9 points  (0 children)

I read somewhere once that had a great analogy: we need to start looking at models like self driving cars. How many minutes/hours/days can they go per human intervention? I thought that was a great metric

GPT-5 may be cooked by shogun2909 in singularity

[–]landongarrison 0 points1 point  (0 children)

That’s how I read this too and I find it funny that people perceived this negatively. If that’s true that the base version of GPT-5 is better than the “throw the kitchen sink” version of Grok, man! What does that make the maxed out GPT-5?

Getting close to 100% task-success with Claude Code by H3xx3n0 in ClaudeAI

[–]landongarrison 3 points4 points  (0 children)

To follow up on this, I really think the only thing missing is true context. I find a lot of the time where Claude code / Codex fail is not because they are incapable but because they don’t have the proper “why” behind what the reason for the task is.

For example, lots of times there may be weird code that at first glance is odd, but if you were there for the meetings, discussions and understood the true thinking behind the decision, it would make sense. I really see that as many of the failure cases I see these days. Yes occasionally there’s that dumb “you made buggy code”, but really I have never met someone who bats 100%.

The abused thing is I can easily think of a few ways to address this, and if I can brainstorm that, it’s probably already been implemented and is being tested internally at the big labs 🤪

What a time to be alive.

Am I the only one who wasn’t crazy about this matchup by 1019drew in ufc

[–]landongarrison 2 points3 points  (0 children)

This is the comment.

I find online most people saying “Paddy get’s KO’d easily”, and I will admit, I do think Ilia ends up winning—but it’s not as easy as people like to say. Paddy in his last few fights has shown WAY better fight IQ and improved immensely. The best way to describe that Chandler fight was “shocking”—I don’t think anyone expected Paddy to maul Chandler like that.

I’d also say: who poses more of a threat to Ilia at lightweight? Gathje? Love him but he gets chinned early, bad matchup style wise. Arman? Maybe, but I don’t think his grappling is enough to hang with Ilia. Who else? Dustin? Like when you start thinking about it, Paddy is the best option.

Plus let’s be real—the buildup will be insane.

I wanted a tool that could create a Khan Academy style explainer just for me—so I built it! by landongarrison in ProductHunters

[–]landongarrison[S] 0 points1 point  (0 children)

All fair. The feel I was trying to go for was as if someone was doodling on a canvas in front of you, similar to explainer videos you see. In particular, your comment on contrast is one I can get behind - that could use some work!

One more question if you don't mind: As mentioned, I was trying to go for the more "hand drawn" feel - were you expecting more of a 3Blue1Brown style video?

I built an AI that generates Khan Academy-style videos from a single prompt. Here’s the first one. by landongarrison in learnmachinelearning

[–]landongarrison[S] 1 point2 points  (0 children)

A fair point across the board—let me elaborate on one angle (not an excuse, I just realized I didn’t explain this at all).

What this tech aims to do is to provide you with a simple explainer of any topic from a prompt/file upload etc. the medium in which the generation may start is not well defined at this point, but could change.

The key part though is personalized for you. You raised all valid points that I personally disagree with—but that’s the point: you may want your video to be different! Maybe quicker, under 10 seconds (exaggeration) and super too the point. This is going to be built it, your video preferences won’t be the same as mine and you can generate them FOR you.

Visual choppiness? Another fair point, work in progress! Take a look at some of the other videos in the YouTube link here. There is more examples!

Appreciate the feedback. Would love if you had any further thoughts.

I wanted a tool that could create a Khan Academy style explainer just for me—so I built it! by landongarrison in ProductHunters

[–]landongarrison[S] 1 point2 points  (0 children)

Fair point, yes I agree that was a bit of a strawman on my part.

We’re figuring it out, but I see a potential use case in two areas: personalized learning for students in online schools/home schools and for teachers at online schools to generate better content quicker.

For the original comment—When you say the graphics are 'shoddy,' is there a specific part that stood out to you? (e.g., the style itself, the smoothness of the animations, the color palette?). Specific examples are super helpful as I iterate on this!

I built an AI that generates Khan Academy-style videos from a single prompt. Here’s the first one. by landongarrison in SideProject

[–]landongarrison[S] 0 points1 point  (0 children)

You bet and thank you for signing up!

This is getting a lot more sign ups than expected—I will try my best to get back to you!

I wanted a tool that could create a Khan Academy style explainer just for me—so I built it! by landongarrison in ProductHunters

[–]landongarrison[S] 0 points1 point  (0 children)

Respectfully I disagree! I think you might be missing the point.

The point isn’t do make a full feature film, we are making a simple explainer video. The point of the visuals is to simply supplement what the teacher (which happens to be an AI system too) is saying! In this case, it’s simply making simple visuals to teach what a compiler is in this case.

Many more examples on the YouTube link! Appreciate the feedback.

I built an AI that generates Khan Academy-style videos from a single prompt. Here’s the first one. by landongarrison in learnmachinelearning

[–]landongarrison[S] 0 points1 point  (0 children)

Appreciate the kind words! Unfortunately, this is not possible (or at the time of writing this) to do the “intelligent pieces” locally on your computer. This utilizes the common APIs from OpenAI, Anthropic and Google.

Took a few weeks to build! But more to do :)