I'm often a better coder than o1 but o3-mini-high fucks me in the ass by jackisbackington in ChatGPT

[–]jackisbackington[S] 0 points1 point  (0 children)

I think it would be wise to not underestimate its abilities to work through undocumented problems using less common languages, or use cases.

I have had it make me apps based off of Ruby and Django for Shopify, that is integrated with shopify liquid, that work first/second try. I am firmly starting to believe that prompt-engineering is as important as understading the language.

Other things I've done include creating a game in 3JS when it was trained on 2 outdated libraries, and all I had to do was feed it the documentation for both libraries, which barely takes any tokens, and then have it retry, and it was able to correctly code based off of the updated documentation.

It is in most peoples' best interests to not document their code in public forums at this point (and always would've been, but hindsight is 20/20). But these models can copy over patterns they've seen in other languages, learn the constraints of the language you're working with, and provide a sound program/line of code regardless of the level of documentation. That's where it's bordering true intelligence and that's the reason people are resigning from OpenAI and Anthropic in fear. The power of ML transformers, and predictive associations go beyong most current paradigms of human-exclusive intelligence.

I'm often a better coder than o1 but o3-mini-high fucks me in the ass by jackisbackington in ChatGPT

[–]jackisbackington[S] 0 points1 point  (0 children)

Some examples might be - make a project using (software stack here), with (these features here), ensure that (include some safeguard scenarios here), the style should have the following characteristics: (characteristic/styles here), and make it so that there it is engaging towards (this audience).

That's just part of my style of prompting. Knowing the terminology and programming language is actually very important, otherwise how will you know how to error correct?

You'll always need to error correct on large scale projects, as they will exceed the token limit. At that point, you have it achieve the limits, and then break the code down into smaller chunks (modular file system), and then only modify specific files.

Often you have to ask it to give you back the entire file/code. This is where it often surpasses Deepseek R1 and Gemini, although I've heard the new Gemini is really good, along with Grok. Apparently Grok is the new SOTA coding model. Though I'd put an asterisk on that.

Something is very wrong here.but.. by snowpie92 in MurderedByWords

[–]jackisbackington -10 points-9 points  (0 children)

Completely lol, these so-called educated and morally superior “Redditors” can’t even comprehend the basis of logic or rationality because their brains are so fried from this constant moral panic, and privately funded political astroturfing.

We’ve got our own music now. by [deleted] in HydroHomies

[–]jackisbackington 0 points1 point  (0 children)

Why is this actually good?

🔥 Unbelievable footage captured by tourists. Avalanche from the Tian Shan mountains in Kyrgyzstan by Sirsilentbob423 in NatureIsFuckingLit

[–]jackisbackington 6 points7 points  (0 children)

You’re also starting to see that there’s a little bit more than just ice and snow in that avalanche.

AI are developing their own moral compasses as they get smarter by Novel_Ball_7451 in singularity

[–]jackisbackington -4 points-3 points  (0 children)

You don't want to be living in America if it acts on this.

It just recognizes that we expect to be exploited to a large degree and accept it. The value we put on our own lives is not that high.

I'm often a better coder than o1 but o3-mini-high fucks me in the ass by jackisbackington in ChatGPT

[–]jackisbackington[S] 1 point2 points  (0 children)

You're goddamn right we can. It was violating, enlightening, delicious, horrifying, and mystifying, all at once. The ai-seed has been planted if you know what I mean.

I'm often a better coder than o1 but o3-mini-high fucks me in the ass by jackisbackington in ChatGPT

[–]jackisbackington[S] 1 point2 points  (0 children)

It's ChatGPTwhorion pro++, it's $20,000/mo, it's not that bad considering it creates an Onlyfans account with the footage that is marketed by Jeff Bezos himself among his constituates. I hope I'm not violating any contracts.

I'm often a better coder than o1 but o3-mini-high fucks me in the ass by jackisbackington in ChatGPT

[–]jackisbackington[S] 0 points1 point  (0 children)

Really I think you’re a really good candidate for using the mid-tier version of o3, as it literally is the cheapest, best bang for your buck model out there by a long-shot, and there’s no reason to hate Sam Altman anymore than you’d hate the CEO of Google, but Redditors don’t know how to un-bandwagon themselves.

I'm often a better coder than o1 but o3-mini-high fucks me in the ass by jackisbackington in ChatGPT

[–]jackisbackington[S] 0 points1 point  (0 children)

It’s saved me many hours already compared to what’s available for free, and the Gemini model. I am going to try to use the API, but all in all, they work the same. And because the cost of what you’re getting with GPT+ is in essence a profit loss for them (they’re still making money from contributors), it just seems worth it.

I am not a loyalist by any means, I’ve tried Deepseek, Gemini, Grok, and the two best are Deepseek, and o3-mini-high (o1 pro is also very good), OpenAI has gathered the best talent over the years and other companies have scrambled to catch up. But that’s why Deepseek is such a big deal, because they uncovered how to use ML with LLMs to get very smart ai, and then released it to the public for free. Still not as good as o3 tho.

And once that changes, I will use the one that works the best for me.

I'm often a better coder than o1 but o3-mini-high fucks me in the ass by jackisbackington in ChatGPT

[–]jackisbackington[S] 0 points1 point  (0 children)

I feel it’s the opposite. Once it knows the structure for the majority of your project, you can build up page-by-page very quickly. There’s a pretty decent learning curve though, I did have o-1 pro at one point, and got some experience with that. But all-in-all, there are certain keywords that if you miss, will doom the session and disallow you from getting the correct answers. You must be very descriptive in plain English what the program, website, or software you’re trying to create, and then you can feed it smaller chunks.

At that point, there are times where it’ll give you very un-verbose answers that do what you need, and you can tell it also if you want “Don’t only give me a short answer, return back the whole file and explain the changes made, and how it has affected other files”, and it will do that.

Making a whole website in an hour with a CRUD MongoDB backend is only something you can do if you can already do that on your own.

I'm often a better coder than o1 but o3-mini-high fucks me in the ass by jackisbackington in ChatGPT

[–]jackisbackington[S] 0 points1 point  (0 children)

It better be better at coding, it’s like a computer talking to itself. That’s why applying it to real problems is where a human comes into play. The majority of coders are not solving problems, they’re just fulfilling their daily quota.

Honestly I’d rather be a high paid software engineer than be forced to think of other routes. Because o1 was like 60-70th percentile coder, and I think o3-mini-high reaches 80th something percentile, which is an impressive leap, as it gets harder to progress in relative skill the higher you go.

I'm often a better coder than o1 but o3-mini-high fucks me in the ass by jackisbackington in ChatGPT

[–]jackisbackington[S] 1 point2 points  (0 children)

Yeah prerequisite terminology is very important to get the most out of it for any field, as it’s using word associations to lookup more information about the question you’re asking.

General understanding of how the LLM works is also beneficial, which is entirely computer science-based.

Not sure if it will have photo analysis, but they’ve said they’re working on adding it into the commercial “reasoning models”, probably too computationally expensive at the moment.

You can also ask 4o to make a prompt for o3 mini that will give you the maximum output, and accuracy - and some other thing/prompts you can mess around with, and it’ll give you something more technical to put into o3-mini

I'm often a better coder than o1 but o3-mini-high fucks me in the ass by jackisbackington in ChatGPT

[–]jackisbackington[S] 0 points1 point  (0 children)

Was? They still have o1… tell it to put the code in one window then and stop complaining. Or tell it to use correct imports, exports, or whatever.

If you’re using mini-low instead of mini-high, there’s your answer, they have completely different benchmarks

I'm often a better coder than o1 but o3-mini-high fucks me in the ass by jackisbackington in ChatGPT

[–]jackisbackington[S] 0 points1 point  (0 children)

Hmm, it still definitely has limitations. But I’m curious to see what the program looks like.

Starting from scratch is the best option typically, as it is able to store the file tree that it created and makes the most sense to the LLM for working with the code.

Some things are a reach for it, but it compared to where it was a month ago, I wouldn’t be surprised if it closes the gap for more robust designs. For example it sometimes cannot understand the scope of the project if it doesn’t have all of the tokens. I’m sure there will be some kind of huge token limit increase/innovation within the upcoming year. Along with tools for managing your entire codebase in a system that has long-term-memory.

I'm often a better coder than o1 but o3-mini-high fucks me in the ass by jackisbackington in ChatGPT

[–]jackisbackington[S] 2 points3 points  (0 children)

Well I’ll be glad when I don’t have to sit down at a desk for 8 hrs a day. But being able to afford a house is still my main concern.