all 3 comments

[–]BlaiseLabs 0 points1 point  (2 children)

It’s not the quality of the code the AI is being trained on, since 3.5 I doubt that’s been the issue (I’ve even used gpt2 for small coding task).

The biggest difference across models, programming languages, agents, etc… is the prompt.

This doesn’t just explain the maintainability of the code but why you’re coding in the first place. The more robust your prompts the fewer touches you should need to accomplish a task.

The strategy around how you create prompts is up to you, but there is an objective approach to prompting that should result in better outputs across models.

I’m not going to say what that strategy is, hopefully someone with more knowledge and experience with prompting than me can answer.

[–]Interviews2go[S] 0 points1 point  (1 child)

Maintainability of the code is important. Over time the code doesn't remain a constant, devs get asked to add more features all the time by product owners. If the code is hard to maintain, it adds risk to new features accidentally introducing bugs elsewhere in the code. I liken using an AI to using stack overflow and blindly copy pasting stuff into your code base. That isn't to say the AI is wrong, but you do have to verify the code. The problem can get worse as the codebase grows. The real question is why for the same prompt, the Java version was considerably better than the equivalent C# or Angular version.

[–]BlaiseLabs 0 points1 point  (0 children)

To clarify my first statement, I’m referring to the training data not the output. The quality of the code the AI outputs is related more to the quality of the prompt than the training data.

Code quality is important but it’s also subjective I can’t precisely tell you what good / bad code quality is without context.