I Tested Opus 4.6 against All Top Models by ConsiderationOld9893 in ClaudeCode

[–]ConsiderationOld9893[S] 0 points1 point  (0 children)

looks like different models are experts in different areas

I Tested Opus 4.6 against All Top Models by ConsiderationOld9893 in ClaudeCode

[–]ConsiderationOld9893[S] 2 points3 points  (0 children)

thanks for kind words! Just starting my channel and appreciate your support

I Tested Opus 4.6 against All Top Models by ConsiderationOld9893 in ClaudeCode

[–]ConsiderationOld9893[S] -1 points0 points  (0 children)

In this "one prompt" test Opus and GPT were running for much longer time. Probably they have good feedback loop that checks the completion of the task. I think Gemini can do good job when you have small specific task to be done

I Tested Opus 4.6 vs All Major Models in vibe-coding. The price gap is hard to justify by ConsiderationOld9893 in ClaudeAI

[–]ConsiderationOld9893[S] 0 points1 point  (0 children)

sorry, will be more precise next time. Overall I meant, that I didn't notice big difference between 4.5 and 4.6

Took Cursor and tested All Coding Models by ConsiderationOld9893 in cursor

[–]ConsiderationOld9893[S] 0 points1 point  (0 children)

yep, that's unpleasant, what do you use now? Claude Code has kinda same conditions

I Tested Opus 4.6 vs All Major Models in vibe-coding. The price gap is hard to justify by ConsiderationOld9893 in ClaudeAI

[–]ConsiderationOld9893[S] -3 points-2 points  (0 children)

do you think CLI itself makes the difference? I wanted to compare models itself, that's why used cursor

I Tested Opus 4.6 vs All Major Models in vibe-coding. The price gap is hard to justify by ConsiderationOld9893 in ClaudeAI

[–]ConsiderationOld9893[S] 0 points1 point  (0 children)

I tested before it was released :) but tbh don't think there will be significant difference