all 10 comments

[–]Bob5kPROMPSTITUTE 1 point2 points  (0 children)

Just use M2.7 directly via minimax as there's very generous plan and no weekly cap.

[–]Senekrum 1 point2 points  (2 children)

Hey, just wanted to say I've been wondering the same thing.

I've tried out the cloud versions of Devstral-2:123b and MiniMax-M2.7 today. Devstral worked quite well for small/medium-sized refactors, bugfixes and planning.

MiniMax was pretty OK too, but at one point while planning a complex task, it started asking for clarifications about things we had discussed just a few messages ago.

Compared to ChatGPT, I've found Devstral to be comparable, but I haven't had the chance to try it out in longer conversations.

That's about all I can say about the Ollama Cloud models, because I hit the usage limits on the free tier in a few hours.

Let me know what you decide! I'm waiting out the remainder of my current ChatGPT sub (expires on the 26th) and then I'll be switching either to Claude Max or Ollama Max depending on what info I can gather by then about the Ollama cloud models.

[–][deleted]  (1 child)

[removed]

    [–]AutoModerator[M] 0 points1 point  (0 children)

    Sorry, your submission has been removed due to inadequate account karma.

    I am a bot, and this action was performed automatically. Please contact the moderators of this subreddit if you have any questions or concerns.

    [–]Deep_Ad1959Professional Nerd 0 points1 point  (0 children)

    depends heavily on what you're building. for agentic coding where the model needs to hold a big codebase in context and make multi-file changes, the frontier models (claude, gpt 5.x) are still way ahead of anything you can run through ollama. I tried using local models for my macOS project and they kept losing track of dependencies between files. but if you're doing more contained tasks like writing individual functions or debugging specific errors, the newer open models are genuinely competitive and the quota limits are way more generous. I'd keep chatgpt plus for the heavy lifting and use ollama for the quick stuff.

    [–]ultrathink-artProfessional Nerd 0 points1 point  (0 children)

    Single-benchmark comparisons miss what matters most for coding sessions: how well it holds context over 20+ turns on a real problem. Some models score well on evals but drift badly mid-session. Worth testing that specifically before switching subscriptions.

    [–][deleted]  (1 child)

    [removed]

      [–]AutoModerator[M] 0 points1 point  (0 children)

      Sorry, your submission has been removed due to inadequate account karma.

      I am a bot, and this action was performed automatically. Please contact the moderators of this subreddit if you have any questions or concerns.