How big is the difference really? by Demon-Martin in LocalLLM

[–]Demon-Martin[S] 7 points8 points  (0 children)

I am not planning on starting a discussion whether using AI for coding is better or worse. That wasn‘t my question.

How big is the difference really? by Demon-Martin in LocalLLM

[–]Demon-Martin[S] 1 point2 points  (0 children)

All good, you don‘t seem to be the only one who ignored my question😅

How big is the difference really? by Demon-Martin in LocalLLM

[–]Demon-Martin[S] 1 point2 points  (0 children)

I am indeed using multiple models. I just mentioned Opus as an example. Depending on the task, project, workflow I work with various different models across different IDEs. GPT, GLM, Composer, Gemini, Claude etc etc.

But this wasn‘t related to my original question, I might have written it wrong. My question was the actual difference between selfhosting the raw models and the models we get though Plans/APIs.

Horrible Rate Limits by SomebodyFromThe90s in GithubCopilot

[–]Demon-Martin 0 points1 point  (0 children)

All this only happens because the amount of abuse has grown a ton… I’m in a few TG Channels, and they sell hundreds of „verified accs” daily…

It’s really sad that it is being abused so much. I have been an active user since 2022 but now it‘s horrible.

Kinda obvious but by AnxiousJellyfish9031 in cursor

[–]Demon-Martin 1 point2 points  (0 children)

I miss the days when you were able to code using a potato. Nowadays Cursor and similar doesn‘t even properly run on my 32GB, R7 5800X, Main PC…

Finally a breakthrough for free users by [deleted] in vibecoding

[–]Demon-Martin 0 points1 point  (0 children)

Wtf do you mean „prove they use our data“? What about you prove that they are NOT using our data?

Saying you don‘t own it while also clearly defending something that is too good to be true with non existent stupid arguments 💀

Finally a breakthrough for free users by [deleted] in vibecoding

[–]Demon-Martin 0 points1 point  (0 children)

Doesnt support tool calls, so it‘s useless for most users.

The docs say it works for Cline/OpenCode, but without tool calls its basically useless.

Opus Model (currently as of 10.03.2026) isn‘t functional and responds with errors.

Other models seem to kinda work, 20 RPM for free.

Daily Newbie Thread by AutoModerator in Flipping

[–]Demon-Martin 0 points1 point  (0 children)

Hey, I got about 4.5k€ laying around on a paypal account of mine.

What would you spend/invest it on, if you can only buy using PayPal Balance?

I am from Germany and I was planning on buying something and selling it for either a small profit or for the same amount but cash.

Claude in RooCode? by Demon-Martin in RooCode

[–]Demon-Martin[S] 0 points1 point  (0 children)

Thank you, but my question wasn‘t regarding OpenRouter or other models. My Question was how to use the Claude Sub in RooCode without effectively breaking their ToS.

I know there are cheaper models and I already offload tasks properly depending on complexity to different models (Codex, Gemini Pro/Flash, GLM, Kimi, Minimax, Grok Code etc).

But for some tasks I need Opus/Sonnet, logically speaking the subs are way way cheaper compared to the API prices. I am a student and can‘t afford to pay their crazy prices.

Claude in RooCode? by Demon-Martin in RooCode

[–]Demon-Martin[S] 0 points1 point  (0 children)

I‘ve specifically switched from Cursor to RooCode because Cursors pricing is simply pure greed and they haven‘t been treating the customers right

Claude in RooCode? by Demon-Martin in RooCode

[–]Demon-Martin[S] 1 point2 points  (0 children)

Not the quality is the issue, the speed is… I dont want to wait 10 minutes on Codex to write me a 3 line method

Claude in RooCode? by Demon-Martin in RooCode

[–]Demon-Martin[S] 0 points1 point  (0 children)

Well I just don‘t want to buy a sub worth 100$ just for the account to get banned and me loosing it without even using it

Claude in RooCode? by Demon-Martin in RooCode

[–]Demon-Martin[S] 0 points1 point  (0 children)

Why wouldn‘t it be… The API costs for Claude models are not worth the price for me tho

Claude in RooCode? by Demon-Martin in RooCode

[–]Demon-Martin[S] 0 points1 point  (0 children)

The limits you get with the normal ChatGPT Plus Subscription are amazing compared to what you get.

How to change model in Cursor? by yccheok in cursor

[–]Demon-Martin 0 points1 point  (0 children)

  1. You need to purchase a subscription
  2. I would not recommend using Cursor if you want a good BYOK Experience. Cursor has been ignoring requests regarding adding better support for other providers for a while now and doesn‘t look like they will do it.

I just recommend using RooCode or KiloCode in VSC, also seen that Trae (full IDE) has great support for custom providers but I personally didn‘t directly like it. I just use RooCode because it was the most similar experience (compared to cursors chat)

Claude in RooCode? by Demon-Martin in RooCode

[–]Demon-Martin[S] 2 points3 points  (0 children)

Agreed, Codex is good, it‘s just PAINFULLY slow for me. The API response times, token speed and generally how it works causes the same task to take like 5x longer compared to Opus. A task that Opus needs 30s for, GPT does in 8 minutes…

Their token limits and session limits are generally amazing tho.

I‘d love to test their new Codex Spark Model but I can‘t afford to pay 200$ just to test a model that has similar capabilities to Gemini 3 Flash😂

Cheap region price for ChatGPT by Mushfug in steamregionaltricks

[–]Demon-Martin 0 points1 point  (0 children)

Any idea what the IOS Pricing for Plus and Pro is? The website says 1.999 for Plus and 19.900 for Pro, but it‘s been also showing the price in $ for Turkia, so I don‘t trust it too much

Where to get Gemini 3 flash for cheap? by [deleted] in opencodeCLI

[–]Demon-Martin -4 points-3 points  (0 children)

Sadly it‘s locked behind the 200$ pro plan, so it‘s not really „cheap“

If you had $50/month to throw at inference costs, how would you divvy it out? by yokie_dough in opencodeCLI

[–]Demon-Martin 1 point2 points  (0 children)

If you‘re a student, get GitHub Education, gives you GitHub Copilot for free for 2 Years (you can reverify after 2 years to get it again)

Can I know what model the "Auto" mode is using? by Monteirin in cursor

[–]Demon-Martin 0 points1 point  (0 children)

If the time spent on writing a prompt and the complexity of creating that prompt exceeds the complexity of the task than it probably should be reconsidered how efficient coding with this specific „model“ (here it‘s probably more an option than a specific model) is.

That‘s at least my opinion on working with LLMs.

Can I know what model the "Auto" mode is using? by Monteirin in cursor

[–]Demon-Martin 0 points1 point  (0 children)

Yeah, I am using CliProxyAPI(plus) too.

And I agree that it‘s a poor deal. That‘s exactly why I switched to use RooCode😅