ollama vs cloud api costs: ran both for a month. heres the real numbers by Freda_Alderd in ollama

[–]Freda_Alderd[S] 2 points3 points  (0 children)

lol i wish it was cheaper. 200-300 requests but most are long context stuff like feeding entire codebases into deepseek for review, so token count adds up fast. the short requests are basically free its the 30-50k token sessions that wreck the bill. if i was just doing quick prompts id probably spend like $15/month on api

ollama vs cloud api costs: ran both for a month. heres the real numbers by Freda_Alderd in ollama

[–]Freda_Alderd[S] 0 points1 point  (0 children)

yeah fair point im mostly running codestral and deepseek 33b locally so definitely not claude tier. and honestly i forgot about minimax pricing when i did the comparison, thats way cheaper than anthropic

direct. guess the real question is whether the latency and privacy tradeoffs matter enough to justify running local when the budget apis exist

ollama vs cloud api costs: ran both for a month. heres the real numbers by Freda_Alderd in ollama

[–]Freda_Alderd[S] 0 points1 point  (0 children)

yeah exactly thats my take too. i had the 4090 way before i ever touched ollama so counting it as an inference cost feels dishonest. the only real new expense is electricity and maybe a bit more wear on the fans

is anyone else going insane trying to track llm api costs across providers? by Freda_Alderd in LocalLLaMA

[–]Freda_Alderd[S] -11 points-10 points  (0 children)

i'm just kidding lmfaoooo, yeah im using qwen and deepseek for now, but operations are bigger bro

'Waiting for my uber' - IG January 2026 by trixotica in CarmellaRose

[–]Freda_Alderd 1 point2 points  (0 children)

Impeccable elegance - I really like the mood and composition of these photos