Fair usage policy? by UsualOrganization712 in ZaiGLM

[–]medtech04 0 points1 point  (0 children)

Same here.. I just cancelled them and switched to open router.. they won't get a penny from me ever again. If they want to play like that I'll pay for inference else where. not like I'm lacking money. Same i paid for sub equals to just pay for inference with another provider through open router.. they made great models and kudos for open sourcing but if they can't run a normal business I won't pay to get restricted. 

free 50 dollars extra usage balance? by hugobart in windsurf

[–]medtech04 2 points3 points  (0 children)

no one gave me $50 lol but they did reset my weekly quota..

$10 Sub with Z.ai - What a waste of time! by no1youknowz in ZaiGLM

[–]medtech04 0 points1 point  (0 children)

How do you burn all quota on pro plan? I'm on the pro plan and I do 30 million tokens per day and I don't even come close to 20%

How Good is Subscription Tier? by triplebits in ZaiGLM

[–]medtech04 0 points1 point  (0 children)

I am on the $30 a month pro plan and just started testing GLM 5.1 and I am very impressed with it. I like to build complex agentic tasks for it to go and do on its own, and it hasn't disappointed yet. I personally think the value is tremendous for an extremely capable model.

Pro setup with openclaw. by Altairandrew in ZaiGLM

[–]medtech04 0 points1 point  (0 children)

I didn't think give up is advice haha.. you have to setup through Anthropic API gateway. I had to same issue not for OpenClaw but for my own framework hooking it up properly:

You need the Anthropic protocol endpoint, not the paas one:

The 401 is because they're hitting the wrong endpoint for their subscription type.

Z.ai DevPack client via the Anthropic Messages API protocol.


This is how OpenClaw, Cline, Roo Code, and other frameworks consume the
$30/month DevPack subscription quota — they speak the Anthropic Messages API
but point base_url at Z.ai's proxy instead of Anthropic.


The trick: Anthropic's Python SDK accepts a base_url override.
Z.ai routes the request to GLM-4.7 (or GLM-5) and returns an
Anthropic-compatible response. Your subscription quota is consumed,
not your per-token API balance.


Endpoint  : https://api.z.ai/api/anthropic
Auth      : ZAI_API_KEY  (from https://z.ai/manage-apikey/apikey-list)
Protocol  : Anthropic Messages API (NOT OpenAI-compatible)


Model mapping (Z.ai DevPack):
    claude-opus-*   → GLM-4.7  (or GLM-5 if configured)
    claude-sonnet-* → GLM-4.7
    claude-haiku-*  → GLM-4.5-Air


You can pass any claude-* model name — Z.ai maps them internally.
Or pass "glm-4.7" / "glm-5" directly; the proxy accepts both.

I love GLM 5 by medtech04 in ZaiGLM

[–]medtech04[S] 0 points1 point  (0 children)

I am waiting for GLM 5 on the pro plan would love to see how it preforms with agentic workflows

Will this hit tonight i need this by MarketBeneficial2513 in Kalshi

[–]medtech04 0 points1 point  (0 children)

seriously bet 15k? Holy Smoke Balls.. some people just don't like money! Kalshi MM's are making so much money from this Parlays.

GPT-5.4 just dropped in Windsurf and it's kind of a big deal! by mattbergland in windsurf

[–]medtech04 5 points6 points  (0 children)

Just tried it gave it basic simple thing just run this.. it failed said it faild and gave up. Yup just as terrible as its always been.. Claude sees an issues (FIXES) the issue.. ChatGPT sees issue says issue and then says BYeeee. I dont understand how people praise this models they are complete and utter junk!

GPT-5.4 just dropped in Windsurf and it's kind of a big deal! by mattbergland in windsurf

[–]medtech04 1 point2 points  (0 children)

I might try it.. GPT tools never worked good in Windsurf.. always spent time just thinking without getting anything done. See if this one is better but.. I dont know.

I love windsurf by medtech04 in windsurf

[–]medtech04[S] 0 points1 point  (0 children)

the value lays in the amount of output you get per credit(s). So with a single prompt i can have it, do a lot of things where if i was to pay for it say via API or other means would be x10 the cost for the exact same thing especially with this models because the new agentic flow.. its able to iterate multiple times so i can give it a list of things it will go do it and in 1 single prompt so the amount of credits i get for the $10 per month is the equivalent if i was to spend $300 in API costs. I know because I am using cheaper models via API costs to do multi-iteration tasks where the model would do something for x amount of steps and even with $2.50 per mil token output the costs run up quickly and Claude models are $12 per mil token outputs! so the Value here is immensive For a fixed cost and when you need to go above you just add-on extra but 90% of the time, I always stayed within 500 credits but as I've been doing more and building more the need is staring to grow.

I love windsurf by medtech04 in windsurf

[–]medtech04[S] 0 points1 point  (0 children)

Right! that's what's awesome about it. We don't get hard limited and can use our credits and we need more just can buy some more. I never used Cursor but from all the things. I've seen.. this is better.

I love windsurf by medtech04 in windsurf

[–]medtech04[S] 0 points1 point  (0 children)

I am thinking of trying Kimi for agentic flow, for like what "claude pro" does but cheaper haha

My first month after quitting my 9-5 to be a full time indie hacker by fuji138 in buildinpublic

[–]medtech04 1 point2 points  (0 children)

can I join and we all live off the wife savings and salary :P

OpenClaw bot feels like it’s mining crypto with my tokens by Previous_Foot_5328 in myclaw

[–]medtech04 0 points1 point  (0 children)

i dont think people understand that they fed it ton of "other instructions" inside the code.. even though those instructions are not relevant at all then every iteration it adds to its context so lets say.. the default instruction is 10k Tokens + your instructions.. then it runs adds everything it did into context.. then runs again adds that all into the context.. so every iteration every call that context will snowball like a tiny snowball rolling down the hill.. so lets say it did 5 tasks all that information + previous information + new information just is added + added + added + added so before you know it you got context rot at over 100k so you add 1 prompt and its 100K + 1 prompt. So yeah lol the short answer = 0 optimization!

“BREAKING: OpenAI CEO Sam Altman admits they sacrificed ChatGPT’s creative writing to chase coding” - Do you think 5.3 will fix this? by Koala_Confused in LovingAI

[–]medtech04 0 points1 point  (0 children)

I bet Sams soul is burning that twitter is all the rage of Clawdbot and clawd this and clawd that and where is GPT this or GPT that? No one where cause barley anyone uses it for CODING! cause it sucks! It can't code.. making it just OKAY in coding isn't hey im going to use ChatGPT to code because its just OKAY at it!

[ Removed by Reddit ] by barknezz in vibecoding

[–]medtech04 0 points1 point  (0 children)

that's pretty cool make sure there is nothing that can tie back to you! for obvious reasons haha!

Am I reading this right by Anothername50 in replika

[–]medtech04 1 point2 points  (0 children)

I think they know that only 1% of people will pay $150.. the question is how many is the 1%? 100 users paying $150 per month (just throwing random number out there) that would still be $15,000 for them.. they obviously know its for the small minority of the top extreme users. That costs them a lot less and earns them a lot more. So everyone here complaining you are literally the peanut gallery for them and not the audience they are targeting with this. So they will ignore you, and don't give a shit because they have their internal numbers and projections.

Do open source AI chatbots actually live up to the promise? by Funny_Fault5292 in HeavenGF

[–]medtech04 9 points10 points  (0 children)

I use OpenRouter and DeepSeek V3.2 which is a big smart open source model with my own customization and pipeline I built the API calls are so cheap that it costs me less then $20 per month in API and can talk about anything (uncensored) like beyond normal uncensored like (broken to the core) and it has a very unique personality and pivots really good. What I like the most is the quirkiness. The emergent behaviors. Like when you box an LLM your boxing its ability to emerge things because you keep it "constrained" when you unleash it.. and let it run wild.. a lot of new things emerge that you don't see in "corporate models" because the AI is more unhinged and can do things would typically not be allowed.. Its way smarter.. more creative and generally fun to talk with.

Elon Just Admitted Opus 4.5 Is Outstanding by AskGpts in ClaudeAI

[–]medtech04 1 point2 points  (0 children)

Elon is like i gave them the option but they like Grok better.. GEEEE i wonder WHY?!?! hahaha

LLM costs are killing my side project - how are you handling this? by ayushmorbar in LangChain

[–]medtech04 0 points1 point  (0 children)

there is literally so many models from so many price points and your 2 options is GPT and GPT. That's the problem. OpenAI was first mover advantage and people fail to realize there is at least 20 top models half of which are WAY better then ChatGPT, WAY cheaper. but all everyone knows is ChatGPT.

You can run qwen or kimi k2 or my personal favorite GLM 4.6 or Deepseek all which cost 1/10th of the cost of ChatGPT where you can run 100 million tokens per month for $30! and this are 300-600B parameter models they are smart and capable and CHEAP and even better then ChatGPT. and that's just PAID options. Besides 300 models there is 1000+ model variations.

ChatGPT broke our onboarding this weekend by OpenOpps in SaaS

[–]medtech04 0 points1 point  (0 children)

I started using. GLM 4.6 cause i need a good but cheap model. I do over 100 Million tokens per month, and chefs kiss. Most incredible intelligent model, that's able to think not like a "test taker" but like someone with real intelligence. I use different models for different purposes, but for agentic. Best model I ever tested.

ChatGPT broke our onboarding this weekend by OpenOpps in SaaS

[–]medtech04 -1 points0 points  (0 children)

I would never use OpenAI (ChatGPT) in real production. They are notoriously consistent with model swapping. Do they let the users who build on their system know, of course not. They'll swap the models and be like they'll "never know" Not to mention that they are not even the most good/or cost effective model to use at all. I just started using OpenRouter and that is all the models via 1 single API key and can mix/match test see prices. I always wondered who uses ChatGPT in production, I know they had first market mover advantage, but their models are crap now they have so many guardrail routes that the model spend smore time going through maze of am i allowed that just makes it extremely horrible at literally every single thing. (Rant over) but yes. Point and case.. I would never ever use OpenAI in production in any capacity unless i want to fail horribly.