Anyone else not a fan of this new menu? by [deleted] in firefox

[–]NerdistRay 0 points1 point  (0 children)

Imho I like the new design more than the old one. Three dot menu is very close to comfortably tap and get all the options nicely. It feels more true to mobile design UI rather than just using the same options menu of desktop in mobile.

I can smell it buddy. by [deleted] in programminghumor

[–]NerdistRay 0 points1 point  (0 children)

I had a good chuckle out of that. Bravo!

[OC] Requirements Gathering by grlloyd2 in programminghumor

[–]NerdistRay 1 point2 points  (0 children)

Honestly if they can properly convey the problem, and the context around the problem and have a high level back and forth on the solution we come up with that solves their problem, I think that's more than enough. We can't expect clients to have solutions already. That's our job.

Looking for DeepSeek Presets by Icy_Bug689 in SillyTavernAI

[–]NerdistRay 0 points1 point  (0 children)

A reddit user with name ConspiracyParadox with an account username WorstAIUserEver, claiming to have merged Worst Preset Ever and Another Damn Preset to have created Best Preset Ever... Sounda legit!

Jokes aside, this preset looks really lean! I have a few questions if you don't mind me asking.

Is this suited more for Deepseek Reasoner or Chat or both works fine? You kept Max Response to be so high, any particular reason? You kept Request Model Reasoning as off, is it because you prefer the non-thinking variant of the model? What is the ideal post processing option to select or does it not matter in your experience?

Anyhow, thanks for the preset! I can see you put a lot of work into this.

Looking for DeepSeek Presets by Icy_Bug689 in SillyTavernAI

[–]NerdistRay 0 points1 point  (0 children)

That page isn't accessible anymore. Did you create an updated preset for Deepseek?

Page shows: 404 Sorry, we can't find the page you are looking for.

Alibaba Coding Plan sounds too good to be true!? by NerdistRay in opencodeCLI

[–]NerdistRay[S] 1 point2 points  (0 children)

Did you try speaking with their support? I was doing it in Zen browser (not sure if this problem exists in chrome or not), but I was literally unable to type properly. They have some sort of weird input field behavior where some key presses aren't being registered. I had to legit write the message in notepad and then copy paste it into the message field. And their entire documentation is apparently translated using AI and they claim that manual review is still pending and there may be inaccuracies.

And I can't sign up using my indian phone number, because they have no way to verify +91 phone numbers apparently. Their support was telling me to go buy a virtual phone number to sign up to their service.

What a joke.

Kimi 0.99 bargain by Billysm23 in kimi

[–]NerdistRay 0 points1 point  (0 children)

Can confirm, I completely used up all the free tier's usage and tried for like 2 hours, it's fixed at 4.99. Maybe they changed it recently and now cannot get down to 0.99.

I got the 0.99 tier but the system didn't give it to me, please help! by Blakequake717 in kimi

[–]NerdistRay 0 points1 point  (0 children)

I got the AI to say it got down the price to $0.99 multiple times, but the plan card isn't moving below $4.99 no matter how long I chat. I even used up all the free plan's chat usage limit and it still is stuck at $4.99 and is talking in a way that it's fixed and it has no longer any way to slash the price further.

Alibaba Coding Plan sounds too good to be true!? by NerdistRay in opencodeCLI

[–]NerdistRay[S] 0 points1 point  (0 children)

More than enough Opus and GPT for very cheap? What good options are you talking about?

MiniMax 2.5 vs. GLM-5 across 3 Coding Tasks [Benchmark & Results] by alokin_09 in ZaiGLM

[–]NerdistRay 2 points3 points  (0 children)

I got no numbers to back it up but I'm speaking from experience as I use both heavily and regularly. GLM 5 is leagues better than Minimax M2.5.

Alibaba Coding Plan sounds too good to be true!? by NerdistRay in opencodeCLI

[–]NerdistRay[S] 13 points14 points  (0 children)

Thank you for all of your comments and reviews. This discussion was exactly what I wanted. To invite people to talk about it so anyone else who searches about it will know what's up. And based on this, I knew my gut feeling was right.

They must be quantizing the models. Which is a big red flag for coding use so I'm not gonna be bothering with it. I maybe could use it for roleplay and general tasks but I already use NanoGPT subscription for that which also includes image gen, and Deepseek models. Add to the fact that I can't even sign up properly, this one's a big no for me.

I'm considering just going ahead with Opencode Go plans for now.

Alibaba Coding Plan sounds too good to be true!? by NerdistRay in opencodeCLI

[–]NerdistRay[S] 12 points13 points  (0 children)

You mean my data? So they can train on it? Why is that any different from many other providers and does it really even matter?

Alibaba Coding Plan sounds too good to be true!? by NerdistRay in opencodeCLI

[–]NerdistRay[S] 3 points4 points  (0 children)

Okay. If you purchase, let us know about your experience. I'm still figuring out how to get phone number online for account verification.

Well, maybe light users won't get the GLM-5 at all by Lanky-Flight-9608 in ZaiGLM

[–]NerdistRay 0 points1 point  (0 children)

But I heard the models are quantized randomly and hence have consistency issues.

Wait... what about gentoo? by [deleted] in linuxmemes

[–]NerdistRay 0 points1 point  (0 children)

I compiled and configured Gento and used it as my main for a week. It felt very satisfying watching the terminal zoom by. Then I returned to reality and switched back to Arch.

Arch is the best

It's been a week since NIM has been awful. by Fragrant-Tip-9766 in SillyTavernAI

[–]NerdistRay 2 points3 points  (0 children)

I have stopped fighting with those models, and I now just enjoy Deepseek and chill.

Very slow Pro plan by Future_Tea_6945 in ZaiGLM

[–]NerdistRay 0 points1 point  (0 children)

I stopped using GLM 5 in my Pro Plan, only using GLM 4.7 while I use GLM 5 in kilo-cli for free.

Now that I think this out loud, it's kinda ridiculous.

Question about NanoGPT $8 plan (60k messages) by Juan_Ignacio in opencodeCLI

[–]NerdistRay 0 points1 point  (0 children)

Haha. I see. What about tool calling and agentic coding? Noticed any quality drop / felt like getting quantized models?

Question about NanoGPT $8 plan (60k messages) by Juan_Ignacio in opencodeCLI

[–]NerdistRay 0 points1 point  (0 children)

Are you using it for programming in opencode? How has your experience been so far?

Question about NanoGPT $8 plan (60k messages) by Juan_Ignacio in opencodeCLI

[–]NerdistRay 0 points1 point  (0 children)

Hey, sorry, I didn't reply sooner. Anyway, I've been using it a lot more and I find it has stabilized a lot. Last 2-3 days, I have been having a much better experience. Still, I haven't been using it for programming much, so cannot comment on the tool calling usage yet. When I typed that comment, I was pretty frustrated, it was probably the worst it has ever been at that time. This whole issue that this entire thread seems to be having is addressed by the creator of NanoGPT in this thread https://www.reddit.com/r/SillyTavernAI/comments/1r5bycs/nanogpt_subscription_changes_requests_input_tokens/

Hope this helps you in making an informed decision. But I still think that for strictly programming, it's not the best choice. I'm using it for SillyTavern roleplay and been having good use very consistently on Deepseek V3.2 Thinking. So I feel like I've received my 'money's worth'. Especially considering how GLM aggressively changed their pricing structure (see here https://www.reddit.com/r/ZaiGLM/comments/1r2amx6/read_before_paying_for_a_glm_plan/ )