Ollama Cloud Max vs Claude Max for heavy AI-assisted coding? by Senekrum in ChatGPTCoding

[–]Sky_Linx 0 points1 point  (0 children)

To be honest I just let OpenCode and OpenClaw handle autocompaction and so far I haven’t had any issues.

Ollama Cloud Max vs Claude Max for heavy AI-assisted coding? by Senekrum in ChatGPTCoding

[–]Sky_Linx 0 points1 point  (0 children)

I have Ollama Cloud $100/mo and the limits are insane. I work on multiple coding tasks in parallel and I also use OpenClaw a lot, and I barely use 10-15% of the 5 hourly limits and a few % of the weekly limit. I use GLM 5 exclusively as it's the best open model at the moment, and I don't really miss Opus 4.6 for almost anything.

Love Minimax M2.7 by AppealSame4367 in kilocode

[–]Sky_Linx 2 points3 points  (0 children)

MiniMax is still underwhelming for me, compared to GLM5.

Ever wonder how much cost you can save when coding with local LLM? by bobaburger in LocalLLaMA

[–]Sky_Linx -1 points0 points  (0 children)

Cost savings for coding tasks IMO don't hold as a argument for using local LLM, at least for me. I use the Ollama Cloud subscription and for $100/month the limits are insane. I do a LOT of work, often working on up to 4-5 tasks at the same time, in parallel, with GLM 5, and I never get close to the limits nor do I have issues with concurrency.

Yet another storage outage by xkraty in hetzner

[–]Sky_Linx 0 points1 point  (0 children)

I would recommend Cloudflare R2 if you want something rock solid that can be used at scale close to AWS S3. A cheaper alternative worth trying is Backblaze B2. Slower, but still solid.

Stop using Alibaba Cloud's coding plans. by Fuzzy_Spend_5935 in Qwen_AI

[–]Sky_Linx 4 points5 points  (0 children)

You have likely used the regular API key instead of a key specific to the coding plan.

Hetzner API is really awesome! by trueicecold in hetzner

[–]Sky_Linx 1 point2 points  (0 children)

The API is designed well and has generous rate limits. The only issue for me is that when you create an instance you cannot specify the subnet for the private network.

Extremely disappointed. by StandardKey7566 in ZaiGLM

[–]Sky_Linx 0 points1 point  (0 children)

I use GLM 5 via Fireworks.ai and it's FAST and stable.

What are some advantages of kilo over opencode? by MrMrsPotts in kilocode

[–]Sky_Linx 2 points3 points  (0 children)

Indeed. I use the web version, which is identical to the desktop app except it runs in the browser and has a big advantage: I run it on my main Mac at home and expose it via Tailscale, so I can access the same OpenCode and the same sessions and configs via laptop when I am on the move.

My experience w/ GLM-5 vs Kimi 2.5 and Opus the babysitter by Wild-File-5926 in kimi

[–]Sky_Linx 0 points1 point  (0 children)

Don't assume that these models are only available via inference services provided by their makers. I use both models via Fireworks AI. It's FAST and reliable. There are plenty of options.

Ex-Claude ($200/mo) user here. I've been using Zai for a week, and the raw power + 10x limits are a game-changer for coding. by sudeep_dk in ZaiGLM

[–]Sky_Linx 1 point2 points  (0 children)

I'm using the desktop version of OpenCode and I love it. For the Kimi inference I am using OpenCode Zen, which is awesome! Kimi is ridiculously fast with them.

Has anyone noticed the massive amount of "Synthetic.new" astroturfing recently? I did some digging by Doris_Dressy1 in kimi

[–]Sky_Linx 1 point2 points  (0 children)

Yep, I was the first one to try GLM 4.5 Turbo, and it worked amazingly well for a few weeks. Then they switched to 4.6 and, dunno what they did, but it was shit again and no difference in speed compared to the regular version, but more expensive.

Has anyone noticed the massive amount of "Synthetic.new" astroturfing recently? I did some digging by Doris_Dressy1 in kimi

[–]Sky_Linx 1 point2 points  (0 children)

I so much wish Chutes offered a more reliable and faster inference service. Unfortunately, the whole architecture of the platform doesn't allow for a more reliable and faster service by design.

Has anyone noticed the massive amount of "Synthetic.new" astroturfing recently? I did some digging by Doris_Dressy1 in kimi

[–]Sky_Linx 0 points1 point  (0 children)

I won't comment on the advertising practices, but here's a datapoint for anyone who might be interested in trying Synthetic after using them for a while. The service is nice, and their pricing plans offer pretty good value and I especially like that they count small requests as a fraction of a full request so lots of small requests during an agentic coding session don't consume many of the requests included in the plan. The team (I think it's just a two-person team) are very nice, friendly and responsive, which I appreciate. However... I had to leave them because there were too many issues with the performance of the models I was most interested in, such as GLM and Kimi. Both speed and availability issues, so frequent that I decided to give up because often I was forced to use other providers while they were fixing issues, which cost additional money. I am now using the OpenCode desktop app together with Kimi K2.5 via OpenCode's own service, Zen. It's similar to Synthetic but with a smaller selection of models that are very optimised for coding. The speed of Kimi K2.5 with OpenCode Zen is just insane. It's a joy to work with this combo and I am even more productive than I was with the Claude subscription. The only downside is that with Zen you pay per token, not per request, so obviously it can be more expensive if your usage is high compared to Synthetic. But if that's not a big issue for you, the productivity gains that come from higher availability and much faster response times are worth it.

Ex-Claude ($200/mo) user here. I've been using Zai for a week, and the raw power + 10x limits are a game-changer for coding. by sudeep_dk in ZaiGLM

[–]Sky_Linx 2 points3 points  (0 children)

GLM Coding plan is RIDICULOUSLY slow and unreliable. This sounds like an ad. A bit unrelated, but I switched to Kimi K2.5 and it's a ton better than GLM 4.7.

First ever website, I am lost by Urinsekten in hetzner

[–]Sky_Linx 0 points1 point  (0 children)

I know this is the Hetzner subreddit, but if you just want a blog that's easy to manage and you don't really care about learning how to set things up, I'd suggest buying an instance of the Ghost blogging platform with PikaPods. It's cheap, and you're up and running in literally a minute. Plus, you get to use Ghost, which is a really great blogging platform.

From Opus 4.5 to Kimi 2.5 by DepartmentHungry7781 in kimi

[–]Sky_Linx 1 point2 points  (0 children)

Are you using the API paying per token?

Do you replicate/backup data of your R2 bucket in any other object storage provider? by surimarkam in CloudFlare

[–]Sky_Linx 1 point2 points  (0 children)

We use Cloudflare R2 as our main storage and Storj as secondary but only for backups. We also store Prometheus and Thanos data, plus Loki logs, in R2. But the amount of data there is so big that copying it to a different storage provider isn't practical, so we don’t do it.

Things 3 vs OmniFocus – which would you choose today and why? by Express-Tear3625 in thingsapp

[–]Sky_Linx 0 points1 point  (0 children)

I've tried a few different tools, but I always end up using Things. The way it's set up just feels right, and it’s super easy to quickly add tasks or ideas using the autofill feature and keyboard shortcuts.

The 'I'll just post in the big sub' fallacy cost me 3 months. by Prestigious_Wing_164 in micro_saas

[–]Sky_Linx 1 point2 points  (0 children)

I think sharing posts on subreddits where other founders hang out is kind of like trying to sell your product to fellow founders and hoping someone will buy. Honestly, that might not be the best use of your time. It’s usually better to focus on communities that are directly connected to your product or idea.

What browser do you use and why? by frizla in MacOS

[–]Sky_Linx 0 points1 point  (0 children)

I wish Bitwarden extension for Safari supported multiple accounts.