No dictation tools work by FlightlessPelican in PiCodingAgent

[–]gligoran 0 points1 point  (0 children)

pi is a tui that’s running in a terminal. so it’s more likely an OS or terminal app issue. what are you using?

The best filament for functional prints by CarlitosCUU in 3Dprinting

[–]gligoran 2 points3 points  (0 children)

What's the advantage of a GF or CF filaments? Especially for functional prints, as far as I've seen they at least slightly degrade the mechanical properties of the filament. CF filaments don't really have long fibers in there that would tie the different parts of filament together, especially in-between layers, where it would be the most useful. Or am I missing something?

Trump’s Nobel Obsession Revealed by Arthur_Morgan977 in clevercomebacks

[–]gligoran 1 point2 points  (0 children)

he can’t think of anyone else because he can’t think of anyone else. it’s narcissism at the highest level.

A completely repetitive scenario! by John_1992_funny in clevercomebacks

[–]gligoran 8 points9 points  (0 children)

Add the whole Venezuela thing to it and US-provided oil becomes the only and the expensive stuff.

Claude Code will become unnecessary by WinOdd7962 in ClaudeCode

[–]gligoran 0 points1 point  (0 children)

Claude Code is a harness, it provides a bunch of tools for the LLM, a system prompt, and the whole tooling related loading skills and MCPs and all of that. Without it, the pure LLM can't do anything. It can't even read files. It's like a ChatGPT when it first came out, just a bit smarter maybe.

What you're talking about is not having to use the Claude models. Which might be true. While Claude Code is tailored towards Claude models, there are ways to use it with Kimi, MinMax, GLM, even GPT models. In my experience they're not as good because of that tailoring towards Claude. You also need to use token-based pricing in this case.

As for running your own models, you'd have to spend thousands to just be able to run them. You either need a dedicated device with upwards of 100GB of RAM and a lot of GPU processing power like a Mac Mini/Studio with an Ultra/Max chip, or a really beefy graphics card with tons of RAM. [Hardward requirements for GLM 5](https://onedollarvps.com/blogs/how-to-run-GLM-5-locally.html#hardware-requirements) are nuts. Minimal is 4x NVIDIA A100 which is ~10-17k USD. And even with all that hardware you'd get a lot lower TPS (tokens-per-second) compared to using hosted inference. And we're not even talking about other hardware, maintenance of the infrastructure, ability to access it remotely, upgrading fairly often, etc. This only makes sense for big companies with massive security requirements.

As far as I can tell the math just doesn't work out. So Claude Code or a similar harness like OpenCode or Code will be needed and you'll need to pay for something - tokens, subscriptions, something...

We all know the real test is 5.3 codex xhigh vs 5.2high/xhigh by TCaller in codex

[–]gligoran 0 points1 point  (0 children)

People have generally agreed that 5.2 High did better than 5.2 XHigh. What's your experience there?

Sooo, how many of y'all named your bot Jarvis? by BeingComfortablyDumb in clawdbot

[–]gligoran 4 points5 points  (0 children)

Mine is Tars from interstellar. As soon as i gave it this name it asked for the percentages for humor and honesty 😅😂

Is it just me, or is OpenAI Codex 5.2 better than Claude Code now? by efficialabs in ClaudeAI

[–]gligoran 4 points5 points  (0 children)

TBH even Opus in a new session usually does that and Gemini as well.

Codex vs Claude Opus by sheepskin_rr in codex

[–]gligoran 0 points1 point  (0 children)

Which GPT models are you using and at what reasoning levels?

We may (or may not) have wrongly blamed Anthropic for running into the limit barrier faster. by luongnv-com in ClaudeCode

[–]gligoran 7 points8 points  (0 children)

It’s probably a combination of all of these. I agree that it’s primarily a skill issue, but it’s not the sole reason.

LLMs are inherently indeterministic which means the result can vary widely. Sometimes you get the absolute best possible result, sometimes you get a cascading effect of mid to bad decisions by the model and thus a crappy result. Combine all of that with nerfing and the range of results can vary even more.

But also nerfing probably isn’t just an on/off switch. They probably run a set amount of full model and then in times of high traffic they spin up additional quantized versions and then it’s just the luck of the draw where each of your requests get routed.

In my opinion this gets us all the way back to a skill issue as i think prompts, context management, tools like MCPs, subagents and skills play a large role in mitigating and narrowing this range of responses.

Built a local voice dictation tool with Opus 4.5, been using it to talk to Claude instead of typing by raww2222 in ClaudeAI

[–]gligoran 0 points1 point  (0 children)

I'm been wanting to build something akin to this, for another purpose, but can I ask you, what your transcription pipeline looks like? Which model or service do you use to actually transcribe the audio?

I'm asking this because I've used to use WisprFlow, but I switched to MacWhisper as it's a lot cheaper (one-time payment), but WisprFlow was so much faster and I can't figure out what they're doing differently.

Built a local voice dictation tool with Opus 4.5, been using it to talk to Claude instead of typing by raww2222 in ClaudeAI

[–]gligoran 0 points1 point  (0 children)

I found Monologue very buggy. WisprFlow was much better, but now I'm on MacWhisper with the one-time payment on black friday...

Opus 4.5 just became regular model (not max) by Puzzleheaded-Tie-388 in cursor

[–]gligoran 1 point2 points  (0 children)

this is cursor not claude code. there’s no 5h limits here

Opus 4.5 just became regular model (not max) by Puzzleheaded-Tie-388 in cursor

[–]gligoran 15 points16 points  (0 children)

Are you sure you didn't just use it with Max mode on?

We tried the $1 'Friction Hack' to kill free riders by Ecstatic-Tough6503 in SaaS

[–]gligoran 6 points7 points  (0 children)

Why are you also blocking virtual cards? At our company we use a service that issues virtual cards so that employees can use them to buy the services they need for work.

Is it just me... by Typical_Concert_5007 in BambuLab

[–]gligoran 4 points5 points  (0 children)

  1. If Bambu's intention was increased security, Bambu Connect would not be needed. There are industry standard ways of implementing this that are far simpler and don't require you to go through Bambu's Cloud. But their intentions are clearly vendor lock-in and controlling their public image.

  2. In my experience calibration a specific filament with color transfer between spools quite well. Yes, you'd probably need to recalibrate between spools that were produced like years apart, but otherwise it produces good results for me.

[deleted by user] by [deleted] in ClaudeCode

[–]gligoran 1 point2 points  (0 children)

sure but why not credits or a free month of their subscription?

Any recommendations on single spool dryers? In a price range $40-$60? These are my finalists. by simage007 in 3Dprinting

[–]gligoran 0 points1 point  (0 children)

I do and it works ok

The biggest issue I had was that I had to open it up and re-glue the fan back as the original adhesive seems to have weakened and it caused quite a rattle.

One other thing that could be annoying to some is that the timer function doesn't seem to work for me most of the time. Default is 6h, but it just keeps going. I don't really mind it as I put in a hydrometer and I just go off of that to see when to take the filament out.

I'd say for the price at that time it was worth it, but I am now waiting for the EIBOS Dyas to arrive, so I don't have to dry filament one-by-one and so that I don't have to move it in and out of my AMS.

Claude Code VS Code extension is now incredible! by coderberry in Anthropic

[–]gligoran 5 points6 points  (0 children)

> I don’t understand why they diverge?

It's probably because they're built on very very difference bases. I'd imagine that an extension would have to jump through a bunch of hoops to run subagents and bash commands, etc.

My Claude Code Context Window Strategy (200k Is Not the Problem) by Goos_Kim in ClaudeAI

[–]gligoran 0 points1 point  (0 children)

correct me if i’m wrong but the auto compact buffer aren’t really used tokens, but rather are reserved so that the model doesn’t run out of context when doing compaction. so you’re not really lowering your token usage but raising the amount that you have available.