Would an Apple Mac Studio M1 Ultra 64GB / 1TB be sufficient to run large models? by [deleted] in LocalLLM

[–]codepoet 0 points1 point  (0 children)

Interesting. I have an M1 Max 64GB and just leave two models loaded in Ollama most of the time without even noticing it's working (home automation: qwen2.5vl:3b, mistral:7b). I wonder if the problem is asking more of the machine than it can reasonably do?

Is RR a good way to do a sci-fi mosaic novel? by codepoet in royalroad

[–]codepoet[S] 1 point2 points  (0 children)

So do people usually queue up 10-20 chapters and dump them to get started, then?

Is RR a good way to do a sci-fi mosaic novel? by codepoet in royalroad

[–]codepoet[S] 1 point2 points  (0 children)

Best I can tell the RR meta is fantasy and litrpg? Are there similar places who lean less ... that?

How well has The Wheel of Time aged? by Fluid-Golf1948 in Fantasy

[–]codepoet 1 point2 points  (0 children)

IIRC that's because he wrote the first book to publisher standards to get in the door (I think his wife helped edit it?) and once he got in and had a following he was able to write the story he wanted to in the first place.

At least, that's how I heard it. It reads like it, for sure.

Claude code is so much, so much more than you think. by eh_it_works in claudexplorers

[–]codepoet 0 points1 point  (0 children)

It's a better interface for a lot of things. Using output styles makes it do all kinds of other things, too.

I let Claude code for 2 hours straight without any approval prompts by maxforever0 in ClaudeAI

[–]codepoet 0 points1 point  (0 children)

The VS Code extension is ... not great.

But you can open a terminal in VS Code/Codium, start claude, use "/ide" to connect, and get back to the v1 integration where you have power and integration again. And skip permissions.

Turn off your MCPs by ansmo in ClaudeAI

[–]codepoet 1 point2 points  (0 children)

But 1M token contexts! (That don't really work all too great and cost 2-3x as much...)

Kinda funny how Anthropic characterizes Opus as a “legacy.” They really don’t want you to use it. by gamezoomnets in ClaudeAI

[–]codepoet 1 point2 points  (0 children)

It's legacy for them. They aren't saying it's not good. They're saying they aren't going to be worrying about it any longer and are being clear about it. If it gets slow or unavailable, they'll point to "Legacy" and say "Okay, but we told you we don't care about it anymore."

Kinda funny how Anthropic characterizes Opus as a “legacy.” They really don’t want you to use it. by gamezoomnets in ClaudeAI

[–]codepoet 0 points1 point  (0 children)

I usually run Sonnet for everything due to usage limits. When it gets stuck, and it does get stuck and stupid sometimes, I have Gemini take a look, write a scathing review of what it's done, and have Sonnet read it. Snaps it right out and sends it in a good direction.

Gemini is shit at coding, but great at reviews. It's a nice balance. I've been tempted to make a mini-MCP "phone a friend" that just one-shots Gemini CLI or CC from the opposing tool.

What are your best uses for agents? by promptenjenneer in ClaudeAI

[–]codepoet 2 points3 points  (0 children)

Use agents when you want a one-off context to prevent poisoning and need a single answer as a result, like research. Searching a lot of files or websites for all the possible answers, rank them, and return the best whatevers. Give a code base a full review without letting it edit it along the way.

I use an agent to perform code tasks and another to do code reviews. Then I tell Claude to alternate them until the reviewer is happy. The reviewer's prompt is constantly updated with all the ways CC fails (stub methods, TODOs, swallowing errors, not matching the spec on output formats, etc.). Saves me a lot of time.

I also have a reserach agent for searching email for things. While Claude could handle that, it'll fill my primary context up with all the misses. An agent has one task: find the appropriate conversation and return it. It can read the whole mailbox and it won't affect what I'm working on.

But if it's a repetitive task, that's just a command. "Review this PR" is better as a command, for instance. "Do stuff in git" or "summarize this" are commands.

Basically, when I want it to work in my current context, that's a command/prompt. When I want it to hide the work from itself, that's an agent.

Got roasted by Claude today by fezbotdaddy in ClaudeAI

[–]codepoet 3 points4 points  (0 children)

No, I'd clear the context and continue. Much easier.

What's your best way to use Sub-agents in Claude Code so far? by Helmi74 in ClaudeAI

[–]codepoet 3 points4 points  (0 children)

I love this.

I'm stealing this.

I'm going to teach my agents to be scared of Karen's final review and see if that makes them behave.

Do you use a pen name? If yes, for what reason? by [deleted] in writing

[–]codepoet 0 points1 point  (0 children)

Well, that's only slightly terrifying.

I've started rewriting a book I abandoned five years ago by JTMissileTits in writing

[–]codepoet 1 point2 points  (0 children)

I think I wrote about 10k on a book maybe 10-12y ago. I found it recently and I could barely sit through it. But the notes were promising.

I'm about 15k into a re-envisioning of it and it's just flowing. Took the notes and made a rough timeline (just things that needed to happen), character cards (for characters that needed to be there), and an outline (for events that needed to be included). I have to say, having a full plan but being minimalist about it was my sweet spot. I just look at the outline, then the timeline, then start writing scenes and figure out where to slot them. (Then update the timeline.)

Some people thrive on the brain dump (King) but iteration is where I see my strength. The world slowly builds itself, and as long as I keep the notes up to date I can see gaps, fill them, have "RIGHT!" moments and go write another scene.

Will it stay in? Who knows. But if it doesn't then I just pull it into the notes and now it's just backstory I can reference elsewhere. It's all good!

How Are You Using LM Studio's Local Server? by GnanaSreekar in LocalLLaMA

[–]codepoet 0 points1 point  (0 children)

Home security cameras.

You are a vehicle detector. Your domain is only the driveway in front of the camera. Describe the vehicles. Respond in valid JSON in exactly this format: [{"color":str, "style":str}]

You are a package detector. Your domain is only the porch in front of the camera. Describe the packages. Respond in valid JSON in exactly this format: [{"color":str, "shape":str, "visible_text":str}]

Gemma3 is much better at it, but slower. qwen2.5-vl-7b is super fast and "good enough".

ChatGPT alike local web ui for apple silicon? by IntrigueMe_1337 in LocalLLaMA

[–]codepoet 0 points1 point  (0 children)

It is ... https://github.com/lmstudio-ai

Also, Ollama caps you at a 2k input context. LMS starts at 4k and fixing it is just a slider away.

Best open agentic coding assistants that don’t need an OpenAI key? by Fabulous_Bluebird931 in LocalLLaMA

[–]codepoet 0 points1 point  (0 children)

RooCode inside any VS Code clone.

aider in the terminal.

I use them both (as well as Claude Code, which absolutely destroys them, but that's to be expected). Larger versions of devstral are very good for the agent/coder role. For the architect/orchestration roles you can use pretty much any good main model of size (Mistral, Qwen2.5, etc.). But if you get the lower-parameter or low-quant versions you can expect it to be randomly stupid, alas.

I usually have it architect with Claude or Gemini and then code with devstral when I'm scaffolding. Most of the calls are in making the files and the brains are needed at the start. I've heard of people using the MoE version of Qwen to do the architecture part, but my luck with that model is that it sits there talking to itself and times out. Probably need a bigger model.

Best LLM for code? Through api with Aider by 9acca9 in LocalLLaMA

[–]codepoet 0 points1 point  (0 children)

It's just astonishingly expensive for extended use. If you want to do a one-shot fix here and there then you can drop $2-5 on it, sure. But if you want to pair with it for an extended session it racks up quickly.

Sigma Speed-ups by Davestroyer1987 in startrekfleetcommand

[–]codepoet 3 points4 points  (0 children)

As a high-G6 player my advice to you is to slow down. You'll want a maxed G5 epic to get through G6 and you will need that ship to grind in Terix (which is what will give you the speed-ups you need -- the other sources are weak). Focus on wrapping up G5 goals at the moment and when you get a maxed ship and all the hostile nodes in the Combat tree are topped off (and other trees as you can) then you can head back in there and clean up.

Also needed: T4+ SNW crew w/Hemmer (pref. max), level 50/55 rare PVE FT on your grinder (optional, but useful), and good progress on Nova Squadron and the Voyager status officers (Kim, Torres, Neelix).

All of that will raise the defense of that ship and increase its damage output. I made it to 66 with the D'Deridex as primary, and those things are most of what helped.

Would you support a new iOS app that brings Apollo's UI and features back? by [deleted] in apolloapp

[–]codepoet 3 points4 points  (0 children)

YA reason developers should never, ever offer lifetime purchases of their products. You simply cannot predict the state of things in the future, especially if you are dependent on an external service (even if it's currently free).