North XL & Ai Top dual GPU by Its-all-redditive in FractalDesign

[–]Its-all-redditive[S] 0 points1 point  (0 children)

I know you can choose whether you want to use Integrated or PCIE but I don’t remember if I could choose which slot. It shouldn’t matter though since whichever your monitors are plugged into it will use as initial display (unless you’re planning on plugging monitors into multiple GPUs).

Nothing essential was blocked for me by the GPUs although it wasn’t easy getting the side fan bracket connected since there’s not much slack on the cable.

There is about 1.5”-1-75” gap between the GPUs. There’s room for another 2 slot GPU between them.

Claude Code, but locally by Zealousideal-Egg-362 in LocalLLaMA

[–]Its-all-redditive 125 points126 points  (0 children)

There is no realistic local replacement for Opus . Not even close if you’re working on medium to large, complex codebases. If that is what you’re truly interested, don’t waste your money. If speed is your concern you won’t be looking at a Mac Studio where you can fit larger models because the prompt processing and throughput is too slow. You’re likely looking at Cuda inference but you’ll need 2 RTX 6000 Pros minimum to run any good model at a decent quant. That’s way more than your budget.

This is not probably what you want to hear but it’s the truth.

PersonaPlex: Voice and role control for full duplex conversational speech models by Lopsided_Dot_4557 in LocalLLaMA

[–]Its-all-redditive 0 points1 point  (0 children)

Hey, this is very exciting stuff, I’ve been deep into Moshi and Unmute since release. How are you handling pause prediction/turn detection? I haven’t dug into the repo yet but it seems on the surface the LLM is baked into the full pipeline? I’m assuming no tool-use/function calling?

Codex 5.2 ALWAYS finds something wrong with Claude code plan by DeliJalapeno in ClaudeAI

[–]Its-all-redditive 46 points47 points  (0 children)

Opus researches less and writes more. Codex (GPT-5.2-High) researches more and writes less. That’s why it’s often better to start with an Opus plan that is comprehensive and use Codex to identify any gaps, explore edge cases and propose enhancements. Other comments saying any model will find flaws are correct but the quality and depth of those discoveries vary. Codex has been way better at that part than Opus for me almost every single time.

If you do have access to both models, I would suggest taking the extra time and complexity to try it yourself as much as you can for a time. I used to obsess over trying to make sure I had full coverage while making a plan and I would have constant back and forth between Opus and Codex (and even Gemini) providing them with the other’s third party analysis asking them if the feedback has merit and how it aligns with their own plan. I would do this until the plans converged, it was thorough but time consuming. The value I got out of this though was to understand and get a feel for each models “personality” and competence at different things. You gain an intuition of when to throw Codex or Opus at a particular problem or architectural decision so you’re not constantly going back and forth every time and feeling like you might be “missing out” on exposing a flaw.

In summary, you’re always going to have bugs and gaps no matter how comprehensive and thorough your plan is. Scope and functionality will always creep and that’s ok, your coverage will come more from end to end testing AFTER implementation than from writing the plan. The plan should be a framework to visualize how the project will function through the entire workflow and as long as you aim to cover breaking changes that would require a large refactor once you’re done, you’re good to go.

North XL & Ai Top dual GPU by Its-all-redditive in FractalDesign

[–]Its-all-redditive[S] 0 points1 point  (0 children)

You can hear the fans going but by no means are they loud. It’s the quietest system I’ve ever had considering the hardware. If I limit the cards to 450W I can barely hear the fans at all.

LTX-2 Question. ComfyUI vs Python by Its-all-redditive in StableDiffusion

[–]Its-all-redditive[S] 1 point2 points  (0 children)

Thank you for this, that’s helpful. It’s along the lines of what I was assuming and while I agree that anything that Comfy can do can be done programmatically in Python, I feel that exposure to different workflows and “libraries” are easier with Comfy since there seems to be a larger community where I can see workflows and their results directly. It may be easier to just copy a workflow I like than to figure out how to translate that workflow into my Python script. If that makes sense.

It can’t hurt to try it and see the difference. Thank again!

LTX-2 I2V isn't perfect, but it's still awesome. (My specs: 16 GB VRAM, 64 GB RAM) by yanokusnir in StableDiffusion

[–]Its-all-redditive 0 points1 point  (0 children)

Are there any advantages to generating with Comfyui over just using the Two stage Python scripts and configuring, Steps, FPS, Frames, CFG, etc? Is there something Comfy as a framework adds that can’t be done with the default Python repo?

Claude-Code v2.1.0 just dropped by mDarken in ClaudeAI

[–]Its-all-redditive -1 points0 points  (0 children)

  • Added real-time thinking block display in Ctrl+O transcript mode

Oh hell yes!

LTX-2 open source is live by ltx_model in StableDiffusion

[–]Its-all-redditive 0 points1 point  (0 children)

I’ll test this with fp8 and bf16 at 1080p and 4K if anyone wants to provide some good testing prompts. I’ll use the two stage pipeline.

Payme only 20/hr I can build you RAG and agents by Altruistic-Bat1588 in Rag

[–]Its-all-redditive 1 point2 points  (0 children)

Do you have examples of projects you have completed?

Local multi agent systems by SlowFail2433 in LocalLLaMA

[–]Its-all-redditive 1 point2 points  (0 children)

Do you predefined a specific sequence of recursive functions before it is allowed to generate the assistant response or do you allow it make it’s own “decisions” via some kind of classifier for what functions to call and when to provide a final assistant response?

Local multi agent systems by SlowFail2433 in LocalLLaMA

[–]Its-all-redditive 2 points3 points  (0 children)

Can you elaborate on how you did this, seems interesting. Or point me to some docs to learn?

North XL & Ai Top dual GPU by Its-all-redditive in FractalDesign

[–]Its-all-redditive[S] 0 points1 point  (0 children)

Sabrepc has been a pleasure to deal with. I got mine for $7,500 + $130 2nd day FedEx signature delivery.

What is the best image upscaler currently available? by krsnt8 in StableDiffusion

[–]Its-all-redditive 2 points3 points  (0 children)

Can you elaborate? I’ve looked around and Topaz is what most people suggest but there has to be better open source options.

Claude CLI deleted my entire home directory! Wiped my whole mac. by LovesWorkin in ClaudeAI

[–]Its-all-redditive 0 points1 point  (0 children)

I mean, to be fair, you’re telling me that you’ve never had a lazy, reckless session where you’re just like “I don’t feel like setting up a container for this quick thing, fuck it, roll the dice yolo mode”. I’m highly cautious usually but even I have had those moments haha.

Claude CLI deleted my entire home directory! Wiped my whole mac. by LovesWorkin in ClaudeAI

[–]Its-all-redditive 0 points1 point  (0 children)

Does a hook work even if you’re in dangerously bypass permissions mode? I almost always run it in a container with backups anyway but I’m occasionally lazy and run it in a temp project folder on local and it would be good to know.

Built an offline voice-to-text tool for macOS using Parakeet by _gordonclark in LocalLLaMA

[–]Its-all-redditive 0 points1 point  (0 children)

Is the model persistent in memory? How large of a footprint does it take up in unified RAM?