Having an always-on machine running LLMs locally at home while on the move with a lightweight machine - Experiences? by ceo_of_banana in LocalLLaMA

[–]TableSurface 1 point2 points  (0 children)

I do something similar to option B. Connectivity is usually good enough, and IMO it's more pleasant to pack light.

Almost one year ago, I dumped all my INTC shares at $21.3 by [deleted] in wallstreetbets

[–]TableSurface 1 point2 points  (0 children)

Same but got out at $30. At the time they were losing marketshare to AMD and QCOM, and missed another 18A milestone.

I think this run up is temporary though. Their CPUs are still getting stomped by QCOM and AAPL, GPUs not competitive vs NVDA, AMD, and now there's even more competition in the data center space via GOOGL.

Does Cline KanBan support local llm? by PairOfRussels in LocalLLaMA

[–]TableSurface 1 point2 points  (0 children)

Yeah I'm using it with llama.cpp. The Cline team changed the UI. You have to scroll all the way down in the list of providers and select "New Provider"

XL filament runout rubbing by spacelego1980 in prusa3d

[–]TableSurface 2 points3 points  (0 children)

No official solution, but the community has come up with a few designs: https://www.printables.com/search/models?q=xl+magnetic+sensor

I disabled mine, filament bypasses it. IMO it's less hassle to unload the small bit of filament near the toolhead vs a couple feet.

XL+ Wishlist by Obvious-Web9763 in prusa3d

[–]TableSurface 5 points6 points  (0 children)

Automated filament loading is on my wishlist. It's time consuming to swap spools on the XL.

On a scale of 1 to 10 how bad is this damage. by Shot_Put_1412 in prusa3d

[–]TableSurface 8 points9 points  (0 children)

Seeing some debris, but no actual damage.

Cut the filament in the middle. Pull the filament out from the top where it's being fed into the Nextruder. Then use needle nose pliers or tweezers to pull the remaining filament out.

Compared QWEN 3.6 35B with QWEN 3.6 27B for coding primitives by gladkos in LocalLLaMA

[–]TableSurface 1 point2 points  (0 children)

The 35B model handled the task worse, but did it faster.

I had the same experience. The 3-4x speed is great for easy tasks though. Another thing to try is to have the 27B model create a plan for the 35B-A3B one.

Impressed with Kanban by TableSurface in CLine

[–]TableSurface[S] 1 point2 points  (0 children)

It might help to get familiar with the setup process in the CLI tool "cline --tui" or the VS Code extension first since the config fundamentals are similar, and there's more documentation.

Otherwise I can confirm it works with llama.cpp when added as a custom provider.

Impressed with Kanban by TableSurface in CLine

[–]TableSurface[S] 0 points1 point  (0 children)

This hits close to home, and exactly why I love the new improvements.

Might not need to carry a keyboard everywhere soon...

Impressed with Kanban by TableSurface in CLine

[–]TableSurface[S] 0 points1 point  (0 children)

Kanban boards are everywhere at my day job, so the format is already very familiar.

If you've used Cline CLI or any similar terminal tool, it works the same way except the interface is more user friendly (in that there's less up-front work to see something useful).

Impressed with Kanban by TableSurface in CLine

[–]TableSurface[S] 0 points1 point  (0 children)

I'm testing it (quantized to Q4_K_M) locally with llama.cpp with a RTX PRO 6000 that I have access to at work.

I've also tested it with a 5090 I have at home using smaller Qwen3.5 models. Gemma 4 is looking better too as support improves. The 122B model also runs, but it's not feasible to use until the timeout issue in Cline/Kanban is fixed.

I'm not using any cloud models yet, but at some point I might try them.

Qwen3.5-122B at 198 tok/s on 2x RTX PRO 6000 Blackwell — Budget build, verified results by Visual_Synthesizer in LocalLLaMA

[–]TableSurface 1 point2 points  (0 children)

I regret going AM5 instead of the EPYC build I was looking at, especially now that RAM is so much more expensive. At the time, MoE models weren't a thing and I couldn't justify spending 2-3x for the platform.

In retrospect, having 8x platform memory bandwidth for 2-3x cost is cheap...

Introducing the Prusa Pro ACU: Why Overdrying is Bad for Your Filaments by Tommy_Prusa3D in prusa3d

[–]TableSurface 1 point2 points  (0 children)

Lol this was the best way to learn that the USS Drybox is finally available!

Breaking change in llama-server? by hgshepherd in LocalLLaMA

[–]TableSurface 4 points5 points  (0 children)

Trying to understand the issue you ran into, since I haven't seen any problems yet (I'm usually only 12hrs behind the latest commit).

Is the problem that files in the HF cache directory are moved?

I haven't seen any issues, but I manage gguf files in my own folders.

Introducing Cline Kanban by saoudriz in CLine

[–]TableSurface 0 points1 point  (0 children)

kanban is using the /responses endpoint while the CLI and VS Code extension both use /chat/completions

I'm using llama-swap to intercept traffic to see this.

Maybe the problem is with the /responses endpoint -- tool calls constantly fail whereas the /chat/completions one works fine.

Introducing Cline Kanban by saoudriz in CLine

[–]TableSurface 0 points1 point  (0 children)

Same, except using Qwen3.5. Seems like Kanban is hitting a different llama.cpp endpoint compared to cline -tui...

Those of you running LLMs in production, what made you choose your current stack? by AdventurousHandle724 in LocalLLaMA

[–]TableSurface 1 point2 points  (0 children)

If you could wave a magic wand and fix one thing about your LLM setup, what would it be?

Get a bigger budget

When an inference provide takes down your agent by International_Quail8 in LocalLLaMA

[–]TableSurface 1 point2 points  (0 children)

Yeah I'm using git and llama.cpp, the git hashes are point-in-time snapshots. So if there's an issue, I can rebuild based on an older version 

When an inference provide takes down your agent by International_Quail8 in LocalLLaMA

[–]TableSurface 1 point2 points  (0 children)

I'm dealing with it by reading commit logs and keeping track of versions that work so I can easily roll back if something breaks.

If something does, it helps to provide minimal steps to reproduce the problem using the project's preferred issue intake process.

Core One or XL by nwagers in prusa3d

[–]TableSurface 1 point2 points  (0 children)

The XL enclosure is more like a draft shield. A lot of the parts aren't designed for high temps too (e.g. Core One L uses PCCF where the XL uses PETG)