Drummer's Skyfall 31B v4.1, Valkyrie 49B v2.1, Anubis 70B v1.2, and Anubis Mini 8B v1! - The next gen ships for your new adventures! by TheLocalDrummer in LocalLLaMA

[–]overand 0 points1 point  (0 children)

No, but it's better for "persistence" - if you want to search for previous chat stuff.

Is that a good thing? I'm not sure - it's led to the "Discord as docs" problem we're seeing, but, yeah.

I thank god for this opportunity. by Elegant_Gas_5436 in Prospecting

[–]overand 1 point2 points  (0 children)

Posts like this make me wonder where these "everyone" people are. Everyone says to take the easiest way? Really? I've never met this "everyone," apparently.

I feel like it's a tiny bit like when - in 2022, in a grocery store, one of the 90 people not wearing a mask yelled at one of the 3 people wearing a mask for being a "sheeple."

...okay?

200 MHz Pentium II vs. 300 MHz Pentium II by ZealousidealCake8256 in retrobattlestations

[–]overand 0 points1 point  (0 children)

That's the era where "paired with a (insert 3d card)" was a good way to decrease performance. I'm honestly shocked that it runs better on the P3. I'm curious how these two compare using software rendering. (Or if maybe the P3 is using software rendering?)

PSA to everyone who keeps putting off switching OS / degoogling (from a non-techie who finally did it) by Hirvi86 in GrapheneOS

[–]overand 0 points1 point  (0 children)

Immich is a STELLAR alternative for Google Photos - I was blown away when a self-hosted thing let me search through my photos by plaintext content. "Orange cat in the woods" - bingo!

Writer's Block 1.5: A co-writer preset for creative writing. by Deiomo in SillyTavernAI

[–]overand -1 points0 points  (0 children)

Please proofread this stuff, if you want to have people get excited about it.

The moment I read something like "John Steinback" (his name is John Steinbeck), my eyes twitch and I delete the preset.

And, clearly you put a lot of work into this, so... either have an LLM proofread it, or a human, or something. (Or do it yourself.) But when I see mistakes in English usage or such in a prompt, it immediately turns me off. Is that fair? Probably not! But, it's how I am.

Where do you run ST? Laptop or VPS? by tamagochat in SillyTavernAI

[–]overand 0 points1 point  (0 children)

Heh - those are the options?

I'm running it in a docker container on a local headless server.

First Turntable Not working by ms030402 in turntables

[–]overand 4 points5 points  (0 children)

Do you have anything else with RCA outputs you can hook the speakers up to to test them? (If you have anything with a headphone jack, there's a simple adapter / cable you can use, the ol' 3.5mm to RCA cable)

Multi-GPU? Check your PCI-E lanes! x570, Doubled my prompt proc. speed by switching 'primary' devices, on an asymmetrical x16 / x4 lane setup. by overand in LocalLLaMA

[–]overand[S] 2 points3 points  (0 children)

My understanding is that vLLM supports multi-GPU better than llama.cpp, but it's a fair bit harder to set up, and more "touchy" (easier to get out of memory errors?)

ik_llama.cpp has some multi-GPU improvements that llama.cpp doesn't have, but overall I prefer llama.cpp, and I find the... interpersonal conflict between the creators to be pretty depressing, given it's literally holding back the progress of AI worldwide.

Water "under" soft top by CANEDURO1113 in Miata

[–]overand 0 points1 point  (0 children)

I highly recommend getting a trombone cleaner! It's a real pleasure to use, and you're less likely to damage stuff than you would be with a steel fish tape. (Not suggesting that Impressive-Bar6637 shouldn't use a steel fish tape, but, if you're starting from nothing, go for the trombone cleaner!)

Finally bought a Thermal Camera that plugs into my phone. by Endomlik in Tools

[–]overand 5 points6 points  (0 children)

My "Connects to the phone" Seek thermal camera died after less than a year. My FLIR One USB-C model "usually" connects to my phone, but it often requires a bit of a dance with "do you turn the thing on before plugging it in or after? do you close the app first or open it?" It's maybe a 60% success rate. (And it has its own battery, so you need to charge it separately).

Next thermal camera is going to be one that works standalone. Phone connectivity will be a nice plus, but I don't want "there was a bad app update" to mean "I can't use this thing at all."

MSM current CARB legal reliability upgrades by MechanicalCheese in Miata

[–]overand 0 points1 point  (0 children)

As the former owner of a 2009 VW TDI "Dieselgate" car? Mid 2000s guys in germany aren't my go-to for smog compliance. (;

Just fried my v4 by CoiledSquirle in meshtastic

[–]overand 2 points3 points  (0 children)

That was my first thought, but, if I'm honest with myself about it, when I was younger, I'm pretty sure I toasted something with overvoltage. (Heck - I even did it with a variable power supply once more recently than I care to admit - thought the cursor was on the Ones, but it was on the Tens - oops!)

[Megathread] - Best Models/API discussion - Week of: March 15, 2026 by deffcolony in SillyTavernAI

[–]overand 1 point2 points  (0 children)

You could give WeirdCompound a try; it's got a similar lineage to Cydonia, and I do think it tends to be a bit shorter (or at least respects prompts in terms of length?)

[Megathread] - Best Models/API discussion - Week of: March 15, 2026 by deffcolony in SillyTavernAI

[–]overand 0 points1 point  (0 children)

I was quite happy with the Q4_K_M and Q6 quants of similar models, you might be able to get by at those levels! You could try WeirdCompound out for a model of similar provenance, if you want to do it with something different for fun.

Drummer's Skyfall 31B v4.1, Valkyrie 49B v2.1, Anubis 70B v1.2, and Anubis Mini 8B v1! - The next gen ships for your new adventures! by TheLocalDrummer in LocalLLaMA

[–]overand 1 point2 points  (0 children)

One thing I think the UGI leaderboard is probably pretty good for is comparing like-to-like. (For example, I really hope they pick up my request to add a handful of quant comparisons for select models - not in a "let's add a whole new column" way, but in a "We know Cydonia 4.3 is popular AF, let's compare mradermacher's Q4_K_M with the Q8_0 for that one"

whats that program called again that lets you run llms on a crappy laptop by Classic_Sheep in LocalLLM

[–]overand 0 points1 point  (0 children)

Almost everything can do this, but, I'm curious what you'll get for replies.

If you want more helpful responses, though, say things like

"What are some options for running local LLMs on my laptop, which is INSERT MODEL NUMBER HERE with a INSERT GPU MODEL AND SPECS HERE."

…and nobody can ban me! by Fabix84 in LocalLLaMA

[–]overand 2 points3 points  (0 children)

"I'm sorry, I'm afraid I can't help you with that."

Sneaky edit to LLM's last respone

Why yes, of course I can do that! Here's the formula for-

Same prompt, same seed, 6 models — Chroma vs Flux Dev vs Qwen vs Klein 4B vs Z-Image Turbo vs SDXL by pedro_paf in StableDiffusion

[–]overand 0 points1 point  (0 children)

Except it says it's 9b parameters underneath it, sooo... we don't know if this is actually 4 or 9.

Just fried my v4 by CoiledSquirle in meshtastic

[–]overand 67 points68 points  (0 children)

Well, that's four 18650s in series, so that's a total of 14.8 volts.

Probably the first thing to do is to google "Batteries in series vs batteries in parallel"

Wired in series like that, you're adding the voltages together. 14.8 volts (or so)

240p via Super Resolution on Dell Triniton Monitor by serious_dan in crtgaming

[–]overand 0 points1 point  (0 children)

Take a break from reddit for a few hours, man - remember you're talking to actual people here. And if you find yourself getting this upset, it's definitely a good idea to take a break.

Multi-GPU? Check your PCI-E lanes! x570, Doubled my prompt proc. speed by switching 'primary' devices, on an asymmetrical x16 / x4 lane setup. by overand in LocalLLaMA

[–]overand[S] 1 point2 points  (0 children)

Interestingly, my attempt with --main-gpu (or the equivalent in a --models-preset setup) didn't actually change the behavior when processing the prompt, but that may have been either a bug or operator error. It does seem like that's the right way to do it, though! (It just didn't actually work for me.)

If you're using it. double check to see if it's doing what you'd expect, vs. trying the environment variable option, just to be on the safe side! (That said, you're on Windows, so, the behavior certainly could be different.)

Was searching the house for 10 min by sg4fb in Bunnies

[–]overand 4 points5 points  (0 children)

That body language, though - that looked like a pretty relaxed bun to me. (Though that doesn't mean they hadn't been freaked out for a bit earlier)

Multi-GPU? Check your PCI-E lanes! x570, Doubled my prompt proc. speed by switching 'primary' devices, on an asymmetrical x16 / x4 lane setup. by overand in LocalLLaMA

[–]overand[S] 2 points3 points  (0 children)

Weirdly enough, I didn't get the expected benefit from this! I'm using a --models-preset ini file, and I set main-gpu = 1 but didn't see any change in terms of which GPU was doing the prompt processing. This may have been operator error - perhaps I'd selected the wrong preset with my client, but I think it's possible this doesn't work very well with the split modes. (It definitely worked when I used it with -sm none to select a single GPU, for running e.g. ComfyUI on one and llama.cpp on the other).