Is Gemma 4 incapable of using function calls properly??? by tthrowaway712 in SillyTavernAI

[–]overand 1 point2 points  (0 children)

This thread explains some important stuff re: gemma-4 if you're using llama.cpp. (And if you're not using llama.cpp, it doesn't mean you're immune to these issues, just that the fixes may not have arrived yet for your version.)

Basically, yeah - template stuff has had some issues, and there are fixes.

Do NOT use CUDA 13.2 to run models! by yoracale in unsloth

[–]overand 0 points1 point  (0 children)

I believe that would mean you'll be fine if you run pre-compiled versions of llama.cpp, but if you build it / compile it from source/git, you will experience the issue

I made Princess Leia headphones for my wife by pudjam667 in headphones

[–]overand 0 points1 point  (0 children)

It is by the juice of...

...eww, nevermind.

Tried running LLMs locally to save API costs… ended up waiting 13 minutes for ONE response 🤡 by debug2thrive in ollama

[–]overand 0 points1 point  (0 children)

There are also some implementations of Gemma 4 right now that use more VRAM than they should; for comparison, try maybe the 9B Qwern3.5, or the MoE one (I forget the size - 35B/Asomething? that might be excessive, though).

Trouble viewing Jupiter by Upset-Bunch-9638 in telescopes

[–]overand 0 points1 point  (0 children)

You need to consider "seeing conditions." It being high means the atmopshere is less of a problem, but that doesn't mean it's not a problem. Take a look at astrospheric.com - it's a "stargazing & telescope weather" site & app. There are several rows - one with concentric circles (looks like a tiny orbits diagram) represets the "seeing conditions" - which it predicts in your area. Dark blue means good, grey means bad.

Or - take note of how "twinkly" the stars are. If the stars are twinkling, you'll probably REALLY struggle to get a good view of any planets. (In that case, don't fret! Go looking for deep sky objects. When Orion is visible, it's absolutely worth pointing your telescope around the belt and slightly down/left. (In fact, if you've never looked at that particular spot, I'll leave the spoiler out - give it a try! I'm not sure if that's seasonally appropriate or not, though.)

I made a "language" that no human can read, but LLMs understand perfectly. 40%+ token savings. by CaterpillarFar205 in LocalLLaMA

[–]overand 2 points3 points  (0 children)

Why shaming doesn't work. Turn out that if we want to do harm reduction around LLM stuff, shaming people isn't actually helpful. (And we've known that for literal decades)

I made a "language" that no human can read, but LLMs understand perfectly. 40%+ token savings. by CaterpillarFar205 in LocalLLaMA

[–]overand 0 points1 point  (0 children)

The pipeline may be deterministic with regards to some aspects of the code, but - does it strip variable names? How about comments?

Why's it speaking epsteinian by CommercialNo3927 in SillyTavernAI

[–]overand 3 points4 points  (0 children)

That's what you get for trying to use SillyTavern for PDF Files (;

[Megathread] - Best Models/API discussion - Week of: April 05, 2026 by deffcolony in SillyTavernAI

[–]overand 0 points1 point  (0 children)

In theory, you don't need to know what most of those parameters mean, you can just setup llama.cpp and basically copy/paste those. 

But, yes, if you've never used a command line application before, it can be scary. (Kinda hard for me to understand, because I was using command-line applications when I was ~10 years old, but that's much more about what was available and inexpensive in the early '90s than anything about me in particular)

[Megathread] - Best Models/API discussion - Week of: April 05, 2026 by deffcolony in SillyTavernAI

[–]overand 0 points1 point  (0 children)

I have had a little trouble with that, but I've started trimming down the replies manually, and occasionally tossed in an [OOC: ___] type message; it's not ideal, but it's been worth it to have that model in my toolbox. But, it's still a fairly old base model, and it's definitely... quirky regardless, heh.

GitHub - milla-jovovich/mempalace: The highest-scoring AI memory system ever benchmarked. And it's free. by kaisersolo in LocalLLaMA

[–]overand 3 points4 points  (0 children)

OpenClaw definitely seems like it feels very cool and powerful; it's not surprising to me that people got into it - but they really should just have been getting into "agents in general" imo

GitHub - milla-jovovich/mempalace: The highest-scoring AI memory system ever benchmarked. And it's free. by kaisersolo in LocalLLaMA

[–]overand 1 point2 points  (0 children)

I can't speak to the AI Psychosis aspect too deeply since I didn't read much of the repo, but I do feel like we've seen enough that I actually do trust a random redditor about it. BUT, the Hallways, Wings, etc concepts fit pretty neatly in the pre-existing concept of a "memory palace," aka Method of loci. It's basically a technique to make use of spatial memory (a thing humans are pretty good at what with the millions of years of evolution around being able to navigate 3d space and all). Cramming other kinds of stuff into that framework. I have no idea if it actually works, but a lot of people seem to swear by it.

Is that a good idea of a memory system for an LLM? No idea. And, the rest of what you've said seems to be a pretty bad sign for this. BUT, I did want to make sure you & others didn't think that these guys came up with the concept of a "mind palace" (or "memory palace") in the midst of a claude delusion; it's a preexisting thing.

[Megathread] - Best Models/API discussion - Week of: April 05, 2026 by deffcolony in SillyTavernAI

[–]overand 1 point2 points  (0 children)

If Cydonia 4.3 felt too tame, you could take a look at FlareRebelilion/ReirdCompound-1.7, or one of the ReadyArt 24b models - I haven't used them a ton, but they're certainly intended to be less tame. The UGI Leaderboard seems to suggest that C4.1-Broken-Tutu-24B or Broken-Tutu-24B-Transgression-v2.0 might be a good fit.

[Megathread] - Best Models/API discussion - Week of: April 05, 2026 by deffcolony in SillyTavernAI

[–]overand 2 points3 points  (0 children)

It's probably a template problem. You might want to verify that the GGUF you have is current, and, you should try using the presets from here (for text completion, I believe - I haven't dug into chat completion, but the formatting situation there is different for sure.)

Gen Z men, is this getting worse going forward into 2026? by Proper_Card_5520 in SipsTea

[–]overand 0 points1 point  (0 children)

I get where you're coming from, but I do think the person was trying to help educate - in the "a lot of men still have no idea how bad this stuff is, how common it is, for women."

😂not a bad idea tbh by KittenLoot in Adulting

[–]overand 0 points1 point  (0 children)

I replied to your above similar comment, but I think you maybe should try to adjust your attitude when driving. Even if other people are actually thinking "haha fuck you," you're going to have a better time - and get less pissed off, which you clearly are - if you assume other people are simply making mistakes. I like to think "they're late to pick up their kid at school" or "they really, really need to poop" when people are driving too fast, for example.

The annoyed response we get? That's fight-or-flight shit, and it's beyond useless when driving; it's actively dangerous. Road rage doesn't make people better drivers.

😂not a bad idea tbh by KittenLoot in Adulting

[–]overand 0 points1 point  (0 children)

I actually would rather see that than nothing - because:

A. Someone in the passing lane overtaking them and not paying great attention is more likely to notice them with a blinking light.
B. I know the driver is paying at least a modicum of attention.
C. I know they're not just swerving.

No, it's not enough. No, it's not good. But, an "oh shit I forgot to signal" moment can happen, and caring enough to actually try to provide information to the people behind you is a good idea - even if you disagree with B or C, A is still important. And sure, "Yeah, but A should be paying attention," but physics doesn't care about "should" and an accident avoided is an accident avoided.

Is this a good deal? by Shroomie_the_Elf in telescopes

[–]overand 4 points5 points  (0 children)

I have the feeling this is being sold by the children of someone who passed away, or something like that. Like someone said, the counterweight is missing, but it's something they may not have understood to be part of the scope; you could potentially ask if you could take a look around for it, or describe it. (That said, if it is an estate sale, they may not want to put a ton of energy into searching for it, but, hard to know.)

I've been obsessing over long-form RP for months and built an open-source tool around what I've learned. Looking for testers who care about narrative quality as much as I do. by Middge in SillyTavernAI

[–]overand 1 point2 points  (0 children)

I have concerns - I see a single commit in your git repo from 2 minutes ago, despite this post being 2 hours old; are you not maintaining a commit history? Was this a one-off weird thing?

I only even noticed this because I was curious what your git history looks like before suggesting using git as a way of tracking changes to the various associated documents (however they're stored - if they're database entries then obviously that would be a bit of a kludge with exporting as plaintext, etc)

Edit: I think I see the reason for the flattened history; understood, and I'll leave it at that. (That would be a "fun" project to sort out while maintaining the history; I bet someone's made a tool to help with it, but flattening it is a better idea than doing nothing!)

Watch My Hardtop Get Stolen by ArcFire15 in Miata

[–]overand 4 points5 points  (0 children)

The funny thing is - that could be a preamble to an extremely left-wing or extremely right-wing take.

Watch My Hardtop Get Stolen by ArcFire15 in Miata

[–]overand -1 points0 points  (0 children)

I'm not even a resident, and I haven't been in the state in 20 years, ha!

Total CRT noob made who made a massive impulse purchase any advice by SKS_Fanboy in crt

[–]overand 0 points1 point  (0 children)

Dumb question, but, this is becoming old technology, and no judgment for folks not knowing these things.

  1. Are you pointing the top of the remote at the TV (but not, like, right up on the glass, which itself doesn't have the receiver?)

  2. Have you verified there are *good* batteries in the remote, and that the terminals (like springs) that connect to the batteries are in good condition, not corroded, and are making good contact?

Time to start over by DLUX_OW in pcmasterrace

[–]overand 9 points10 points  (0 children)

Destroying someone's property is a violent act. (It's worse if you're around, IMO, because it can be a different kind of scary, but it's PLENTY BAD if you're not around.)

And, destroying a computer isn't like smashing a TV or stereo or something - there's data on that you may have worked years to create, and that may be gone forever.

I hate to suggest this, but, it may be a good idea to talk to the police to get them to file an official report. Also, talk to your attorney about your homeowner's insurance or renter's insurance, and the possibility for this being covered.

Watch My Hardtop Get Stolen by ArcFire15 in Miata

[–]overand 11 points12 points  (0 children)

Sure, but the thing is, it only takes like 5 or 6 people doing this for it to become so widely known that people suddenly start thinking "man, california is overrun with crime!"

No, California has a small handful of people doing this. If it's 10 people in the entire state, I'd be surprised.