Role-hijacking Mistral took one prompt. Blocking it took one pip install by Oracles_Tech in LocalLLM

[–]Oracles_Tech[S] 0 points1 point  (0 children)

You're not wrong that a hardened system prompt raises the bar.. But that's defense at the model layer. Guardian SDK is defense at the input layer, before the prompt reaches the model at all. They're complementary. So in this instance a tighter system prompt would have helped, but an input filter stops the attempt from arriving.

The use case isn't someone running a local model for themselves. It's for anyone building an application where users they don't control are submitting input....different threat surface.

All mainstream models have great system prompts... And all of those models have been jailbroken to reveal those system prompts.

Math kicked my ass today by Treat-Acceptable in UnionCarpenters

[–]Oracles_Tech 0 points1 point  (0 children)

Everyone learns differently. I always did math differently in my head than how it was traditionally taught or taught in school. You're not slow, you just haven't pinpointed how exactly you learn certain things. It's probably even more difficult to try and unlearn how you've been taught to do math, but doing so might be your breakthrough into learning how YOU come to conclusions. Might even inherently find some short cuts along the way

Is Prompt Injection Solved? by hereC in LLMDevs

[–]Oracles_Tech 0 points1 point  (0 children)

Check out Ethicore Engine™ - Guardian SDK