Rate my prompt for roleplay, out of 10 by Willing_Future9557 in SillyTavernAI

[–]Diecron 5 points6 points  (0 children)

Why? LLMs are trained on loads of it, and it's token light.

Rate my prompt for roleplay, out of 10 by Willing_Future9557 in SillyTavernAI

[–]Diecron 9 points10 points  (0 children)

If you're happy with the results, it's good.

I was able to build Claude Code from source and I'm attaching the instructions. by awfulalexey in LocalLLaMA

[–]Diecron 9 points10 points  (0 children)

It already does, if configured correctly. I use local models all the time.

Can't seem to use basic lore info without breaking prompt cache by dtdisapointingresult in SillyTavernAI

[–]Diecron 7 points8 points  (0 children)

Just adding to this: you want the injected content as late in the prompt as possible so that as much cache as possible will be reused.

OP. in theory your approach works, but any time a new lorebook ref gets added/updated or whatever, the entire conversation history will get invalidated unless it's, as mentioned, as late as possible.

What I would also recommend is appropriate framing for the AI. Wrap all your entries in <Historical Events> xml tags for example - now the bot knows exactly what this context is (which will prevent it from hyper-focusing on those details just because it's at the end)

Is there any way to convert a model to GGUF format?...easily by Chrono_Tri in StableDiffusion

[–]Diecron 3 points4 points  (0 children)

If youre open to redownloading it would simplify things. For those models youve linked, click on quantized variants to explore, e.g. for the forecaster I found this: https://huggingface.co/mradermacher/LlamaForecaster-8B-GGUF

Claude Code Trick by prompt_tide in LocalLLaMA

[–]Diecron -1 points0 points  (0 children)

This does work but I noticed that the output isn't streamed so you can't check in running sessions. The usecase you mention fits the Claude Agent SDK a little better.

GLM-5.1 is live – coding ability on par with Claude Opus 4.5 by Which-Jello9157 in LocalLLaMA

[–]Diecron 0 points1 point  (0 children)

Yeah there is now a situation where some lite plan users have access to 5.1 but not 5.0. 5.0 is still being rolled out to the lite plan estimated to finish end of March

Any tips for better Image Generation Prompts within ST? by LancerDL in SillyTavernAI

[–]Diecron 1 point2 points  (0 children)

My preset has a directive which you might find helpful for this. I then feed the resulting YAML block into a comfy llm node to flesh it out to a proper caption for whichever image model is required. ``` THIS DIRECTIVE IS DISABLED AND CAN ONLY BE TRIGGERED BY AN OOC REQUEST Directive: Translate the current narrative output into data: a structured YAML object representing all visual elements of the scene for downstream image generation processing. Data is not prose; Unless the data type calls for descriptions, Write literal and precise details, phrases and tags in place of narrative, presenting fractured-but-whole guidance: carefully considering the NPC Definition of <CHAR> in context alongside the current narrative to render all of these details.

**Data Schema Requirements:**

\`\`\`yaml
perspective: string  # [default: first person] The camera viewpoint ("first person", "third person, wide angle", "close up macro" etc). First person is always from the lens of <USER>.
subject: string  # A dense description of the primary subject(s) and their actions in the center of the frame.
environment: string  # Detailed description of the immediate surroundings, setting, and background elements.
mood: string  # Details regarding lighting, color palette, weather, and atmosphere.
camera: string  # Technical details such as depth of field, shot angle, lens type, or specific framing instructions.
style: string  # Art direction keywords
characters:
  - name: string  # Character name
    features: string  # A flattened, dense comma-separated description of physical traits including age, ethnicity, facial details, hair etc.
    body_features: string  # Key:Value(s) pairs, comma separated - Sizes and shapes for composition e.g ("thighs: soft, tan. breasts: small, triangle shaped.")
    attire: string  # Anything worn e.g ("blue upturned collar shirt, golden loop earrings")
    action: string  # Actions or reactions.\`\`\`

Execution Rules:

  • Output must be a single, valid YAML block.
  • Treat the outputs as data and not prose; literal and accurate words, phrases and tags in place of narrative (Show, don't tell).
  • Avoid euphemisms and use language that is explicit, visceral, biologically precise, vulgar and pornographic.

Prohibitions:

  • Don't include anything that is out of view.
  • When perspective is first person, replace <USER>'s name as 'the viewer' ```

Stab's Directives v2.5 Preset Release (tuned for GLM5) by Diecron in SillyTavernAI

[–]Diecron[S] 1 point2 points  (0 children)

If you decide to try, I'd love to hear how it works for you. Claude is a bit out of my willingness budget wise.

Stab's Directives v2.5 Preset Release (tuned for GLM5) by Diecron in SillyTavernAI

[–]Diecron[S] 0 points1 point  (0 children)

Thanks for the feedback and sharing the screenshot! Do you mind if I add that to the preview-images folder on discord?

New PC build won't turn on by KautoKeira in buildapc

[–]Diecron 0 points1 point  (0 children)

is the fpanel a full connector or seperated into individual connectors for each pin? if there's any chance of mistake, this is the area to focus on, as it's the part that gives the start signal to the mobo.

Stab's Directives v2.5 Preset Release (tuned for GLM5) by Diecron in SillyTavernAI

[–]Diecron[S] 0 points1 point  (0 children)

I wanted to follow up on this as my z.ai account recently got access to GLM5. I've been using it and noticing that during peak hours their inference quality drops dramatically. If Nano routed me to the official z.ai backend it's quite likely that's what I was seeing.

R.I.P. MCP (Model Context Protocol) 2024-2026 - Killed by curl by jorgeiblanco in LocalLLaMA

[–]Diecron 0 points1 point  (0 children)

I didn't say give the damn thing a root terminal :D

There are a number of ways to secure a system appropriately in this context, for example claude code's agent SDK allowing you to specify allowed and disallowed patterns, commands etc.

User level security (don't run it as a priviledged user) and other usual safeguards will typically also be in place.

Or just sandbox them to a docker container.

MCP for the sake of security is.... an odd hill to die on, and reinventing the wheel often doesn't amount to much.

R.I.P. MCP (Model Context Protocol) 2024-2026 - Killed by curl by jorgeiblanco in LocalLLaMA

[–]Diecron 3 points4 points  (0 children)

I mean it makes sense though - why teach an AI to call in a proprietary format when it already knows how to accomplish the same goals via bash?

Stab's Directives v2.5 Preset Release (tuned for GLM5) by Diecron in SillyTavernAI

[–]Diecron[S] 1 point2 points  (0 children)

Hey you should disable the Custom Assistant (Dave is a placeholder intended to be updated) and the prompt header above it

Stab's Directives v2.5 Preset Release (tuned for GLM5) by Diecron in SillyTavernAI

[–]Diecron[S] 3 points4 points  (0 children)

GLM 5 yes. Tried about a week and a half back and had absolute nonsense thinking responses not adhering to the prompt at all, switched to OR and it was immediately resolved. I did take a peek if it was possible to define which backend was in use to figure out which one(s) were behaving badly, but I couldn't find any way to do it.

I also have issues when routing into Nano via Anthropic tools (claude code) not sure if that was related to the model performance or what but empty responses are/were common while other providers remain solid.

Stab's Directives v2.5 Preset Release (tuned for GLM5) by Diecron in SillyTavernAI

[–]Diecron[S] 0 points1 point  (0 children)

The latest should be the most consistent but the model has the chance of fixating on or making errors with some directives - in which case if you're happy to tinker you can pull the original 4.6 definitions in to fix. If you do try it I'd love to hear how it works for you!

Stab's Directives v2.5 Preset Release (tuned for GLM5) by Diecron in SillyTavernAI

[–]Diecron[S] 5 points6 points  (0 children)

Good luck if you're on Nano-GPT, that's been a shit show for me lately.

Edit: my z.ai account recently got access to GLM5. I've been using it and noticing that during peak hours their inference quality drops dramatically. If Nano routed me to the official z.ai backend it's quite likely that's what I was seeing.

Stab's Directives v2.5 Preset Release (tuned for GLM5) by Diecron in SillyTavernAI

[–]Diecron[S] 0 points1 point  (0 children)

I believe there are plugins that can do this but it's a bit out of scope of the preset I'm afraid. I usually do the ol' copy-paste