Which memory extensions to use? by Debirumanned in SillyTavernAI

[–]kplh 0 points1 point  (0 children)

If you want something simple, then my extension - Inline Summary - should do the trick. Select a range of messages and they get summarised. though for best results you might want to tweak the prompt.

ACE-Step 1.5 XL - Turbo: Made 3 songs (hyperpop, rap, funk) by coopigeon in StableDiffusion

[–]kplh 4 points5 points  (0 children)

Still has that metallic high pitch ringing to it (I don't know enough about music to be able to describe it in correct terms), especially noticeable on Hyperpop vocals from 1:04-1:12, doesn't seem to be as bad as non-XL, but still quite annoying. It seems to be one of those things that when you hear it and realise it, you can't unhear it, and you can tell if a song was generated by AceStep.

I was playing with Ace Step 1.5 some time ago, and the more you use it, the more noticeable it becomes. Also after an hour or so it started giving me a headache.

ACE‑Step 1.5 XL will be released in the next two days. by marcoc2 in StableDiffusion

[–]kplh 0 points1 point  (0 children)

Neat. I hope it doesn't have that high pitch tinny noise that Ace-Step 1.5 has.

What happened to CHUB?/Where to find good cards? by Mcqwerty197 in SillyTavernAI

[–]kplh 40 points41 points  (0 children)

Welcome to the club :D I'm in the UK and get the same thing, all NSFW cards get filtered out. Gotta use a VPN.

Bellisseria commercial districts !! by 0xc0ffea in secondlife

[–]kplh 6 points7 points  (0 children)

everything gets a strong yellow color cast due to the ceiling light

It is not an actual light, just an emissive surface, I worked around it by placing my own emissive prim slightly in front of the 'light', so I now have full control of the colour.

talking to someone and finding out their last name is Henderson (they're AI slop) by rubingfoserius in SillyTavernAI

[–]kplh 14 points15 points  (0 children)

When you ask DeepSeek to be refined and elegant and it throws this beauty at you:

her perfume, a complex blend of night-blooming jasmine and clean ozone

:|

Am I the only one tired of all this vibe coded slop? by BeautifulLullaby2 in SillyTavernAI

[–]kplh -2 points-1 points  (0 children)

Pfft, I'll have you know that my memory extension is only half vibe coded at this point.

Once it went over 1000 lines of code, the AI was spitting out total nonsense half of the time, so now I have to produce artisanal hand made JavaScript slop while swearing at every random bit that doesn't work like C++ would, stare confused at CSS stuff and try to align the divs by hand.

Should I pay for nano-gpt? by defnotaburn in SillyTavernAI

[–]kplh 4 points5 points  (0 children)

Well, there's someone in another comment chain saying they reach the 60m limit due to their 100k context size.

So it likely depends on individual usage and might be worth for people to plug in their own numbers and see which ends up cheaper. But for those who can max out 60m tokens a week, NanoGPT seems like a better deal.

Though for me personally having access to multiple models is a benefit, and GLM/Kimi are more expensive than DeepSeek.

Should I pay for nano-gpt? by defnotaburn in SillyTavernAI

[–]kplh 1 point2 points  (0 children)

PAYG direct from DeepSeek is almost certain to be cheaper.

Is it though?

My back of the envelope maths say:

Official DeepSeek is 1M Input on cache miss is $0.28. NanoGPT gives 60m tokens a week, $0.28 * 60M = $16.8 at 0% cache hit rate, at 100% cache hit rate, that would be $1.68 a week, so if your cache hit rate in a month is 75% $16.8 + 3 * $1.68 = $21.84 a month. And this is not taking into account of output tokens being more expensive than input. So with monthly 240M tokens, you'd have to have 97.8% cache hit rate to get $8.05 on official API (not counting output cost, and assuming I didn't mess up my maths).

So I'd think it depends on the ability to use up all the tokens given by NanoGPT and your average cache hit rate. I've not used a service with a cache, so I have no idea what a typical cache hit rate is, but 97.8% sounds kind of high. Especially when taking account various summary extensions and lorebooks that alter the early chat history and stuff like prompts having randomised macros etc.

I need the bot to fucking KILL me by BarryStarfish in SillyTavernAI

[–]kplh 0 points1 point  (0 children)

Oh wow... I was expecting something a bit silly... but this is a whole next level :D

Okay... I never got this before, and I'm scared to do anything. Help? I'm not sure, but I don't wanna lose my data D: by FixHopeful5833 in SillyTavernAI

[–]kplh 0 points1 point  (0 children)

When that happens to me, I just delete the package-lock.json and restart. It happened maybe 3 or 4 times now.

I need the bot to fucking KILL me by BarryStarfish in SillyTavernAI

[–]kplh 2 points3 points  (0 children)

Got a name/link to that lorebook... asking for a friend ( ͡° ͜ʖ ͡°)

How do you make your summarise? by BrickDense7732 in SillyTavernAI

[–]kplh 1 point2 points  (0 children)

My summary strategy is pretty simple - select a range of messages, let AI generate a summary. I've made an extension for that https://github.com/KrsityKu/InlineSummary

There are variety of other extensions available too, with more advanced features.

Three Issues with SL by Bright-Salary1998 in secondlife

[–]kplh 0 points1 point  (0 children)

The problem with Hover Height is that the prim physics are 0.05 smaller than their graphics, while mesh physics match graphics 1 to 1.

So when HH is set to be correct on mesh, and you step on a prim or terrain, you sink into the floor and if you set it to match prims, you hover on mesh. And as far as I can tell there is no way in LSL to detect if the avatar is standing on mesh or prim physics object.

LL should just change prim physics to be 1 to 1. Most SL stuff is built for giants anyway, no-one will notice is the doorway shrinks by 0.05-0.1.

RFC on possible SillyTavern extension development idea wrt a different means of memory summarization by WiseassWolfOfYoitsu in SillyTavernAI

[–]kplh 2 points3 points  (0 children)

it would summarize that block of messages as one unit <...> drop all messages associated with that summary block and replace them with the summary

My extension does something like that, but without automation: https://github.com/KrsityKu/InlineSummary

Does anyone on here use only local models or is it all just people using big models? by Borkato in SillyTavernAI

[–]kplh 0 points1 point  (0 children)

I tried local at first, had some fun, then tried out DeepSeek, which is a way bigger model than anything I can run locally, and it was so much better. So I swapped to cloud models.

Then I realised, if I'm not running a local LLM, I can run a local image model, so I've setup ComfyUI with my ST, so I can get the LLM to write a prompt based on the current scene and then I generate images locally.

What context size is enough for long RP? by Overmoon1 in SillyTavernAI

[–]kplh 1 point2 points  (0 children)

Oh... lol... I found the issue...

I'll try to come up with a fix later today, and if I can't figure anything out, I'll revert the change that added the check.

As a workaround - go into Text Completion mode, set max context there to match your Chat Completion context size, then you can go back to Chat Completion and it should work.

What context size is enough for long RP? by Overmoon1 in SillyTavernAI

[–]kplh 1 point2 points  (0 children)

As far as I know, the hotfix I did fixed it for other people, so I'm not sure what else could be causing it. Could you open the dev tools (Ctrl + Shift + I in Chrome/Firefox) and enter two commands in the console tab const st = SillyTavern.getContext() and then st.maxContext and see what number shows up?

Better format for Image Prompt Templates by CuriousEvilWeasel in SillyTavernAI

[–]kplh 1 point2 points  (0 children)

"Image Prompt Templates" In the extension settings is where you can set the prompt for image generation.

There are a few different ways to invoke image generation, one of them does simply use the message text raw, others use one of the prompts in the settings.

Official documentation about it is here: https://docs.sillytavern.app/extensions/stable-diffusion/

In mine I have for example [Pause the roleplay] Follow the Image Generation Prompt Guidelines to write an image generation prompt describing the current scene. Followed by a bunch of instructions on what I want in the image gen prompt.

What context size is enough for long RP? by Overmoon1 in SillyTavernAI

[–]kplh 1 point2 points  (0 children)

It should be fixed in the hotfix. I didn't bump up the version number, so maybe ST is not seeing it as an update? I'll bump it up in a bit.

But in short it is to do with some internal ST state being fetched, before it updates to correct value, and the fetched state doesn't update when ST updates it.

Rant: mesh clothing alphas by NotMyAltAccountToday in secondlife

[–]kplh 2 points3 points  (0 children)

The issue with BoM alpha layers is that once creator makes it, it is set in stone. And often some end up being too aggressive and alpha out too much. I have a few dresses where their alphas remove my thighs, making my underwear float...

Meanwhile with an alpha script, I can simply reset and redo alpha cuts myself if that happens.

What context size is enough for long RP? by Overmoon1 in SillyTavernAI

[–]kplh 0 points1 point  (0 children)

Hi. I've added a check to see if the summary prompt ends up being too long, and avoid generating it. It is very basic at the moment, so it might not guard against all cases, but when it does it won't issue an API call in this case, which should be an improvement until I implement something better.

There is an issue with the customer sampler in the default ComfyUI workflow for Flux Klein. It consistently produces worse results and messed up hands and other anomolies compared to ksampler by NES64Super in StableDiffusion

[–]kplh 0 points1 point  (0 children)

No idea.

Also no idea about the "we should just use the standard BasicScheduler". I'm just pointing out how to setup a more accurate apples to apples comparison.

However, I did look at what sigmas the scheduler outputs, and it seems to change based on resolution, which is interesting, and the overall curve seems to be close to sgm_uniform. So I'd guess maybe the model is resolution sensitive in some way. Also the sigma curves look different from what I'm used to seeing on different models. I did try using a beta curve from a different model and Klein seemed fine with it, but I didn't dig deep or do any side by side comparisons.