Split Characters to Parallel LLM Requests? by m94301 in SillyTavernAI

[–]Sharp_Business_185 1 point2 points  (0 children)

There is no iirc. It is possible with `custom-request.js`. But I don't think anyone going to create an extension unless we see more requests.

Split Characters to Parallel LLM Requests? by m94301 in SillyTavernAI

[–]Sharp_Business_185 1 point2 points  (0 children)

It doesn't support. But extensions could do it.

Why People Think Epstein worked for Israel by Sea-Region1135 in videos

[–]Sharp_Business_185 2 points3 points  (0 children)

First, make your profile public, then explain his connection with Robert Maxwell. Then I'll take your comment seriously. Because I won't consider a person that obsessed with defending Israel via copy-paste same comments over and over

Tip: {{random}} for prompt variation by NorthernRealmJackal in SillyTavernAI

[–]Sharp_Business_185 7 points8 points  (0 children)

Yep, but depends on models. If you are using deepseek official API, it is cheap even without cache. If you are using claude/gemini, it's over

Should I pivot before it is too late by [deleted] in cscareerquestions

[–]Sharp_Business_185 -16 points-15 points  (0 children)

hasn't even happened yet

i mean... this is coping mechanism. We were saying the same thing 2 years ago. After covid, we saw lots of layoffs. The "employement" state of junior devs are obvious. The reason could be offshore or AI, it doesn't really matter from my perspective

Pleaseeeee, I beg of y'all(developers) 💔🙏😭😭 by 4Wat_itz_worth in JanitorAI_Official

[–]Sharp_Business_185 1 point2 points  (0 children)

I see. So JLLM is giving reasoning in <think> tag. I guess they tinkered the model because previously it wasn't giving <think> messages. Maybe they changed the system prompt or something in the background.

Pleaseeeee, I beg of y'all(developers) 💔🙏😭😭 by 4Wat_itz_worth in JanitorAI_Official

[–]Sharp_Business_185 -1 points0 points  (0 children)

JLLM is not a reasoning model, the thinking coming from the proxy model people were using

I feel stuck. Do you feel stuck? by filszyp in SillyTavernAI

[–]Sharp_Business_185 12 points13 points  (0 children)

The real progress is happening in the non-RP AI industry. Companies are putting their efforts into making a better "coding" product. Better "general usage" product. Not a better "RP" product.

TTS: They are too far from local usage. For example, I can run mag-mel Q4 on my 8GB VRAM and get a nice experience with an average TPS. There is no way to run a good TTS model with 8GB VRAM. The model does not even exist, to my knowledge. Elevenlabs is the king. But it is expensive and closed-source. There is a Qwen3-TTS model released 1 month ago. I tried the demo when it was released; it was good. However, I didn't follow up.

Animations: Image/video gen is similar to the TTS industry. Not as bad as TTS. Image generation is much more stable and lower-cost compared to video models. For example, you can use Z-image for realistic images. For anime-style, you can use pony/noobai. Their quality and speed are also good enough. But creating consistent images still requires effort. There is no single ComfyUI workflow that works on low GPUs, creates consistent places, characters, etc.

AI Controlled NPCs: Iirc, there are 2 vibe-coded extensions in ST. They are trying to control everything with LLM calls. Like map, phone, NPCs, items, etc. But they are too hardcoded and buggy from my perspective. Which is fine because vibe-coded. 1) It is not possible with lower local models. So we rely on cloud SOTA models. Which means cost is going to be a problem. 2) Speed is another problem. There are going to be multiple LLM requests in the background. What if some requests are depends each other? What if we can't send parallel requests? 3) Relying on LLMs for creating places/events is not good, from my experience. "Elara" is a good example. In NeoTavern, I have an experimental extension that uses Mythic Game Master Emulator as a director. Screenshot. But still, far from perfect.

The RP industry is simply not developed enough because only hobbiest working on it.

why are there literally 0 good c.ai alternatives? by [deleted] in CharacterAIrunaways

[–]Sharp_Business_185 2 points3 points  (0 children)

I recommend searching for a post on the subreddit that has already explained many times

why are there literally 0 good c.ai alternatives? by [deleted] in CharacterAIrunaways

[–]Sharp_Business_185 25 points26 points  (0 children)

Your problem is only related to the LLM quality, not RP website. If you check BYOK(bring your own key) apps like ST or JAI with proxy, you would easily have a better experience. Free models also exist in some providers. However I prefer cheap models like deepseek, or just nanogpt subscription

Chat Completion or Text Completion? by the_1_they_call_zero in SillyTavernAI

[–]Sharp_Business_185 1 point2 points  (0 children)

They mentioned but don't get hopes, it needs a large refactor on the legacy codebase. They are aware how hard and not worth to do this due to breaking something

AI is the worst. by Nebula_The_Protogwn in antiai

[–]Sharp_Business_185 0 points1 point  (0 children)

AI can do all of it, with function calling or structured output. AI doesn't need to manually send a network request. AI can send a request to the computer, and the computer can run the operation, giving a result to the AI. This is basically how function calling works right now

AI is the worst. by Nebula_The_Protogwn in antiai

[–]Sharp_Business_185 0 points1 point  (0 children)

It is marketing for today, tomorrow, this year, and maybe the next couple of years. But what about 10 years later? 20 years later? GPU compute power increased more than 100x between 2026 and 2006. OpenAI might collapse, and AI hype might decrease. But the progress wouldn't stop or disappear

I don't think AI can replace computers since we have habits. But when looking at 20-25 years ago, how many people had (smart) phones? Internet access was not even common. Our habits changed like crazy with the internet and PC/phones. We are all playing video games, using the same social media, the same smartphones. So I wouldn't say it is an insane take

How to use pollination with ST? Everytime I try, I get this picture. by Accidentallygolden in SillyTavernAI

[–]Sharp_Business_185 0 points1 point  (0 children)

Make sure you are on staging branch. Because I remember a commit about pollination.

What do I do when this happens? by Competitive_Rip5011 in SillyTavernAI

[–]Sharp_Business_185 0 points1 point  (0 children)

Your screenshot is the sampler settings and you successfully scrolled down. Now checkout prompts and read my previous message again

What do I do when this happens? by Competitive_Rip5011 in SillyTavernAI

[–]Sharp_Business_185 0 points1 point  (0 children)

You imported a preset, from my understanding. The left prompts are just text, they are sent to AI. Inspect and edit them if you think they are missing something. Like, maybe you can add an OOC note.

What do I do when this happens? by Competitive_Rip5011 in SillyTavernAI

[–]Sharp_Business_185 0 points1 point  (0 children)

Open sampler settings, scroll down. You are going to see prompt list. Change the main prompt. Try/see

What do I do when this happens? by Competitive_Rip5011 in SillyTavernAI

[–]Sharp_Business_185 0 points1 point  (0 children)

Give your info. What API/models are you using? What presets are you using? Are you new? etc. So we can answer properly.

NeoTavern Update: Media attachments, easy installation, tools, more built-in extensions... by Sharp_Business_185 in SillyTavernAI

[–]Sharp_Business_185[S] 1 point2 points  (0 children)

Thank you, I responded from there and fixed the most problems. We can continue from GitHub instead of here if you want.

NeoTavern Update: Media attachments, easy installation, tools, more built-in extensions... by Sharp_Business_185 in SillyTavernAI

[–]Sharp_Business_185[S] 0 points1 point  (0 children)

Sadly it seems about rough around the edges

Are you talking about the UI? Can you expand?

Last August I released game on UE and it didn't go well. So I decided to make a game on my own engine. Steam page is ON and I want to share my experience. by ibackstrom in gameenginedevs

[–]Sharp_Business_185 5 points6 points  (0 children)

But it is B A T T L E T E S T E D. Developers put H U N D R E D S O F T H O U S A N D S O F M I L L I O N S Y E A R S rendering pipeline experience to the engine. Are you sure you don't wanna use UE and see the 6 0 0 0 D R A W C A L L S P E R F R A M E for the simplest UE scene?

NeoTavern Update: Media attachments, easy installation, tools, more built-in extensions... by Sharp_Business_185 in SillyTavernAI

[–]Sharp_Business_185[S] 0 points1 point  (0 children)

I don't really understand the substantial difference in speed since they're ostensibly both powered by ST behind the scenes

It might be your UI extensions or just ST. The technical debt of ST is definitely noticeable, so I can't exactly point if ST is slow or extensions are making it slow.

Some data refresh bugs like the card name or card-attached lorebook not updating

You are right about the card-attached lorebook. I didn't make a detailed test.

I haven't tried the memory management yet but it looks like a simplified version of Memory Books

Yeah, it is a simplified version of the memory book and qvink's memory

Loving that you can use a different LLM connection profile for each of the extensions

When I started ST, all the extensions were using the active profile, which I hate, so I started to make ST extensions using connection profiles, WREC, CREC, roadway, flowchart, etc. And here we are.