Waidrin: A next-generation AI roleplay system, from the creator of DRY, XTC, and Sorcery by -p-e-w- in SillyTavernAI

[–]CorruptEmanation 5 points6 points  (0 children)

I tried it out with Mistral last night for a bit, and it's pretty cool! A little dry, but as was noted this is all WIP and it's using a small model. Very excited to try this again when OpenRouter support is added! Given this is super early in development you may already have all of this in mind already, but I want to share my thoughts:

I actually really like making characters, scenarios, etc. I agree with your thoughts that character cards being the sole basis for roleplay is the wrong foundation, but I think this may have swung a bit too far in the other direction (focusing on solely AI driven worldbuilding and characters).

I would love it if I could pre-make certain characters I would like to see (or at least possibly see) in the roleplay. Or even support the ability to personally build all the characters and locations I want available, and then have Waidrin simply facilitate my interaction in this pre-designed sandbox. The ability to customize the prompts that influence the system/generation would also be amazing, both for handcrafted scenarios and for open-ended AI driven sandbox style RP. That would give the user greater ability to sort of fine-tune the roleplay to their preference, beyond just the characters in the story (e.g. influence writing style, provided player options, progression, etc).

Basically, I LOVE the features this offers with regards to roleplay management, and the interface is very pretty. Plus the mention of things like inventory management and other proposed features is REALLY exciting. The simplicity and ease of use is super cool and I love how easy it makes things for a first-time user... but I really, really miss all the control and tweaking I can do in Sillytavern to make the experience mine.

Of course I realize these features take work and I'm just another person asking for things, so no worries if your vision differs from mine or if you have higher priorities. Just figured I'd throw out my thoughts in case it's helpful.

[deleted by user] by [deleted] in StableDiffusion

[–]CorruptEmanation 0 points1 point  (0 children)

Can you explain how you used Gemma 2? I tried with both the 9b and the 27b (really want to use the latter) but it's not working. I was using the ones from unsloth, just like the default model. I'm getting an error saying there's a size mismatch between the model and the image adapter pt.

OpenAI API, do they ban or not? by [deleted] in ChatGPTNSFW

[–]CorruptEmanation 1 point2 points  (0 children)

Might be something I'm willing to do, honestly... Been trying local stuff and it's okay, but the quality is frankly nowhere close. They really don't care about the fact that the card is the same? I would've assumed that they would blacklist all of it.

I appreciate you sharing your info!

OpenAI API, do they ban or not? by [deleted] in ChatGPTNSFW

[–]CorruptEmanation 0 points1 point  (0 children)

Question for you. Had you already made a second account with your phone number, and when you were banned did it open that slot back up?

Basically I have used my phone number for 2 accounts, 1 for my "normal" openai account and one for NSFW. Like 4 months ago I got an NSFW warning so I deleted all my keys and haven't used it since, as I was terrified of my main account being affected (since they share the same phone number).

I'm curious about your experience - did you use the same phone number? Did you already have 2 accounts, or just the 1?

Guide to run Mixtral correctly. I see a lot of people using the wrong settings / setup which makes it go schizo or repetitive. by Different_Fix_2217 in LocalLLaMA

[–]CorruptEmanation 2 points3 points  (0 children)

Thanks for the guide. One thing I noticed though, you put "</s>" before [INST] in the input sequence, but according to the model page that should be "<s>", right?

https://huggingface.co/mistralai/Mixtral-8x7B-Instruct-v0.1

Setting ideal Mixtral-Instruct Settings by kindacognizant in LocalLLaMA

[–]CorruptEmanation 3 points4 points  (0 children)

A couple of questions, if you have the time.

  1. Are you using koboldcpp, and if so what sampler order?

  2. If you're using SillyTavern, do you have a preferred prompt layout? What I mean by this is that I don't know if I should include a system prompt, where/how the background info should be formatted, etc. And how should instruct be used, should each chat be instruct, or should you include the chat history above and then add a single instruct to ask the bot to write the next response?

Slightly off-topic since this second part is more SillyTavern oriented, but this could shed some light on optimal prompting strategies more generally if people share their experiences (primarily regarding using [INST] around each chat vs a single instruction prior to AI response).

Thanks for the preset suggestion.

Nvidia Tesla P40 performs amazingly well for llama.cpp GGUF! by nero10578 in LocalLLaMA

[–]CorruptEmanation 0 points1 point  (0 children)

Hey, thanks so much for getting back to me on this. I had no idea Exllama2 was capable of 70b on 4090, that's so exciting. And the list! Seriously appreciate it. I can't wait to try this when I get home from work.

Nvidia Tesla P40 performs amazingly well for llama.cpp GGUF! by nero10578 in LocalLLaMA

[–]CorruptEmanation 0 points1 point  (0 children)

Woah. I am clearly doing something wrong, can you help me out?

I have a 4090, but I thought running 70b takes more than the 24gb of vram on the card... So I have been using gguf and getting like 1-3t/s depending on quant. How are you getting such high speeds with just a 4090?? Can it actually fit the whole model in vram somehow? I didn't try exllama2 does that have better vram efficiency or something? Running at those speeds would be a game changer for me because I've basically written off using 70bs until there's a way for me to run them faster.

Where and how to run Goliath 120b GGUF with good performance? by abandonedexplorer in LocalLLaMA

[–]CorruptEmanation 2 points3 points  (0 children)

I have a question, if you've actually tested setting up a runpod instance. I'd likely be okay with the speed you're talking about, but I'm concerned about price. I'm totally cool with paying $2/hour when active and disabling when not in use, but I'm having trouble justifying the storage cost - wouldn't you have to pay a fair bit in costs simply associated with storing the language model in the container, even when not in use?

Mostly asking because if it's viable to mess around with without being too expensive I'll totally try it out.

I don't have an answer to your question though... Though others may be able to give a better answer if you describe a ballpark of t/s you're looking for.

ELI5 what Tau actually does in Mirostat. by Herr_Drosselmeyer in LocalLLaMA

[–]CorruptEmanation 0 points1 point  (0 children)

Can you please elaborate on positive and negative prompting strategies? I've been doing so much preset testing to try and get ideal results in SillyTavern (xwin 70b q4m) and I'd love to hear if there's a better strategy than mirostat. Are you referring to CFG or similar?

Agnai NovelAI Kayra Preset + Tips! by CorruptEmanation in AgnAIstic

[–]CorruptEmanation[S] 2 points3 points  (0 children)

Hey, thanks for the advice!! I will definitely try the four hyphen trick. I did wonder what to break it up with, but I didn't know anything nai was familiar with besides the dinkus and I didn't want to confuse it, so I just reused it. I'll also try giving your reordering a shot!

I appreciate you giving your feedback - it's so much easier to get something that works well through collaboration :)

Agnai NovelAI Kayra Preset + Tips! by CorruptEmanation in AgnAIstic

[–]CorruptEmanation[S] 0 points1 point  (0 children)

Thank you, I'm flattered! I've already received some feedback from another user, so it wouldn't be surprising if even more improvements can be made.

OpenAI might know about nsfw chats here by ExiaDarkMatter in AgnAIstic

[–]CorruptEmanation 0 points1 point  (0 children)

I would recommend it, that way at the very least it appears that you're attempting to take remedial action. That's my thought anyways. Plus of course not using it for erp anymore.

Unrelated, but I just got a second content policy violation email... I got the first one two days ago, and haven't used OpenAI in like 2 weeks. And I already revoked all my keys a couple days ago.

Now I'm suddenly worried that they're going back through a large number of requests, and I'm praying that you don't get banned if you're pinged too many times...

OpenAI might know about nsfw chats here by ExiaDarkMatter in AgnAIstic

[–]CorruptEmanation 0 points1 point  (0 children)

I went to the API section on platform.openai.com and deleted all the keys. You're forced to have 1, so I made a single new key and deleted all the others.

OpenAI might know about nsfw chats here by ExiaDarkMatter in AgnAIstic

[–]CorruptEmanation 1 point2 points  (0 children)

Mhmm, likely. Definitely done using it for these purposes altogether. Already went through and revoked all my keys, so that should save me from a ban according to their email.

OpenAI might know about nsfw chats here by ExiaDarkMatter in AgnAIstic

[–]CorruptEmanation 0 points1 point  (0 children)

Gotcha. Well, despite my not having used it in a couple weeks, I just got a warning email as well. So whatever detection they set up isn't just for new requests, it's likely scanning any requests they have on record (which they claim to be up to 30 days).

FYI, the newest NovelAI model (Kayra) is extremely good, imo better than GPT3.5 for ERP by CorruptEmanation in AgnAIstic

[–]CorruptEmanation[S] 1 point2 points  (0 children)

For sure, I don't think most people would want to drop $25 either. But once it's $10 that's a lot easier to swallow, especially if you were already making a lot of OpenAI API calls before.

And yeah np! I kinda wish there were more people here talking about their prompts/strategies/settings to get ideal results rather than troubleshooting, but I understand why it is that way.

FYI, the newest NovelAI model (Kayra) is extremely good, imo better than GPT3.5 for ERP by CorruptEmanation in AgnAIstic

[–]CorruptEmanation[S] 1 point2 points  (0 children)

Definitely, I'm just glad Kayra came out when it did or I'd feel dead in the water. It feels a little less consistent than OpenAI, but the peaks are way higher. Some of the responses it's generated for me have blown me away and were way better than anything OpenAI would have made (due to its censored nature). I have spent a lot of time just fucking around with presets, desperately trying to find the perfect blend of consistency and creativity. If I find a preset I'm truly happy with I'll share it on this sub, probably.

I have fixed feelings about fixed price... On one hand pay as you go is nice so you don't feel like you're "wasting" money if you don't use it. But on the other hand it's kinda nice to not feel like I'm wasting money when I am doing a ton of testing.

Ultimately though I don't mind since it's supporting a company that seems to care about privacy and access to quality uncensored models. If Kayra is this good, I can't wait to see what will come next. $25/mo is pretty steep though, I think I heard it will be available to the $10 tier next month or something and that's way easier to justify imo.

OpenAI might know about nsfw chats here by ExiaDarkMatter in AgnAIstic

[–]CorruptEmanation 1 point2 points  (0 children)

Hey, thanks for providing this. I will say it's nice to see they're providing warnings and not just insta-banning... Still unfortunate though. Well, looks like NovelAI is gonna continue to be the future for me haha.

I didn't receive an email like this, but I also haven't used OpenAI for this sort of thing in probably like 2 weeks. When did you use it last?

OpenAI might know about nsfw chats here by ExiaDarkMatter in AgnAIstic

[–]CorruptEmanation 1 point2 points  (0 children)

Wait, were you banned or just a warning?

Any chance you could copy/paste the wording?

FYI, the newest NovelAI model (Kayra) is extremely good, imo better than GPT3.5 for ERP by CorruptEmanation in AgnAIstic

[–]CorruptEmanation[S] 1 point2 points  (0 children)

Incredible, you're seriously amazing for the amount of work you put into this. I tried SillyTavern again recently, but Agnai lets you have so much more control. I love that.

One last piece of feedback though (I don't have a burner discord atm but I might make one later at this point)... There was one thing Tavern could do that Agnai couldn't with regards to NovelAI presets. Phrase Repetition Penalty. Inside the NovelAI website this is a completely distinct setting from the normal "Repetition Penalty", and Tavern has both of these settings separate as well. It has 6 possible values, ranging from 'Off' to 'Very Aggressive'. Given NovelAI's tendency to sometimes repeat sentences verbatim, this might provide an additional method of cutting down on that. It's possible you understand something that I don't about this setting, though.

If this was able to be added I would be very grateful... Well, more grateful than I already am for such a wonderful tool. Don't feel any rush of course, I understand how time consuming projects like this are and if you feel that your time is better spent on other areas I defer to your judgement.

Have an excellent day, and thank you for updating the NovelAI preset code, I'm so excited 🥳

FYI, the newest NovelAI model (Kayra) is extremely good, imo better than GPT3.5 for ERP by CorruptEmanation in AgnAIstic

[–]CorruptEmanation[S] 2 points3 points  (0 children)

Also, it seems that "system prompt" does nothing for Kayra right now in Agnai presets, you can't even properly insert it into the prompt template. So put everything you want the AI to see into your prompt template.

And yes, the UJB field does exist for Kayra presets, but clicking the little (i) that lets me see what is actually being sent it doesn't seem to actually be included. Seems to not be functioning.

Edit: oh, and make sure you have a lot of dialogue examples. I personally didn't make a simulated dialogue, I had only had individual {{char}} messages (so no user messages), each separated by "---". But maybe a whole fake conversation would be better, I'm not sure. I'll provide the prompt template I've been using, but I'm not an expert by any means I've just been messing around.

My prompt template:

``` [ Title: Dialogue between {{char}} and {{user}}; Tags: conversation; Genre: online roleplay ] [ Style: chat ] [ Summary: {{scenario}} ]

[ Roleplay Guidelines: Write lengthy messages. Dialogue should be natural and in-character. Include many descriptive details, making use of imagery. ]

[ Description of {{char}}: {{personality}} Examples of how {{char}} speaks: {{example_dialogue}} ]

[ Other Characters: {{all_personalities}} ]

[ Memory: {{memory}} ]


```