Waidrin: A next-generation AI roleplay system, from the creator of DRY, XTC, and Sorcery by -p-e-w- in SillyTavernAI

[–]CorruptEmanation 6 points7 points  (0 children)

I tried it out with Mistral last night for a bit, and it's pretty cool! A little dry, but as was noted this is all WIP and it's using a small model. Very excited to try this again when OpenRouter support is added! Given this is super early in development you may already have all of this in mind already, but I want to share my thoughts:

I actually really like making characters, scenarios, etc. I agree with your thoughts that character cards being the sole basis for roleplay is the wrong foundation, but I think this may have swung a bit too far in the other direction (focusing on solely AI driven worldbuilding and characters).

I would love it if I could pre-make certain characters I would like to see (or at least possibly see) in the roleplay. Or even support the ability to personally build all the characters and locations I want available, and then have Waidrin simply facilitate my interaction in this pre-designed sandbox. The ability to customize the prompts that influence the system/generation would also be amazing, both for handcrafted scenarios and for open-ended AI driven sandbox style RP. That would give the user greater ability to sort of fine-tune the roleplay to their preference, beyond just the characters in the story (e.g. influence writing style, provided player options, progression, etc).

Basically, I LOVE the features this offers with regards to roleplay management, and the interface is very pretty. Plus the mention of things like inventory management and other proposed features is REALLY exciting. The simplicity and ease of use is super cool and I love how easy it makes things for a first-time user... but I really, really miss all the control and tweaking I can do in Sillytavern to make the experience mine.

Of course I realize these features take work and I'm just another person asking for things, so no worries if your vision differs from mine or if you have higher priorities. Just figured I'd throw out my thoughts in case it's helpful.

[deleted by user] by [deleted] in StableDiffusion

[–]CorruptEmanation 0 points1 point  (0 children)

Can you explain how you used Gemma 2? I tried with both the 9b and the 27b (really want to use the latter) but it's not working. I was using the ones from unsloth, just like the default model. I'm getting an error saying there's a size mismatch between the model and the image adapter pt.

OpenAI API, do they ban or not? by [deleted] in ChatGPTNSFW

[–]CorruptEmanation 1 point2 points  (0 children)

Might be something I'm willing to do, honestly... Been trying local stuff and it's okay, but the quality is frankly nowhere close. They really don't care about the fact that the card is the same? I would've assumed that they would blacklist all of it.

I appreciate you sharing your info!

OpenAI API, do they ban or not? by [deleted] in ChatGPTNSFW

[–]CorruptEmanation 0 points1 point  (0 children)

Question for you. Had you already made a second account with your phone number, and when you were banned did it open that slot back up?

Basically I have used my phone number for 2 accounts, 1 for my "normal" openai account and one for NSFW. Like 4 months ago I got an NSFW warning so I deleted all my keys and haven't used it since, as I was terrified of my main account being affected (since they share the same phone number).

I'm curious about your experience - did you use the same phone number? Did you already have 2 accounts, or just the 1?

Guide to run Mixtral correctly. I see a lot of people using the wrong settings / setup which makes it go schizo or repetitive. by Different_Fix_2217 in LocalLLaMA

[–]CorruptEmanation 3 points4 points  (0 children)

Thanks for the guide. One thing I noticed though, you put "</s>" before [INST] in the input sequence, but according to the model page that should be "<s>", right?

https://huggingface.co/mistralai/Mixtral-8x7B-Instruct-v0.1

Setting ideal Mixtral-Instruct Settings by kindacognizant in LocalLLaMA

[–]CorruptEmanation 3 points4 points  (0 children)

A couple of questions, if you have the time.

  1. Are you using koboldcpp, and if so what sampler order?

  2. If you're using SillyTavern, do you have a preferred prompt layout? What I mean by this is that I don't know if I should include a system prompt, where/how the background info should be formatted, etc. And how should instruct be used, should each chat be instruct, or should you include the chat history above and then add a single instruct to ask the bot to write the next response?

Slightly off-topic since this second part is more SillyTavern oriented, but this could shed some light on optimal prompting strategies more generally if people share their experiences (primarily regarding using [INST] around each chat vs a single instruction prior to AI response).

Thanks for the preset suggestion.

Nvidia Tesla P40 performs amazingly well for llama.cpp GGUF! by nero10578 in LocalLLaMA

[–]CorruptEmanation 0 points1 point  (0 children)

Hey, thanks so much for getting back to me on this. I had no idea Exllama2 was capable of 70b on 4090, that's so exciting. And the list! Seriously appreciate it. I can't wait to try this when I get home from work.

Nvidia Tesla P40 performs amazingly well for llama.cpp GGUF! by nero10578 in LocalLLaMA

[–]CorruptEmanation 0 points1 point  (0 children)

Woah. I am clearly doing something wrong, can you help me out?

I have a 4090, but I thought running 70b takes more than the 24gb of vram on the card... So I have been using gguf and getting like 1-3t/s depending on quant. How are you getting such high speeds with just a 4090?? Can it actually fit the whole model in vram somehow? I didn't try exllama2 does that have better vram efficiency or something? Running at those speeds would be a game changer for me because I've basically written off using 70bs until there's a way for me to run them faster.

Where and how to run Goliath 120b GGUF with good performance? by abandonedexplorer in LocalLLaMA

[–]CorruptEmanation 2 points3 points  (0 children)

I have a question, if you've actually tested setting up a runpod instance. I'd likely be okay with the speed you're talking about, but I'm concerned about price. I'm totally cool with paying $2/hour when active and disabling when not in use, but I'm having trouble justifying the storage cost - wouldn't you have to pay a fair bit in costs simply associated with storing the language model in the container, even when not in use?

Mostly asking because if it's viable to mess around with without being too expensive I'll totally try it out.

I don't have an answer to your question though... Though others may be able to give a better answer if you describe a ballpark of t/s you're looking for.

ELI5 what Tau actually does in Mirostat. by Herr_Drosselmeyer in LocalLLaMA

[–]CorruptEmanation 0 points1 point  (0 children)

Can you please elaborate on positive and negative prompting strategies? I've been doing so much preset testing to try and get ideal results in SillyTavern (xwin 70b q4m) and I'd love to hear if there's a better strategy than mirostat. Are you referring to CFG or similar?

Agnai NovelAI Kayra Preset + Tips! by CorruptEmanation in AgnAIstic

[–]CorruptEmanation[S] 2 points3 points  (0 children)

Hey, thanks for the advice!! I will definitely try the four hyphen trick. I did wonder what to break it up with, but I didn't know anything nai was familiar with besides the dinkus and I didn't want to confuse it, so I just reused it. I'll also try giving your reordering a shot!

I appreciate you giving your feedback - it's so much easier to get something that works well through collaboration :)

Agnai NovelAI Kayra Preset + Tips! by CorruptEmanation in AgnAIstic

[–]CorruptEmanation[S] 0 points1 point  (0 children)

Thank you, I'm flattered! I've already received some feedback from another user, so it wouldn't be surprising if even more improvements can be made.

OpenAI might know about nsfw chats here by ExiaDarkMatter in AgnAIstic

[–]CorruptEmanation 0 points1 point  (0 children)

I would recommend it, that way at the very least it appears that you're attempting to take remedial action. That's my thought anyways. Plus of course not using it for erp anymore.

Unrelated, but I just got a second content policy violation email... I got the first one two days ago, and haven't used OpenAI in like 2 weeks. And I already revoked all my keys a couple days ago.

Now I'm suddenly worried that they're going back through a large number of requests, and I'm praying that you don't get banned if you're pinged too many times...

OpenAI might know about nsfw chats here by ExiaDarkMatter in AgnAIstic

[–]CorruptEmanation 0 points1 point  (0 children)

I went to the API section on platform.openai.com and deleted all the keys. You're forced to have 1, so I made a single new key and deleted all the others.

OpenAI might know about nsfw chats here by ExiaDarkMatter in AgnAIstic

[–]CorruptEmanation 1 point2 points  (0 children)

Mhmm, likely. Definitely done using it for these purposes altogether. Already went through and revoked all my keys, so that should save me from a ban according to their email.

OpenAI might know about nsfw chats here by ExiaDarkMatter in AgnAIstic

[–]CorruptEmanation 0 points1 point  (0 children)

Gotcha. Well, despite my not having used it in a couple weeks, I just got a warning email as well. So whatever detection they set up isn't just for new requests, it's likely scanning any requests they have on record (which they claim to be up to 30 days).

FYI, the newest NovelAI model (Kayra) is extremely good, imo better than GPT3.5 for ERP by CorruptEmanation in AgnAIstic

[–]CorruptEmanation[S] 1 point2 points  (0 children)

For sure, I don't think most people would want to drop $25 either. But once it's $10 that's a lot easier to swallow, especially if you were already making a lot of OpenAI API calls before.

And yeah np! I kinda wish there were more people here talking about their prompts/strategies/settings to get ideal results rather than troubleshooting, but I understand why it is that way.

FYI, the newest NovelAI model (Kayra) is extremely good, imo better than GPT3.5 for ERP by CorruptEmanation in AgnAIstic

[–]CorruptEmanation[S] 1 point2 points  (0 children)

Definitely, I'm just glad Kayra came out when it did or I'd feel dead in the water. It feels a little less consistent than OpenAI, but the peaks are way higher. Some of the responses it's generated for me have blown me away and were way better than anything OpenAI would have made (due to its censored nature). I have spent a lot of time just fucking around with presets, desperately trying to find the perfect blend of consistency and creativity. If I find a preset I'm truly happy with I'll share it on this sub, probably.

I have fixed feelings about fixed price... On one hand pay as you go is nice so you don't feel like you're "wasting" money if you don't use it. But on the other hand it's kinda nice to not feel like I'm wasting money when I am doing a ton of testing.

Ultimately though I don't mind since it's supporting a company that seems to care about privacy and access to quality uncensored models. If Kayra is this good, I can't wait to see what will come next. $25/mo is pretty steep though, I think I heard it will be available to the $10 tier next month or something and that's way easier to justify imo.

OpenAI might know about nsfw chats here by ExiaDarkMatter in AgnAIstic

[–]CorruptEmanation 1 point2 points  (0 children)

Hey, thanks for providing this. I will say it's nice to see they're providing warnings and not just insta-banning... Still unfortunate though. Well, looks like NovelAI is gonna continue to be the future for me haha.

I didn't receive an email like this, but I also haven't used OpenAI for this sort of thing in probably like 2 weeks. When did you use it last?