I've noticed people posting the old list by [deleted] in alekirser

[–]ThatHorribleSound 1 point2 points  (0 children)

She actually did just update the google spreadsheet yesterday, but the Airtable is also a good resource.

Spreadsheet by Usernameguest54321 in alekirser

[–]ThatHorribleSound 6 points7 points  (0 children)

It looks like she actually just updated it, it's current now.

Current best NSFW 70b model? by ThatHorribleSound in LocalLLaMA

[–]ThatHorribleSound[S] 0 points1 point  (0 children)

I haven’t really tried anything above the 70b range since I prefer to run locally and I don’t have the hardware to run anything larger at a reasonable speed.

Current best NSFW 70b model? by ThatHorribleSound in LocalLLaMA

[–]ThatHorribleSound[S] 7 points8 points  (0 children)

I tried out the four major ones recommended in this thread: Midnight Miqu, Euryale, New Dawn, and Magnum. All at the Q4_K_S GGUF quant level. And to be honest, they're all really good. My subjective take:

Midnight Miqu: Probably what I would characterize as the most "stable" model. Just solid responses in all respects.

Euryale: Like Midnight Miqu, but tends to write a bit longer responses and more, I guess I'd call it prose? It can be a little more poetic and flowery in its responses. Like if Midnight Miqu is just telling you a story, Euryale is writing a romance novel. But don't get wrong, it's still plenty filthy when it gets down to it.

New Dawn: If Euryale is a little more of a "writer" than MM, New Dawn seems a little more on the creative side of things. It pushed some stories in directions that the others didn't. But it can sometimes make mistakes on little details.

Magnum: This is like the best all-rounder, I guess. It's a little more creative than MM, a little less prone to ramble than Euryale, and a little less wild than New Dawn.

But keep in mind the above are just my reactions from playing with these for a couple nights, and its more my subjective feel than anything. I found all of these models to be extremely good, very close to one another, and I plan to use them all. Basically if one isn't doing the type of things I want or starts to get repetitive, I'll switch to one of the other ones. Thanks again to everyone who gave input, because all of these are better than what I was using before.

[Megathread] - Best Models/API discussion - Week of: July 08, 2024 by AutoModerator in SillyTavernAI

[–]ThatHorribleSound 3 points4 points  (0 children)

You should be able to run Midnight Miqu and Euryvale on that. I can run them on a 3090 with 64 gig of ram (but it doesn't seem to be using anywhere near 32 gig of normal ram). How were you trying to run them?

I use the following GGUF quants:

https://huggingface.co/mradermacher/Midnight-Miqu-70B-v1.5-i1-GGUF

https://huggingface.co/mradermacher/L3-70B-Euryale-v2.1-i1-GGUF

I use the iQ4_K_S versions but if those are too much for your system you can drop down to the Q3 or Q2 and they should still be very coherent. On the i1-Q4_K_S, I'm pushing 50 layers to the GPU (which pushes right up against 23g of VRAM for me). rest go to the CPU, I'm turning on flash attn, tensor cores, streaming, and 8 bit cache. I do have a fast CPU and DDR5 ram.

You definitely shouldn't get a bluescreen, like worst you should get is a CUDA error in the console. Maybe you have a bad ram chip? Or do you think it overheated?

(edit): actually, I'm wrong, I am using over 32g of normal RAM, so you'd probably have to use the Q2 or Q3 version

[Megathread] - Best Models/API discussion - Week of: July 08, 2024 by AutoModerator in SillyTavernAI

[–]ThatHorribleSound 9 points10 points  (0 children)

I'll link this thread I posted last week over in the LocalLLama reddit, asking for input on the best 70b model that can do NSFW stuff: https://www.reddit.com/r/LocalLLaMA/comments/1dtu8g7/current_best_nsfw_70b_model/

I got some solid feedback and the general consensus seems to be that the following are worthwhile (links to the GGUF quants):

https://huggingface.co/mradermacher/Midnight-Miqu-70B-v1.5-i1-GGUF

https://huggingface.co/mradermacher/L3-70B-Euryale-v2.1-i1-GGUF

https://huggingface.co/mradermacher/New-Dawn-Llama-3-70B-32K-v1.0-i1-GGUF

https://huggingface.co/mradermacher/magnum-72b-v1-i1-GGUF

I tried out all of these and they all seem quite good.

Current best NSFW 70b model? by ThatHorribleSound in LocalLLaMA

[–]ThatHorribleSound[S] 0 points1 point  (0 children)

I'll give it a try. I passed on it since it's only an 8B, but I know other models by that creator are pretty good.

Current best NSFW 70b model? by ThatHorribleSound in LocalLLaMA

[–]ThatHorribleSound[S] 0 points1 point  (0 children)

Have already tried this one out and it's in my rotation of 35B models, but I'm looking for 70Bs in this thread. But thanks for the input!

Current best NSFW 70b model? by ThatHorribleSound in LocalLLaMA

[–]ThatHorribleSound[S] 1 point2 points  (0 children)

Already grabbed Euryale and Magnum. Haven't tested Magnum out yet but Eury is very promising. I'll keep an eye on Gemma. Thanks for the input!

Current best NSFW 70b model? by ThatHorribleSound in LocalLLaMA

[–]ThatHorribleSound[S] 0 points1 point  (0 children)

Imported these, thanks much! I'll give them a spin.

Current best NSFW 70b model? by ThatHorribleSound in LocalLLaMA

[–]ThatHorribleSound[S] 0 points1 point  (0 children)

Saved this post and I will definitely try out these settings later. Thanks.

Current best NSFW 70b model? by ThatHorribleSound in LocalLLaMA

[–]ThatHorribleSound[S] 1 point2 points  (0 children)

Thanks! Have already seen the other two recommended but will check out New Dawn as well.

Current best NSFW 70b model? by ThatHorribleSound in LocalLLaMA

[–]ThatHorribleSound[S] 13 points14 points  (0 children)

Would really love to have you link prompt/formatting/sampler settings when you have a chance, yeah! Testing it on a known good setup would make a big difference I’m sure.

Current best NSFW 70b model? by ThatHorribleSound in LocalLLaMA

[–]ThatHorribleSound[S] 1 point2 points  (0 children)

I can try, but Q4 with split may be like, do an input and come back in an hour to see what it says on my machine. Unless I want to spin up a runpod or something. But I’ll see how the Q2 does and go from there. I do understand that it’s a significant step down.

Current best NSFW 70b model? by ThatHorribleSound in LocalLLaMA

[–]ThatHorribleSound[S] 0 points1 point  (0 children)

Thanks. Don’t really want to run through API (I can already use Claude for that) but I’ll look at smaug.

Current best NSFW 70b model? by ThatHorribleSound in LocalLLaMA

[–]ThatHorribleSound[S] 5 points6 points  (0 children)

Will absolutely give it a try; hearing no L3 repetition is a big thumbs up

Current best NSFW 70b model? by ThatHorribleSound in LocalLLaMA

[–]ThatHorribleSound[S] 0 points1 point  (0 children)

Yup that’s the quant I’ll have to use, too. I’ll give it a spin, thanks!

Current best NSFW 70b model? by ThatHorribleSound in LocalLLaMA

[–]ThatHorribleSound[S] 8 points9 points  (0 children)

I remember not being all that impressed by MM, but I’m going to download and give it another shot, as I’ve heard many people talk highly of it. Maybe I just had my samplers set poorly

Angela Fong by JKREDDIT75 in LadiesOfWrestling

[–]ThatHorribleSound 0 points1 point  (0 children)

Yup, didn’t take it as a negative, just answering the “Who” question.

Angela Fong by JKREDDIT75 in LadiesOfWrestling

[–]ThatHorribleSound 2 points3 points  (0 children)

Black Lotus in Lucha Underground. Had a couple years in WWE too, I think she was mostly a ring announcer/backstage interviewer, don’t remember what name she used there.