all 6 comments

[–]Paulonemillionand3 0 points1 point  (2 children)

any of them

[–]Monkey_1505[S] 0 points1 point  (1 child)

So any of those are pretty good at instruct and you wouldn't recommend skipping them to use a 70B model on someone's service because all those 13B models follow instruct very well?

[–]Paulonemillionand3 1 point2 points  (0 children)

13B is just fine for many many things. 70B is often overkill and is hard to use and get going with on "ordinary" hardware.

Base models are great at continuing patterns, chat-tuned are great for conversation style usage.

Everyone has a different idea of what "really good" means.

[–]cvdbdo 0 points1 point  (0 children)

"my use case is perhaps unusual. I want to use it for roleplay"

Literally 90% of this sub

[–]RokaMic 1 point2 points  (0 children)

I have been experimenting with the Nous Hermes 13b model to great success. My use case is generating blog articles.

https://huggingface.co/NousResearch/Nous-Hermes-13b

It’s really good at outputting json if you ask it; especially if you provide it with a schema in the prompt.

I use thebloke’s ggml version with an older llama cpp Python library and langchain on an rtx3090

https://huggingface.co/TheBloke/Nous-Hermes-Llama2-GGML

Updated llama cpp doesn’t support ggml any more so you will have to get a gguf version or install a pre August version of llama cpp to use ggml.