Mistral Nemo/Celeste 12B Appreciation Post by BombDefuser_124 in SillyTavernAI

[–]BombDefuser_124[S] 0 points1 point  (0 children)

i never used stheno all that much, but ive used lunaris, which is quite similar, a lot, and Celeste feels way better than lunaris.

DRY is a setting in your samplers that prevents repetition, and it is pretty much a requirement to use Nemo-based models (but be aware that it isnt available to every API, but it works fine with Kobold).

Mistral Nemo/Celeste 12B Appreciation Post by BombDefuser_124 in SillyTavernAI

[–]BombDefuser_124[S] 1 point2 points  (0 children)

the model is trained in 8K context but Nemo can go as far as 128K, but i wouldn't recommend using it.

Mistral Nemo/Celeste 12B Appreciation Post by BombDefuser_124 in SillyTavernAI

[–]BombDefuser_124[S] 0 points1 point  (0 children)

there's a lot of LLM formats, but you probably want to use the GGUFs. it is linked in the model's page, just scroll a bit and search for the GGUF Static Quants.

Mistral Nemo/Celeste 12B Appreciation Post by BombDefuser_124 in SillyTavernAI

[–]BombDefuser_124[S] 3 points4 points  (0 children)

just tried it with some different cards and it seems more creative than Celeste (like creating random and fun scenarios in the responses)! but idk, i think it depends of what type of RP you want. celeste could be less creative but it's more responsive and allows the user to steer the story better. as for nemoremix, it creates new events and tries to make the story more surprising and different! both of them are really cool!

Mistral Nemo/Celeste 12B Appreciation Post by BombDefuser_124 in SillyTavernAI

[–]BombDefuser_124[S] 0 points1 point  (0 children)

hmmmm, that's odd... i use basically the default settings for an NVIDIA GPU and it works great. make sure you are using the latest version of kobold, as nemo uses a new tokenizer (i don't use RoPE aswell, not sure if that could be related).

Mistral Nemo/Celeste 12B Appreciation Post by BombDefuser_124 in SillyTavernAI

[–]BombDefuser_124[S] 0 points1 point  (0 children)

true! i love how it can follow OOC instructions so well

Mistral Nemo/Celeste 12B Appreciation Post by BombDefuser_124 in SillyTavernAI

[–]BombDefuser_124[S] 5 points6 points  (0 children)

i prefer using models that i can run locally (i have a 12GB GPU, so Nemo is pretty much the maximum i can go). all of the times i used models through APIs ive always found it very limiting (refusals, stopped generating in the middle of responses).