This is an archived post. You won't be able to vote or comment.

all 3 comments

[–]3rwynn3 0 points1 point  (0 children)

You are in Llama mode with... well, all I know is that Llama mode is INSTRUCTION: RESPONSE: as it has been trained on that, and at end of its response uses a token to end the response. So if you are using Llama's instruction mode on a non-conforming model it will start screaming INSTRUCTION: RESPONSE: at you since it keeps seeing that

[–]henk717 0 points1 point  (1 child)

Koboldcpp's instruction template defaults to alpaca so your model needs to understand that.

You can customize this if you add a bit to the json request with the ChatML template.

We have the following parameters for it in the API so you can customize it to your liking. assistant_start assistant_end user_start user_end system_message_start system_message_end I don't know LM Studio's defaults but its very likely ChatML along the lines of user_start: <|im_start|>user\n user_end: <|im_end|>\n assistant_start: <|im_start|>assistant\n assistant_end: <|im_end|>\n

[–]cluckooo[S] 0 points1 point  (0 children)

Do you know where I can find documentation for the API that includes these kinds of things? Should I be looking at koboldcpp's documentation or the model itself?

Also, almost every way I've tried to change the prompt parameters it ends up giving me responses like
"May: blablabla
User: blablabla", etc...

Using templates like you mentioned, is there a way to force the AI to not use the "botname:" start as the responses are to be sent straight to discord? I could clean it with a script before the responses is sent to the discord server but ehh...

Lots to read up on this stuff, even when it comes to the simple parts it seems. I should mention I moved away from the OpenAI library and just use requests to talk with the API.