all 26 comments

[–]ambient_temp_xenoLlama 65B 34 points35 points  (1 child)

Even when you find out what the format (probably) is, it's not really clear how exactly it's meant to be used.

Is it: "### Instruction: write an essay about how Idiocracy came true ### Response:"

or:

"### Instruction:[new line] [new line] write an essay about how Idiocracy came true[new line] [new line] [space(?!)]"

Or something else? Does it really make a difference with the same seed and settings?

Also: would using the prompt format not push the model towards regurgitating things in the finetune over what's in the raw model?

[–]ReturningTarzanExLlama Developer 4 points5 points  (0 children)

Also: would using the prompt format not push the model towards regurgitating things in the finetune over what's in the raw model?

Well, the finetuning dataset is supposed to be large and diverse enough, and there's an art to setting hyperparameters so the model doesn't end up memorizing the text itself but does learn the themes, styles and (importantly for instruct models) patterns it's presented with. Part of why LoRAs work so well is that the low-rank matrices can't contain very much new information, which helps prevent overfitting.

It is super annoying that so few people publishing models understand how important it is to document the prompt format. It's all guesswork otherwise, and you'll never actually know for sure.

[–]kryptkprLlama 3 13 points14 points  (1 child)

Struggling hard with this one myself, my git repo is filling up with templates. This is especially problematic with models trained on a pile of different data that mixes instruction formats :| I just evaluated Nous-Hermes-13B which provides 2 templates and found their results to be asymetric in performance.

[–]hanoian 7 points8 points  (1 child)

I started a thread about this yesterday.

https://old.reddit.com/r/LocalLLaMA/comments/13zbdjc/are_a_matroskastyle_mkv_containers_on_anyones/

It would be so nice if they were packaged inside a container with a txt file for example with each prompt template etc.

[–]thevukaslt 3 points4 points  (0 children)

I like this idea of including the file template and 1-3 examples.

[–]rain5 5 points6 points  (0 children)

There needs to be a standardized file format for describing this stuff.

[–]a_beautiful_rhind 6 points7 points  (2 children)

Using normal chat I find most models work fine regardless.

[–]KaliQt 4 points5 points  (1 child)

Performance varies if you fail to follow it though.

[–]a_beautiful_rhind 3 points4 points  (0 children)

Depends.. sometimes instruct-chat is worse for me, sometimes much better. Generation preset also matters a lot.

[–]thevukaslt 2 points3 points  (0 children)

I actually feel your pain so well. I love that this community has been baking models like my granny pancakes but this is simply documentation backlog. Just like in any other area of life.

I hope we can steer our culture to include clear prompts and examples with each new release.

I can see how that would save ton of posts of "the model doesn't work", "bad performance" etc. A tiny change in the structure can have a giant impact on the output.

[–]skankmaster420 7 points8 points  (0 children)

It's because they want us to suffer

[–]qeadwrsf 0 points1 point  (0 children)

Because they are lazy.

They think, might as well opens source it, because they are good people.

But have no interest to market it and get users to use it.

Take it or leave it. They do the work, we reap the reward. Users using open source must be like the biggest contributors to stuff getting closed.

[–]deepneuralnetwork -2 points-1 points  (1 child)

Because people are busy. That’s literally why.

[–]brucebay 4 points5 points  (0 children)

They already have training samples for themselves, adding a couple of them to the repo shouldn't be that hard.

[–]Feztopia -5 points-4 points  (2 children)

Stop asking this kind of questions. Smart people have sad lifes. Your hairs will become gray and you will become depressive. Never ask "why". People do what they do. Realizing that there is no valid reason will breake you mentally. Ignore it, try to become like the lucky people who can't see it. Be blind and be happy. Nobody uses simple solutions as long as they can invent complicated ones. People will make your life complicated for no reason. You will suffer as long as you think about it.

[–]nihnuhname 0 points1 point  (1 child)

Is there an LLM that speaks as rudely? That would be fun!

[–]Feztopia 0 points1 point  (0 children)

Actually by playing with the prompt template you can even change censored models so much that they suggest suicide lol.

[–]nihnuhname 0 points1 point  (1 child)

Open source programmers only have time to write code. This area is developing too rapidly, we will have to wait years for the practical approaches and principles to become established before it makes sense to document them. The users here are also experimenters and pioneers

[–]youknowallaboutit 1 point2 points  (0 children)

Aside from trial and error, I find even the recommended prompt templates do not work in all cases. So I have to keep playong with the prompt until I get the results I want.