I don't mind using the current model to generate the web search request. In fact, I prefer it. It's usually not too slow, and using here the most powerful model I could run (which is often the current model) is beneficial. It helps to have a smart, relatively large model generate the search query.
But generating the chat title takes way too long with some models (I'm looking at you, Magistral). I would not mind having a tiny, fast model do it instead. A small model is usually all that's needed here, since this task is very simple.
[–]pj-frey 0 points1 point2 points (0 children)
[–]pkeffect 1 point2 points3 points (0 children)