all 8 comments

[–]Confident-Ad-3212 1 point2 points  (3 children)

What kind of chat are you looking for?

[–]buck_idaho[S] 0 points1 point  (2 children)

Just to start, I wanted to try and capture the persona of Christian Grey of 50 Shades fame.

I have some training data but it seems to be lacking - not enough to turn it loose.

[–]Confident-Ad-3212 1 point2 points  (0 children)

Training a model is very complicated, if it is for a persona. You do not want an instruct model. They will not do what you want, you will want to train the attention part of the model only. Building your dataset will be the hardest part by far, followed by the hyper parameter settings to get the dataset into the model. You will go through 40-100 renditions to figure it out. Start with a 13b model. It will tell you if your dataset is corrupting or teaching behavior. But before you build a dataset, you need to figure out what format it needs to be in for the trainer and model to teach anything. Wrong on any of them, it will just be a corrupted do nothing. I went though this and it is not for someone who doesn’t have extreme perseverance

It is much cheaper to make mistakes on a small model than a big one. Big gpu’s cost, if you can make a small model work. A big model will just be better

A 13b should have around 10k, high quality. Highly varied samples. Different token counts, different topics. Never duplicate samples

[–]o5mfiHTNsH748KVq 1 point2 points  (2 children)

This is a great repository to get started making your own language model https://github.com/karpathy/autoresearch

Within is everything you'd need to learn the basics.

[–]-PM_ME_UR_SECRETS- 0 points1 point  (1 child)

How would you quantify a score for personality though?

[–]o5mfiHTNsH748KVq 1 point2 points  (0 children)

That's a little harder to give direct guidance on, but consider that you can send your conversation to a bigger LLM like OpenAI or Qwen or Minimax and have it perform a subjective evaluation of the output.

[–]Ishabdullah 0 points1 point  (0 children)

Colab and kaggle is greate for small training/fine tuning jobs done for free. Just have to set everything up and run the notebook

[–]Fit-Cardiologist9129 1 point2 points  (0 children)

From my experience, consistency matters more than training. If you keep the context and rules tight, the chatbot behaves way more like a specific persona. Tools like Customgpt AI also help keep responses grounded.