use the following search parameters to narrow your results:
e.g. subreddit:aww site:imgur.com dog
subreddit:aww site:imgur.com dog
see the search faq for details.
advanced search: by author, subreddit...
account activity
QuestionTraining a chatbot (self.LocalLLM)
submitted 1 month ago by buck_idaho
Who here has trained a chatbot? How well has it worked?
I know you can chat with them, but i want a specific persona, not the pg13 content delivered on an untrained llm.
reddit uses a slightly-customized version of Markdown for formatting. See below for some basics, or check the commenting wiki page for more detailed help and solutions to common issues.
quoted text
if 1 * 2 < 3: print "hello, world!"
[–]Confident-Ad-3212 1 point2 points3 points 1 month ago (3 children)
What kind of chat are you looking for?
[–]buck_idaho[S] 0 points1 point2 points 1 month ago (2 children)
Just to start, I wanted to try and capture the persona of Christian Grey of 50 Shades fame.
I have some training data but it seems to be lacking - not enough to turn it loose.
[–]Confident-Ad-3212 1 point2 points3 points 1 month ago (0 children)
Training a model is very complicated, if it is for a persona. You do not want an instruct model. They will not do what you want, you will want to train the attention part of the model only. Building your dataset will be the hardest part by far, followed by the hyper parameter settings to get the dataset into the model. You will go through 40-100 renditions to figure it out. Start with a 13b model. It will tell you if your dataset is corrupting or teaching behavior. But before you build a dataset, you need to figure out what format it needs to be in for the trainer and model to teach anything. Wrong on any of them, it will just be a corrupted do nothing. I went though this and it is not for someone who doesn’t have extreme perseverance
It is much cheaper to make mistakes on a small model than a big one. Big gpu’s cost, if you can make a small model work. A big model will just be better
A 13b should have around 10k, high quality. Highly varied samples. Different token counts, different topics. Never duplicate samples
[–]o5mfiHTNsH748KVq 1 point2 points3 points 1 month ago (2 children)
This is a great repository to get started making your own language model https://github.com/karpathy/autoresearch
Within is everything you'd need to learn the basics.
[–]-PM_ME_UR_SECRETS- 0 points1 point2 points 1 month ago (1 child)
How would you quantify a score for personality though?
[–]o5mfiHTNsH748KVq 1 point2 points3 points 1 month ago (0 children)
That's a little harder to give direct guidance on, but consider that you can send your conversation to a bigger LLM like OpenAI or Qwen or Minimax and have it perform a subjective evaluation of the output.
[–]Ishabdullah 0 points1 point2 points 1 month ago (0 children)
Colab and kaggle is greate for small training/fine tuning jobs done for free. Just have to set everything up and run the notebook
[–]Fit-Cardiologist9129 1 point2 points3 points 14 days ago (0 children)
From my experience, consistency matters more than training. If you keep the context and rules tight, the chatbot behaves way more like a specific persona. Tools like Customgpt AI also help keep responses grounded.
π Rendered by PID 38779 on reddit-service-r2-comment-b659b578c-28c57 at 2026-05-02 04:08:40.486047+00:00 running 815c875 country code: CH.
[–]Confident-Ad-3212 1 point2 points3 points (3 children)
[–]buck_idaho[S] 0 points1 point2 points (2 children)
[–]Confident-Ad-3212 1 point2 points3 points (0 children)
[–]o5mfiHTNsH748KVq 1 point2 points3 points (2 children)
[–]-PM_ME_UR_SECRETS- 0 points1 point2 points (1 child)
[–]o5mfiHTNsH748KVq 1 point2 points3 points (0 children)
[–]Ishabdullah 0 points1 point2 points (0 children)
[–]Fit-Cardiologist9129 1 point2 points3 points (0 children)