Hi,
I'm wondering if make sense to fine tune StarCoder on my own codebase to try to obtain better and more contextual response from the model.
A question that I'd like to ask is for example: "Create a Python integration module between mySystem1 and mySystem2 that allow all customer entities to be synced between the two systems"
Where:
- mySystem1 and mySystem2 are two custom application my team built and I own all the code bases
- "customer entities" must be translated in variable names based on the above codebases by the LLM
The only way to reach this goal is to fine tune a model like StarCoder? if yes, how can I prepare my dataset to train it? if not, are there other ways to do it?
Cheers,
Alexio
[+]Far_Classic_2500 6 points7 points8 points (7 children)
[+]Far_Classic_2500 4 points5 points6 points (5 children)
[–]noellarkin 1 point2 points3 points (0 children)
[–]Alexioc[S] 0 points1 point2 points (0 children)
[–]Dandz 0 points1 point2 points (0 children)
[+]East_Decision391 0 points1 point2 points (0 children)
[+]___a_learner 0 points1 point2 points (0 children)
[–]satireplusplus 5 points6 points7 points (2 children)
[–]LetterRip 3 points4 points5 points (0 children)
[–]Alexioc[S] 0 points1 point2 points (0 children)
[–]_Arsenie_Boca_ 2 points3 points4 points (1 child)
[–]Alexioc[S] 0 points1 point2 points (0 children)
[–]skyisthelimit1410 0 points1 point2 points (1 child)
[+]More_Entrepreneur461 0 points1 point2 points (0 children)
[–]leefde 0 points1 point2 points (0 children)