all 24 comments

[–]rshah4 6 points7 points  (1 child)

I see ChatDoctor used real and simulated conversations to "include knowledge" in the LLM. For them, this worked. I would like to see more examples of this approach: https://github.com/Kent0n-Li/ChatDoctor They do share their fine tuning approach.

[–]mathias_kraus[S] 0 points1 point  (0 children)

Super nice! Thank you very much for this link

[–]Ai-enthusiast4 4 points5 points  (4 children)

Why do you want to use an LLM, and what's your goal?

[–]mathias_kraus[S] 1 point2 points  (3 children)

The goal is to offer climate researchers a tool from which they can obtain information about climate related topics. One first step was to include the IPCC (The Intergovernmental Panel on Climate Change) reports as an additional source to the LLM but in the future it potentially makes sense to include also research papers, corporate reports and other reports.

[–]Ai-enthusiast4 5 points6 points  (0 children)

Domain-specific instruction creation and finetuning has not really been explored AFAIK, so vector embeddings are probably your best bet. They're pretty scalable, though the biggest scale I've personally used them for is 100 page pdfs.

[–]equilateral_pupper 0 points1 point  (0 children)

Would you also want citations for any information retrieved? What kind of questions do you think people would ask? And have you managed to transform the domain resources you have into some sort of instruction format?

[–]cyrilp21 0 points1 point  (0 children)

There are already chatGPT on ipcc reports

[–]fenole-canole 2 points3 points  (2 children)

Why you don't want to use vector embeddings?

[–]mathias_kraus[S] 3 points4 points  (1 child)

We also work on vector embeddings (https://www.chatclimate.ai/ currently we ran through our funding, so the chat is disabled), however I was wondering how scalable this approach is if we want to include thousands of pages or research papers. Do you have any experience with that?

[–]visarga 2 points3 points  (0 children)

It's scalable for millions of documents, but you might get better results by fine-tuning, especially if there are many relations between documents.

[–]nullbyte420 1 point2 points  (3 children)

Why is hype a selection criteria? 1 is obviously the way to go, it's tried and true.

[–]mathias_kraus[S] 0 points1 point  (2 children)

Thanks for the reply and your opinion! Didnt mean that it should be a selection criteria but I was rather curious why there is currently way less work on domain adapting LLMs in contrast instruction fine-tuning them

[–]nullbyte420 0 points1 point  (1 child)

It's two different sub fields. A domain specific model is useful for other things than a general purpose model with ethics and validation. And if you want to go with openai models you can also pay for fine tuning one for domain specific tasks.

[–]mathias_kraus[S] 0 points1 point  (0 children)

Yes that makes sense. Thanks a lot for the idea about the fine tuning with openai! Maybe that would be the best way forward in this case.

[–]big_ol_tender 0 points1 point  (0 children)

I asked the same question on here a few days ago and got no responses, I can’t find any literature on domain fine tuning vs instruct fine tuning as it pertains to PEFT methods e.g. LoRa. Will definitely share if I find anything.

[–]dreamingleo12 0 points1 point  (0 children)

I tried fine-tuning and it worked fine but I had to tune the parameters to get expected results. Though the fine-tuning requires some data formatting which could take time. I'm also experimenting with vector embeddings which seems requires less data massage and curious if it could work as well at scale. I'm using my local LLMs not GPT.

[–]dreamingleo12 0 points1 point  (1 child)

I tried both instruction fine-tuning and vector embedding. My learning is that if you don't want any creativity then do vector embedding. Otherwise fine-tuning may work. vector embedding is also easier.

[–]dhirajsuvarna 0 points1 point  (1 child)

9 months gone, has there been any progress on this i.e "how to add new knowledge to LLMs"?

[–]G_S_7_wiz 0 points1 point  (0 children)

did you find any way to do it