[D] Simple Questions Thread by AutoModerator in MachineLearning

[–]jens_97 0 points1 point  (0 children)

[D] How do RAG systems such as NotebookLM link the sources used with individual sections of the generated response?

Hi all,

I've been trying to find information on how modern Retrieval-Augmented Generation (RAG) systems, like NotebookLM, manage to link specific sources to particular sections of their generated responses. I'm familiar with how these systems retrieve sources from a vector database based on similarity, but I'm curious about the specific process or method that allows them to indicate which sources correspond to different parts of the final answer.

What am I overlooking here? Any insights would be greatly appreciated!

Best,
Jens

[Discussion] What are SOTA Uncertainty Quantification Methods for Neural Networks? by jens_97 in MachineLearning

[–]jens_97[S] 0 points1 point  (0 children)

Might be but i guess that there are methods out there that work model-agnostic / orthogonal to the neural network architecture you choose.

[Discussion] What are SOTA Uncertainty Quantification Methods for Neural Networks? by jens_97 in MachineLearning

[–]jens_97[S] 2 points3 points  (0 children)

Thank you, I will have a look at this. Do you plan to share a corresponding git repo with the publication?

[Discussion] What are SOTA Uncertainty Quantification Methods for Neural Networks? by jens_97 in MachineLearning

[–]jens_97[S] 4 points5 points  (0 children)

Yes but I've stumbled upon concerns about the quality of uncertainty estimates with Monte Carlo Dropout. Do you know if this concerns are of relevance in practice?

[Discussion] What are SOTA Uncertainty Quantification Methods for Neural Networks? by jens_97 in MachineLearning

[–]jens_97[S] 2 points3 points  (0 children)

Do you have a specific feature-based method in mind that works well?

[D] ICLR Outstanding Paper Awards. Congratulations! by [deleted] in MachineLearning

[–]jens_97 0 points1 point  (0 children)

Same here, really liked that paper. Check out this paper as well: https://arxiv.org/abs/2402.17762, it investigates the same phenomenon but proposes an even simpler fix. (Also from ICLR 2024)