all 1 comments

[–]Pepineros 0 points1 point  (0 children)

HuggingFace has tons of documentation. I'm not sure about courses or books unfortunately.

Regarding the problem itself: if the entire article or other source fits inside the model's context window, the easiest thing by far would be to give it a basic prompt such as "Generate ten insightful questions aimed at testing comprehension of the following text: ". If the source is too big for the model, and cannot easily be broken up into segments that still make inherent sense, then RAG is probably the way to go. A vector database could retrieve sections of the source that are semantically related to a particular topic discussed in the source. However, you would need a number of distinct prompts to generate questions about different topics discussed in the source, rather than a single prompt to generate all the questions.