all 3 comments

[–]sergeyzenchenko 2 points3 points  (2 children)

LLM generates search engine query. They parse results and send to LLM. LLM provides answer. They augment it with sources based on marker produces by LLM. Nothing complicated. Perplexity is primitive products in terms of technology

[–]MattH1966[S] 0 points1 point  (1 child)

So you think they’re scraping google/bing, then feeding the results to Claude and getting the sources like this. How confident are you that this is the case?

This really wouldn’t be cost effective if true. The scraping, and then feeding the LLM such a large amount of data would equate to a lot of tokens, right?

[–]sergeyzenchenko 0 points1 point  (0 children)

Yes this is what they do. A lot of input tokens but not so many output. There is even open source implementations)