We have all seen AI-based searches available on the web like Copilot, Perplexity, DuckAssist etc, which scour the web for information, present them in a summarized form, and also cite sources in support of the summary.
But how do they know which sources are legitimate and which are simple BS ? Do they exercise judgement while crawling, or do they have some kind of filter list around the “trustworthyness” of various web sources ?
So scrapping “popular websites” plus “someone said this is a good source for topic X” plus wikipedia? And summarizing over them all? That sounds like a very bad idea, because it’s very fragile to poisoning?
Ya I can see AI resulting in many deaths if people start trusting it for things like “is this mushroom edible”?
Isn’t that how all ranking works everywhere? How else can it rank sources?
My point is “summarizing over all of those” and “poisoning”.
Source of category 1 says cheese is made from XYZ and yellow
Source from category 2 confirms 1 in different words and adds that it has holes
Source from category 3 confirms 2 and adds that its also blue, not only yellow
Source 4 talks about blue cheese only
Poisoning would mean that in the summary cheese is yellow with blue holes.