Martin Bernklau is a German journalist who reported for decades on criminal trials. He looked himself up on Bing, which suggests you use its Copilot AI. Copilot then listed a string of crimes Bernk…
are there mechanisms known to researchers that Microsoft’s not using that can prevent this type of failure case in an LLM without resorting to whack-a-mole with a regex?
are there mechanisms known to researchers that Microsoft’s not using that can prevent this type of failure case in an LLM without resorting to whack-a-mole with a regex?
Yeah there’s already a lot of this in play.
You run the same query multiple times through multiple models and do a web search looking for conflicting data.
I’ve had copilot answer a query, then erase the output and tell me it couldn’t answer it after about 5 seconds.
I’ve also seen responses contradict themselves later paragraphs saying there are other points of view.
It would be a simple matter to have it summarize the output it’s about to give you and dump the output of it paints the subject in a negative light.
lol. like that’s a fix
(Hindenburg, hitler, great depression, ronald reagan, stalin, modi, putin, decades of north korea life, …)