• self@awful.systems
    link
    fedilink
    English
    arrow-up
    0
    ·
    2 months ago

    are there mechanisms known to researchers that Microsoft’s not using that can prevent this type of failure case in an LLM without resorting to whack-a-mole with a regex?

    • linearchaos@lemmy.world
      link
      fedilink
      English
      arrow-up
      0
      ·
      2 months ago

      Yeah there’s already a lot of this in play.

      You run the same query multiple times through multiple models and do a web search looking for conflicting data.

      I’ve had copilot answer a query, then erase the output and tell me it couldn’t answer it after about 5 seconds.

      I’ve also seen responses contradict themselves later paragraphs saying there are other points of view.

      It would be a simple matter to have it summarize the output it’s about to give you and dump the output of it paints the subject in a negative light.

      • froztbyte@awful.systems
        link
        fedilink
        English
        arrow-up
        1
        ·
        2 months ago

        It would be a simple matter to have it summarize the output it’s about to give you and dump the output of it paints the subject in a negative light.

        lol. like that’s a fix

        (Hindenburg, hitler, great depression, ronald reagan, stalin, modi, putin, decades of north korea life, …)