• milicent_bystandr@lemm.ee
        link
        fedilink
        English
        arrow-up
        0
        ·
        3 months ago

        Because too frequently it gives plausible-sounding but completely unfounded statements.

        Also it can go more darkly wrong, and all the extra checks and safeguards don’t always protect it.

          • milicent_bystandr@lemm.ee
            link
            fedilink
            English
            arrow-up
            1
            ·
            24 days ago

            Because a human can understand the situation, and the person they’re talking to, and reply with wisdom, rather than just parroting what seems like what they heard before.

            • anus@lemmy.world
              link
              fedilink
              English
              arrow-up
              1
              ·
              9 days ago

              Are you saying that humans don’t parrot what seems like what they heard before?

              • milicent_bystandr@lemm.ee
                link
                fedilink
                English
                arrow-up
                1
                ·
                9 days ago

                Oh we absolutely do. And we tell lies, and we misunderstand, and miscommunicate.

                But not all the time, and not everyone. So if your friend if they’d like dinner, you expect the answer to be true to what they want, not just whatever sounds good to the general population. If you read a scientific journal, you expect the scientists to represent the facts and even the meaning of their research, not parrot some ideas from a half-forgotten textbook. And if you see a professional counsellor, you expect them to have a good understanding of human nature, and to genuinely empathise with your situation, and have good ways to help you out.

                And of course all three of those examples fail sometimes, which is why as part of life we learn who we can trust and to what extent.

                • anus@lemmy.world
                  link
                  fedilink
                  English
                  arrow-up
                  1
                  ·
                  7 days ago

                  I would argue that all of the cases you presented fail at a comparable rate compared to foundational LLMs