• @ALostInquirer@lemm.ee
    link
    fedilink
    English
    171 year ago

    Why do tech journalists keep using the businesses’ language about AI, such as “hallucination”, instead of glitching/bugging/breaking?

      • @Blackmist@feddit.uk
        link
        fedilink
        English
        141 year ago

        Honestly, it’s the most human you’ll ever see it act.

        It’s got upper management written all over it.

      • @ALostInquirer@lemm.ee
        link
        fedilink
        English
        -1
        edit-2
        1 year ago

        (AI confidently BSing)

        Isn’t it more accurate to say it’s outputting incorrect information from a poorly processed prompt/query?

        • @vithigar@lemmy.ca
          link
          fedilink
          English
          261 year ago

          No, because it’s not poorly processing anything. It’s not even really a bug. It’s doing exactly what it’s supposed to do, spit out words in the “shape” of an appropriate response to whatever was just said

          • @ALostInquirer@lemm.ee
            link
            fedilink
            English
            3
            edit-2
            1 year ago

            When I wrote “processing”, I meant it in the sense of getting to that “shape” of an appropriate response you describe. If I’d meant this in a conscious sense I would have written, “poorly understood prompt/query”, for what it’s worth, but I see where you were coming from.

    • @Danksy@lemmy.world
      link
      fedilink
      English
      301 year ago

      It’s not a bug, it’s a natural consequence of the methodology. A language model won’t always be correct when it doesn’t know what it is saying.

      • @ALostInquirer@lemm.ee
        link
        fedilink
        English
        81 year ago

        Yeah, on further thought and as I mention in other replies, my thoughts on this are shifting toward the real bug of this being how it’s marketed in many cases (as a digital assistant/research aid) and in turn used, or attempted to be used (as it’s marketed).

        • @Danksy@lemmy.world
          link
          fedilink
          English
          81 year ago

          I agree, it’s a massive issue. It’s a very complex topic that most people have no way of understanding. It is superb at generating text, and that makes it look smarter than it actually is, which is really dangerous. I think the creators of these models have a responsibility to communicate what these models can and can’t do, but unfortunately that is not profitable.

      • @ALostInquirer@lemm.ee
        link
        fedilink
        English
        41 year ago

        It’s not a bad article, honestly, I’m just tired of journalists and academics echoing the language of businesses and their marketing. “Hallucinations” aren’t accurate for this form of AI. These are sophisticated generative text tools, and in my opinion lack any qualities that justify all this fluff terminology personifying them.

        Also frankly, I think students have one of the better applications for large-language model AIs than many adults, even those trying to deploy them. Students are using them to do their homework, to generate their papers, exactly one of the basic points of them. Too many adults are acting like these tools should be used in their present form as research aids, but the entire generative basis of them undermines their reliability for this. It’s trying to use the wrong tool for the job.

        You don’t want any of the generative capacities of a large-language model AI for research help, you’d instead want whatever text-processing it may be able to do to assemble and provide accurate output.