• catloaf@lemm.ee
    link
    fedilink
    English
    arrow-up
    110
    arrow-down
    11
    ·
    2 months ago

    To lie requires intent to deceive. LLMs do not have intents, they are statistical language algorithms.

    • CosmoNova@lemmy.world
      link
      fedilink
      English
      arrow-up
      21
      arrow-down
      2
      ·
      2 months ago

      It’s interesting they call it a lie when it can’t even think but when any person is caught lying media will talk about “untruths” or “inconsistencies”.

    • moakley@lemmy.world
      link
      fedilink
      English
      arrow-up
      7
      ·
      2 months ago

      I’m not convinced some people aren’t just statistical language algorithms. And I don’t just mean online; I mean that seems to be how some people’s brains work.

      • Chozo@fedia.io
        link
        fedilink
        arrow-up
        33
        arrow-down
        4
        ·
        2 months ago

        Read about how LLMs actually work before you read articles written by people who don’t understand LLMs. The author of this piece is suggesting arguments that imply that LLMs have cognition. “Lying” requires intent, and LLMs have no intention, they only have instructions. The author would have you believe that these LLMs are faulty or unreliable, when in actuality they’re working exactly as they’ve been designed to.

        • thedruid@lemmy.world
          link
          fedilink
          English
          arrow-up
          1
          arrow-down
          5
          ·
          2 months ago

          So working as designed means presenting false info?

          Look , no one is ascribing intelligence or intent to the machine. The issue is the machines aren’t very good and are being marketed as awesome. They aren’t

          • Chozo@fedia.io
            link
            fedilink
            arrow-up
            7
            arrow-down
            1
            ·
            2 months ago

            So working as designed means presenting false info?

            Yes. It was told to conduct a task. It did so. What part of that seems unintentional to you?

            • thedruid@lemmy.world
              link
              fedilink
              English
              arrow-up
              1
              arrow-down
              5
              ·
              edit-2
              2 months ago

              That’s not completing a task. That’s faking a result for appearance.

              Is that what you’re advocating for ?

              If I ask an llm to tell me the difference between aeolian mode and Dorian mode in the field of music , and it gives me the wrong info, then no it’s not working as intended

              See I chose that example because I know the answer. The llm didn’t. But it gave me an answer. An incorrect one

              I want you to understand this. You’re fighting the wrong battle. The llms do make mistakes. Frequently. So frequently that any human who made the same amount of mistakes wouldn’t keep their job.

              But the investment, the belief in a.i is so engrained for some of us who so want a bright and technically advanced future, that you are now making excuses for it. I get it. I’m not insulting you. We are humans. We do that. There are subjects I am sure you could point at where I do this as well

              But a.i.? No. It’s just wrong so often. It’s not it’s fault. Who knew that when we tried to jump ahead in the tech timeline, that we should have actually invented guardrail tech first?

              Instead we let the cart go before the horses, AGAIN, because we are dumb creatures , and now people are trying to force things that don’t work correctly to somehow be shown to be correct.

              I know. A mouthful. But honestly. A.i. is poorly designed, poorly executed, and poorly used.

              It is hastening the end of man. Because those who have been singing it’s praises are too invested to admit it.

              It simply ain’t ready.

              Edit: changed “would” to “wouldn’t”

              • Chozo@fedia.io
                link
                fedilink
                arrow-up
                7
                ·
                2 months ago

                That’s not completing a task.

                That’s faking a result for appearance.

                That was the task.

                • thedruid@lemmy.world
                  link
                  fedilink
                  English
                  arrow-up
                  1
                  arrow-down
                  4
                  ·
                  2 months ago

                  No, the task was To tell me the difference in the two modes.

                  It provided incorrect information and passed it off as accurate. It didn’t complete the task

                  You know that though. You’re just too invested to admit it. So I will withdraw. Enjoy your day.

      • catloaf@lemm.ee
        link
        fedilink
        English
        arrow-up
        27
        arrow-down
        2
        ·
        2 months ago

        I’ve read the article. If there is any dishonesty, it is on the part of the model creator or LLM operator.

      • gravitas_deficiency@sh.itjust.works
        link
        fedilink
        English
        arrow-up
        17
        arrow-down
        5
        ·
        2 months ago

        You need to understand that lemmy has a lot of users that actually understand neural networks and the nuanced mechanics of machine learning FAR better than the average layperson.

        • venusaur@lemmy.world
          link
          fedilink
          English
          arrow-up
          2
          arrow-down
          2
          ·
          2 months ago

          And A LOT of people who don’t and blindly hate AI because of posts like this.

        • thedruid@lemmy.world
          link
          fedilink
          English
          arrow-up
          1
          arrow-down
          6
          ·
          2 months ago

          That’s a huge, arrogant and quite insulting statement. Your making assumptions based on stereotypes

            • thedruid@lemmy.world
              link
              fedilink
              English
              arrow-up
              1
              arrow-down
              1
              ·
              2 months ago

              No. You’re mad at someone who isn’t buying that a. I. 's are anything but a cool parlor trick that isn’t ready for prime time

              Because that’s all I’m saying. The are wrong more often than right. They do not complete tasks given to them and they really are garbage

              Now this is all regarding the publicly available a. Is. What ever new secret voodoo one. Think has or military has, I can’t speak to.

              • gravitas_deficiency@sh.itjust.works
                link
                fedilink
                English
                arrow-up
                2
                ·
                2 months ago

                Uh, just to be clear, I think “AI” and LLMs/codegen/imagegen/vidgen in particular are absolute cancer, and are often snake oil bullshit, as well as being meaningfully societally harmful in a lot of ways.