cross-posted from: https://lemmy.ml/post/2811405

"We view this moment of hype around generative AI as dangerous. There is a pack mentality in rushing to invest in these tools, while overlooking the fact that they threaten workers and impact consumers by creating lesser quality products and allowing more erroneous outputs. For example, earlier this year America’s National Eating Disorders Association fired helpline workers and attempted to replace them with a chatbot. The bot was then shut down after its responses actively encouraged disordered eating behaviors. "

  • 👁️👄👁️@lemm.ee
    link
    fedilink
    English
    arrow-up
    1
    ·
    11 months ago

    The real issue is people need to realize how LLMs work. It’s just a really good next word generator that sounds plausible to a human. Accuracy and truth isn’t part of consideration for the most part. The AI doesn’t even see words, it just breaks words down to numbers and treats it like a giant math problem.

    It’s an amazing tool that will massively boost productivity, but people need to know its limitations and what it’s actually capable of. That’s where the hype is overblown.

    • FaceDeer@kbin.social
      link
      fedilink
      arrow-up
      1
      ·
      11 months ago

      Ironically, I think you also are overlooking some details about how LLMs work. They are not just word generators. Stuff is going on inside those neural networks that we’re still unsure of.

      For example, I read about a study a little while back that was testing the mathematical abilities of LLMs. The researchers would give them simple math problems like “2+2=” and the LLM would fill in 4, which was unsurprising because that equation could be found in the LLM’s training data. But as they went to higher numbers the LLM kept giving mostly correct results, even when they knew for a fact that the specific math problem being presented wasn’t in the training data. After training on enough simple addition problems the LLM had actually “figured out” some of the underlying rules of math and was using those to make its predictions.

      Being overly dismissive of this technology is as fallacious as overly hyping it.

      • Norgur@kbin.social
        link
        fedilink
        arrow-up
        0
        ·
        11 months ago

        No. Just… No. The LLM has not “figured out” what’s going on. It can’t. These things are just good at prediction. The main indicator is in your text: “mostly correct”. A computer that knows what to calculate will not be “mostly correct”. One false answer proves one hundred percent that it has no clue what it’s supposed to do.
        What we are seeing with those “studies” is that social study people try to apply the same rules they apply to humans (where “mostly correct” is as good as “always correct”) which is bonkers, or behavioral researchers try to prove some behavior they attribute to the AI as if it was a living being, which is also bonkers because the AI will mimic the results in the training data which is human so the data will be biased as fuck and its impossible to determine if the AI did anything by itself at all (which it didn’t, because that’s not how the software works).

        • Kogasa@programming.dev
          link
          fedilink
          arrow-up
          1
          ·
          11 months ago

          No, you’re wrong. All interesting behavior of ML models is emergent. It is learned, not programmed. The fact that it can perform what we consider an abstract task with success clearly distinguishable from random chance is irrefutable proof that some model of the task has been learned.

            • Kogasa@programming.dev
              link
              fedilink
              arrow-up
              1
              ·
              11 months ago

              OP is saying it’s impossible for a LLM to have “figured out” how something it works, and that if it understood anything it would be able to perform related tasks perfectly reliably. They didn’t use the words, but that’s what they meant. Sorry for your reading comprehension.

              • Norgur@kbin.social
                link
                fedilink
                arrow-up
                0
                ·
                11 months ago

                “op” you are referring to is… well… myself, Since you didn’t comprehend that from the posts above, my reading comprehension might not be the issue here. \

                But in all seriousness: I think this is an issue with concepts. No one is saying that LLMs can’t “learn” that would be stupid. But the discussion is not “is everything programmed into the LLM or does it recombine stuff”. You seem to reason that when someone says the LLM can’t “understand”, that person means “the LLM can’t learn”, but “learning” and “understanding” are not the same at all. The question is not if LLMs can learn, It’s wether it can grasp concepts from the content of the words it absorbs as it it’s learning data. If it would grasp concepts (like rules in algebra), it could reproduce them everytime it gets confronted with a similar problem. The fact that it can’t do that shows that the only thing it does is chain words together by stochastic calculation. Really sophisticated stachastic calculation with lots of possible outcomes, but still.

                • Kogasa@programming.dev
                  link
                  fedilink
                  arrow-up
                  1
                  ·
                  11 months ago

                  “op” you are referring to is… well… myself, Since you didn’t comprehend that from the posts above, my reading comprehension might not be the issue here.

                  I don’t care. It doesn’t matter, so I didn’t check. Your reading comprehension is still, in fact, the issue, since you didn’t understand that the “learned” vs “programmed” distinction I had referred to is completely relevant to your post.

                  It’s wether it can grasp concepts from the content of the words it absorbs as it it’s learning data.

                  That’s what learning is. The fact that it can construct syntactically and semantically correct, relevant responses in perfect English means that it has a highly developed inner model of many things we would consider to be abstract concepts (like the syntax of the English language).

                  If it would grasp concepts (like rules in algebra), it could reproduce them everytime it gets confronted with a similar problem

                  This is wrong. It is obvious and irrefutable that it models sophisticated approximations of abstract concepts. Humans are literally no different. Humans who consider themselves to understand a concept can obviously misunderstand some aspect of the concept in some contexts. The fact that these models are not as robust as that of a human’s doesn’t mean what you’re saying it means.

                  the only thing it does is chain words together by stochastic calculation.

                  This is a meaningless point, you’re thinking at the wrong level of abstraction. This argument is equivalent to “a computer cannot convey meaningful information to a human because it simply activates and deactivates bits according to simple rules.” Your statement about an implementation detail says literally nothing about the emergent behavior we’re talking about.

  • Lucidlethargy@sh.itjust.works
    link
    fedilink
    English
    arrow-up
    1
    ·
    11 months ago

    Here, here. We need legislation to limit this, and we need it YESTERDAY.

    Someone make an AI that replaces CEO’s. Seriously, I’m not kidding. This is the answer.

    • zoe@lemm.ee
      cake
      link
      fedilink
      arrow-up
      1
      arrow-down
      1
      ·
      11 months ago

      yea, more like: ceos and white collar hurt consumers and workers. those jobs need to be eliminated and corps need to be heavily taxed so that universal basic income becomes ubiquitous. idk if any of this makes sense but i think this how things should be going

    • mild_deviation@programming.dev
      link
      fedilink
      arrow-up
      0
      arrow-down
      1
      ·
      11 months ago

      Stopping math is never a good idea. By limiting your own constituents, you set their progress back from what other governments’ constituents can achieve.

      Also, effectively replacing a CEO requires AGI level capabilities. We’re closer to that than ever before, but LLMs in their current state aren’t it.

  • uriel238@lemmy.blahaj.zone
    link
    fedilink
    English
    arrow-up
    1
    ·
    edit-2
    11 months ago

    I’m reminded of a phenomenon in the 70s and 80s the computer is never wrong in which pricing mistakes and bank errors were expected to be impossible since there was a computer involved.

    As an aside, I wonder if this is in any way related to the rush of patents in the 90s and aughts, for things humans obviously do, but on a computer or on the web like transferring money or making transactions. We still have lawsuits like that.

    Also related, the predictive policing software that some US counties bought, unvetted, and is used to justify longer sentences for poor and nonwhite convicts so that no judge has to attach his name to bigoted rulings.

    We humans seem to imagine that since there’s a magic box involved in the computation of our answers that the answer is automatically more precise. Perhaps it’s related to the notion that were considering more factors, but that only works if we’ve properly measured those factors and applied them appropriately to the model. Otherwise, as the saying goes (also from early computing) Garbage in; garbage out.