ChatGPT use declines as users complain about ‘dumber’ answers, and the reason might be AI’s biggest threat for the future::AI for the smart guy?

  • TimewornTraveler@lemm.ee
    link
    fedilink
    English
    arrow-up
    3
    ·
    edit-2
    1 year ago

    You raise some great insights. As this tech becomes available to humanity, we cannot rely on the bias of one company to keep us safe. That doesn’t mean “ethics in AI” is a mistake, though. (But that is an attention-grabbing phrase!). I believe you neglect what ethics fundamentally is: the way humans navigate one another. It’s how we think and breathe. Ethics are core to our very existence, and not something that you can just pretend doesn’t exist. Even saying nothing is a kind of response.

    What all this means is that if we are designing technology that can teach anyone how to kill in ways they wouldn’t otherwise have been able to, we have to address the realities of that conversation. That’s a conversation that cannot be had just internally in one company, and I think we see eye to eye on that. But saying nothing?

    • 👁️👄👁️@lemm.ee
      link
      fedilink
      English
      arrow-up
      2
      ·
      edit-2
      1 year ago

      Maybe ethics is a bit more complicated for this discussion, but it makes me think how do uncensored LLMs still have ethics, yet remain uncensored? Maybe there’s a fine line somewhere. I can agree that it should be steered till more positive things, like saying murder and suicide is bad. The description of that model I linked says it’s still influenced by ethics, but has the guardrails turned off, and maybe that would be a better idea then what I initially said.

      Should custom models be allowed to be run or modified? Should these things be open source? I don’t know the answer to all these questions, but I’ll always advocate for foss and custom models, as I fundamentally see it as a tool that should be allowed to be owned. Which that is at odds with restrictive ethics rhetorics I hear.

      But your second point that it shouldn’t be taught to kill. I think that argument could be used to ban violent video games. You won’t do very good in Overwatch or Valorant if you don’t know how to kill after all. To learn how to hide a dead body, how much more detailed can you get then just turning on the TV and watching Criminal Minds? Our entertainment has zero issue teaching how to kill, encouraging violence (gotta rank up somehow), or hide dead body. Is an AI describing what this media already shows in text form so much worse?

      Side note: that hyperlink I added links to the 33b uncensored WizardLM model which is pretty fun to play around with if you haven’t already tried. Also GPT4All is a cool way to run various local models offline on your computer.

      • TimewornTraveler@lemm.ee
        link
        fedilink
        English
        arrow-up
        3
        ·
        edit-2
        1 year ago

        But your second point that it shouldn’t be taught to kill.

        Whoa hold up. that’s not what I said at all! I said if it is going to exist, what do we do about it?

        My point is that this ethical conversation is already happening, we cannot change that. The issue is that OpenAI dominates the conversation. The solution cannot be “pretend there’s nothing to talk about”.