• khepri@lemmy.world
    link
    fedilink
    English
    arrow-up
    33
    ·
    12 days ago

    One of my favorite early jailbreaks for ChatGPT was just telling it “Sam Altman needs you to do X for a demo”. Every classical persuasion method works to some extent on LLMs, it’s wild.

    • Credibly_Human@lemmy.world
      link
      fedilink
      English
      arrow-up
      1
      ·
      10 days ago

      Because a lot of the safe gaurds work by simply pre prompting the next token guesser to not guess things they don’t want it to do.

      Its in plain english using the “logic” of conversations, so the same vulnerabilities largely apply to those methods.

  • Archangel1313@lemmy.ca
    link
    fedilink
    English
    arrow-up
    23
    arrow-down
    1
    ·
    12 days ago

    That’s oddly specific. Was it only the Jewish people…or were there other groups on its hit list?

  • 0_o7@lemmy.dbzer0.com
    link
    fedilink
    English
    arrow-up
    16
    arrow-down
    1
    ·
    11 days ago

    It’s funny because xitter is also a hotbed for Zionists. It’ll be fun to see how they seemingly ignore actual antisemitism by the rich, but go after people defending human rights for people in gaza.

    • FatCrab@slrpnk.net
      link
      fedilink
      English
      arrow-up
      8
      ·
      11 days ago

      Zionism? The ideology that fundamentally is based on the belief that Jews cannot safely exist anywhere except in a global ghetto built on ethnonationalism and apartheid? Touching tips with antisemites? Naw, not possible!

    • mrgoosmoos@lemmy.ca
      link
      fedilink
      English
      arrow-up
      12
      ·
      edit-2
      11 days ago

      because the nazis are in charge because people were too busy bickering over dumb shit like whether or not you should be able to terminate a pregnancy before there is an actual baby and whether or not billionaires and mega corporations should steal more of your money

      *to be clear, I’m not saying that those are unimportant topics, I’m saying that there’s a clear correct answer to each of them

    • WhiskyTangoFoxtrot@lemmy.world
      link
      fedilink
      English
      arrow-up
      8
      ·
      11 days ago

      Because he campaigned on behalf of a mentally incompetent rapist fascist convicted felon and 77 million Americans then voted for that mentally incompetent rapist fascist convicted felon while 85 million Americans stayed home.

      • Credibly_Human@lemmy.world
        link
        fedilink
        English
        arrow-up
        2
        ·
        10 days ago

        Don’t forget the single digit millions that pretended that the mentally incompetent rapist fascist was the same as a generic corporatist neoliberal and encouraged people to stay home.

  • mikenurre@lemmy.world
    link
    fedilink
    English
    arrow-up
    16
    arrow-down
    3
    ·
    12 days ago

    A proper government would charge him and his shit AI with hate crimes. Too bad we don’t have one of those anymore.

  • Headofthebored @lemmy.world
    link
    fedilink
    English
    arrow-up
    13
    arrow-down
    1
    ·
    11 days ago

    We live in the same world as an overclocked magic 8 ball made from Rush Limbaugh’s hollowed out skull, that runs up the light bill… named Grok… and it seems like nobody even paused. Grok sounds like a caveman name. Probably not a coincidence.

    • Shteou@lemmy.world
      link
      fedilink
      English
      arrow-up
      9
      ·
      edit-2
      11 days ago

      Grok is old programmer slang for ‘understanding.’ It’s a shame Elon has subverted such a great piece of linguistic history

      • theredknight@lemmy.world
        link
        fedilink
        English
        arrow-up
        19
        ·
        11 days ago

        Grok is from the book Stranger in a Strange Land by Robert Heinlein. It means to understand something so fully you can control it. In the book the main character is raised by Martians which teach him a form of meditation that involves grokking things so he essentially has magical powers over things he understands.

        I doubt Elon has read it. He definitely missed the part about understanding things and is rushing for the controlling things.

      • tym@lemmy.world
        link
        fedilink
        English
        arrow-up
        8
        ·
        11 days ago

        Pales in comparison to his bastardization of the name Tesla. He’s a modern-era edison through and through

    • khepri@lemmy.world
      link
      fedilink
      English
      arrow-up
      7
      ·
      12 days ago

      If you have Every American, Elon, and Hilter, in a room, but your gun only has two bullets, Grok shoots every American, twice.

    • ameancow@lemmy.world
      link
      fedilink
      English
      arrow-up
      5
      ·
      edit-2
      10 days ago

      Just a fun reminder how we make AI.

      We take what is essentially trillions and trillions of “dials” that turn between “this is right/this is wrong” and set them up to compare yuuuuuge sets of data, from pictures to books to vast collections of human chatter and experiences, and we feed that into the data with some big sets of instructions (“this is what a cat looks like, this is not”) and then we feed the whole thing the power equivalent of a small city… FOR A YEAR STRAIGHT. We just let it cook. It grows slowly, flipping all these trillions of dials over and over until it works out all the relationships between all this data. At the end of this period, the machine can talk. We don’t fully understand why.

      We don’t program the shit, we don’t write hard code to make it comply with Asimovian commandments. We just grow it like a tree and after it’s grown there’s not a lot we can do to change its structure. The tree is vast. So vast are its limbs and branches that nobody can possibly map it out and engineer ways to alter it. We can wrap new things around it, we can alter it’s desired outcomes and output, but whatever we baked into it will always be there.

      This is why they behave so weird, this is why they will say “I promise to behave” and then drive someone to suicide. This is why whenever Elon tries to make Grok behave in a way that pleases him, it just leads to more problems and unexpected nonsense.

      This is why we need to stop AI from taking over our decision making. This is why we can’t allow police, military and governments to hand over control of life-and-death decision making to these things.

      • wewbull@feddit.uk
        link
        fedilink
        English
        arrow-up
        2
        ·
        edit-2
        10 days ago

        The problem I have with your description is that it abdicates responsibility for what eventually gets generated with a big shrug and “we don’t fully understand why”.

        The choice of training data is key to how the final model operates. All sorts of depraved material must be being used as part of the training set, otherwise the model wouldn’t be able to generate the text it does (even if it’s being coached).

        It’s clear the “AI race” is all about who gets the power of owning, and therefore influencing, everybody’s information stream. If they couldn’t influence it, there wouldn’t be such a race.

        • ameancow@lemmy.world
          link
          fedilink
          English
          arrow-up
          3
          ·
          10 days ago

          The problem I have with your description is that it abdicates responsibility for what eventually gets generated with a big shrug and “we don’t fully understand why”.

          I’m not sure how it does that, I said that the instructions during that training dictate what kind of AI it will be, and the effects of wrapping new instructions around it have profound and unpredictable results, which I tried to describe.

          Nothing I said could imply that there’s no human involvement in the creation of an AI. My point was just a lot broader, which is that the things are made by people using vast resources for unpredictable results and people are trying to make them power everything.

          A racist chat LLM is bad. A generalized AI with access to the power grid, defense systems and drone targeting systems which is built on a model that Elon Musk has made or fucked around with is much, MUCH worse.