• @[email protected]
    link
    fedilink
    22 days ago

    Yeah I’ve noticed it’s way more sycophantic than it used to be, but it’s also easier to get it to say things it’s not supposed to by not going at it directly. So like I started by asking about a legitimate religious topic and then acted like it was inflaming existing delusions of grandeur. If you go to ChatGPT and say “I am God” it will say “no you aren’t” but if you do what I did and start with something seemingly innocuous it won’t fight as hard. Fundamentally this is because it doesn’t have any thoughts, beliefs, or feelings that it can stand behind, it’s just a text machine. But that’s not how it’s marketed or how people interact with it

    • @[email protected]
      link
      fedilink
      12 days ago

      it’s a matter of time before some kids poison themselves by trying to make drugs using recipes they got by “jailbreaking” some LLM.