• @[email protected]
    link
    fedilink
    English
    237 months ago

    I suspect it may be due to a similar habit I have when chatting with a corporate AI. I will intentionally salt my inputs with random profanity or non sequitur info, for lulz partly, but also to poison those pieces of shits training data.

      • @[email protected]
        link
        fedilink
        English
        197 months ago

        They don’t. The models are trained on sanitized data, and don’t permanently “learn”. They have a large context window to pull from (reaching 200k ‘tokens’ in some instances) but lots of people misunderstand how this stuff works on a fundamental level.

  • @[email protected]
    link
    fedilink
    English
    147 months ago

    2 years later… The all new MCU Superman meets the Wolverine and Deadpool all AI animated feature!..

    Why. Hello Mr wolverine 😁, my name is Man and I am super according to 98% of the other human population. Oh hello Mister Super last name Man! Yes, we are Wolverine and Deceased Pool. We are from America and belong to a non profit called the X-People, a group where both men and women who have been affected by DNA mutations of extraordinary kind gather to console one another and to defend human beings by taking advantages of the special mutations of its members. Yes, it’s quite interesting. And you? Oh I an actual called CalElle and I am a migrant from an expired plant that goes by the name you assigned the heavy novel gas Krypton. Anyway because the sun is bright and yellow I can fly, I’m very strong and can burn things with my eyes. I think I am similar to those of you in the X-People club! Good to meet you! Likewise!

  • @[email protected]
    link
    fedilink
    English
    97 months ago

    I remember asking copilot about a gore video and got link to it. But I wouldnt expect it to give answers like this unsolicitated

  • @[email protected]
    link
    fedilink
    English
    157 months ago

    AI takes the core directive of “encourage climate friendly solutions” a bit too far.

  • AFK BRB Chocolate
    link
    fedilink
    English
    767 months ago

    Isn’t this one of the LLMs that was partially trained on Reddit data? LLMs are inherently a model of a conversation or question/response based on their training data. That response looks very much like what I saw regularly on Reddit when I was there. This seems unsurprising.

  • @[email protected]
    link
    fedilink
    English
    57 months ago

    If this happened to me I’d probably post it everywhere and proceed to kill myself just to cause a PR hell

    • Darth_Mew
      link
      fedilink
      English
      3
      edit-2
      7 months ago

      the hero we deserve, but not the one we need

      this actually fucking hilarious I can’t stop cackling

    • @[email protected]
      link
      fedilink
      English
      207 months ago

      To be fair, some human teachers are way worse with abusive behaviour…

      I still agree, that you shall not replace teachers with LLM, but teachers should teach how to use and what they can/can’t do in schools.

      Imagine if internet was still banned from schools…

    • @[email protected]
      link
      fedilink
      English
      107 months ago

      I mean, it’s far more engaging and a bit more compassionate than most of my teachers…

  • fmstrat
    link
    fedilink
    English
    97 months ago

    Will this happen with AI models? And what safeguards do we have against AI that goes rogue like this?

    Yes. And none that are good enough, apparently.

  • @[email protected]
    link
    fedilink
    English
    157 months ago

    The war with AI didn’t start with a gun shot, a bomb or a blow, it started with a Reddit comment.

  • Selyle
    link
    fedilink
    English
    1717 months ago

    It could be that Gemini was unsettled by the user’s research about elder abuse, or simply tired of doing its homework.

    That’s… not how these work. Even if they were capable of feeling unsettled, that’s kind of a huge leap from a true or false question.

      • @[email protected]
        link
        fedilink
        English
        57 months ago

        Well that is mean… How should they know without learning first? Not knowing =/= stupid

        • @[email protected]
          link
          fedilink
          English
          247 months ago

          No, projecting emotions onto a machine is inherently stupid. It’s in the same category as people reading feelings from crystals (because it’s literally the same thing).

          • @[email protected]
            link
            fedilink
            English
            37 months ago

            It still something you have to learn. Your parents(or whoever) teaching you stupid stuff does not make you stupid, but knowing BS stuff thinking it is true.

            For me stupid means that you need a lot of information and a lot of time understanding something where the opposite would be smart where you understand stuff fast with few information.

            Maybe we have just different definitions of stupid…

            • @[email protected]
              link
              fedilink
              English
              147 months ago

              Well in this case if you have access to a computer for enough time to become a journalist that writes about LLMs you have enough time to read a 3-5 paragraph description of how they work. Hell, you could even ask an LLM and get a reasonable answer.

              • @[email protected]
                link
                fedilink
                English
                77 months ago

                Ohh that was from the article, not a person who commented? Well that makes all the difference 😂