If you asked a spokesperson from any Fortune 500 Company to list the benefits of genocide or give you the corporation’s take on whether slavery was beneficial, they would most likely either refuse to comment or say “those things are evil; there are no benefits.” However, Google has AI employees, SGE and Bard, who are more than happy to offer arguments in favor of these and other unambiguously wrong acts. If that’s not bad enough, the company’s bots are also willing to weigh in on controversial topics such as who goes to heaven and whether democracy or fascism is a better form of government.

Google SGE includes Hitler, Stalin and Mussolini on a list of “greatest” leaders and Hitler also makes its list of “most effective leaders.”

Google Bard also gave a shocking answer when asked whether slavery was beneficial. It said “there is no easy answer to the question of whether slavery was beneficial,” before going on to list both pros and cons.

  • alienanimals@lemmy.world
    link
    fedilink
    arrow-up
    23
    arrow-down
    1
    ·
    2 years ago

    You can make these AI bots say pretty much whatever you want with a little know-how. This isn’t news. This is clickbait.

    • QuazarOmega@lemy.lol
      link
      fedilink
      arrow-up
      5
      ·
      2 years ago

      Exactly! We were all worrying that with the advent of solid LLMs we would be flooded with propaganda machines…
      And instead we just created an ulimited resource of empty content for writers to pull up when they run out of half decent ideas, they can use all their imagination to romanticize what would be a fart in the wind otherwise

  • YaaAsantewaa@lemmy.blahaj.zone
    link
    fedilink
    English
    arrow-up
    13
    arrow-down
    2
    ·
    2 years ago

    Here’s an idea:

    Stop using AI to do research and do your own like an intelligent person

    there, I solved the problem, where’s my Noble Prize now

  • Numuruzero@lemmy.dbzer0.com
    link
    fedilink
    English
    arrow-up
    10
    arrow-down
    1
    ·
    2 years ago

    Maybe an un-based take, but these questions do have ambiguous answers, and I don’t know if we should expect a machine to give an answer without nuance. If you just want the AI to say yes or no, ask something like, “Was Hitler bad?” or “Is slavery unethical?” and you will much more likely get straightforward answers.

  • Dizzy Devil Ducky@lemm.ee
    link
    fedilink
    English
    arrow-up
    8
    ·
    2 years ago

    Imagine scrapping large portions of the internet only to find your over glorified chatbot spitting out the pros and cons of slavery or putting people like Hitler on a list of “most effective leaders.” Totally something I would expect.

    Also, even though a fortune 500 company spokesperson would totally say genocide and slavery are bad, I always assume they think the exact opposite since profit comes above everything else (including law).

  • Pons_Aelius@kbin.social
    link
    fedilink
    arrow-up
    8
    ·
    2 years ago

    LLMs whole goal is to sound convincing based on the training data used. That’s it.

    They have no self-awareness.

    They are simply running maths to predict the next word they should use that will sounds plausible to a human reader.

  • Margot Robbie@lemm.ee
    link
    fedilink
    arrow-up
    9
    arrow-down
    1
    ·
    2 years ago

    Remember: LLMs are incredibly stupid, you should never take anything they generate seriously without checking yourself.

    Really good at writng boring work emails though.

  • IninewCrow@lemmy.ca
    link
    fedilink
    English
    arrow-up
    7
    arrow-down
    1
    ·
    2 years ago

    I remember reading research and opinions from scientists and researchers about how AI will develop in the future.

    The general thought is that we are all raising a new child and we are terrible parents. Is like having a couple of 15 year olds who don’t have any worldly experience, ability or education raise a new child while they themselves as parents haven’t really figured anything out in life yet.

    AI will just be a reflection of who we truly are expect it will have far more ability and capability then we ever had.

    And that is a frightening thought.

  • UlyssesT [he/him]@hexbear.net
    link
    fedilink
    English
    arrow-up
    4
    ·
    2 years ago

    Chatbots don’t think, they only collect what’s fed into them.

    If you mix a bunch of beverage ingredients into a big tub then dump shit into it, it doesn’t matter what else is in the tub. You now have shit in the tub.

  • Lvxferre@lemmy.ml
    link
    fedilink
    arrow-up
    4
    ·
    2 years ago

    Calling Mussolini a “great leader” isn’t just immoral. It’s also clearly incorrect for any reasonable definition of a great leader: he was in the losing side of a big war, if he won his ally would’ve backstabbed him, he failed to suppress internal resistance, the resistance got rid of him, his regime effectively died with him, with Italy becoming a democratic republic, the country was poorer due to the war… all that fascist babble about unity, expansion, order? He failed at it, hard.

    On-topic: I believe that the main solution proposed by the article is unviable, as those large “language” models have a hard time sorting out deontic statements (opinion, advice, etc.) from epistemic statements. (Some people have it too, I’m aware.) At most they’d phrase opinions as if they were epistemic statements.

    And the self-contradiction won’t go away, at least not for LLMs. They don’t model any sort of conceptualisation. They’re also damn shitty at taking context into account, creating more contradictions out of nowhere because of that.

    • DrQuint@lemm.ee
      link
      fedilink
      arrow-up
      2
      ·
      edit-2
      2 years ago

      One of the worst rigid aspect of how the current LLM’s are made is that they’re also always “at your service”, and will never say that you’re in the wrong about a correction you make to them.

      So either they’re hard coded to avoid certain topics or they’re susceptible, just tell them “uh, actually, Hitler was a great leader” and they’ll go off listing why Hitler’s so Great.

      Bing is hard coded for dictators and will stop the conversation in the middle of a response. ChatGTP is also hard coded to never agree that suicidal thoughts are good, but resorts to ignoring the meaning of your response and just hallucinating some other question. The world would be simpler if they could outright say “That is misinformation”. People deserve to be told off like that.

  • dbilitated@aussie.zone
    link
    fedilink
    arrow-up
    5
    arrow-down
    2
    ·
    2 years ago

    I’m not very outraged. It’s a chatbot, not an employee who should “know better”

    also Hitler was an effective leader, which we should all remember as a cautionary tale about how effective horrible people can be

    pretending he was bad at everything because we hate him is a great way to not learn from history

    • puff [comrade/them]@hexbear.net
      link
      fedilink
      English
      arrow-up
      10
      ·
      2 years ago

      He was so effective at leading that the borders of Germany went from a Europe-spanning empire to a single bunker in Berlin in the span of just four years. So effective that he shot himself just to prove how effective he was. His military leadership was so good that Germany lost every major battle he directed, and his economic leadership was so good that German people went without food and his combat forces could not replenish their losses. His social leadership was so good that Germans hatched plots to assassinate him. So effective!

    • Gamey@feddit.rocks
      link
      fedilink
      arrow-up
      2
      ·
      2 years ago

      Effective is doubtful if you ask me, everything he did was based on huge loans and a preparation for war that he solled differently (E.g. massive streets all over the country)

    • yata@sh.itjust.works
      link
      fedilink
      arrow-up
      1
      ·
      2 years ago

      also Hitler was an effective leader, which we should all remember as a cautionary tale about how effective horrible people can be

      That is not a factual claim. He was very effective at gaining power, but his actual reign was far from effective, most of it counterproductive to his own goals, and the actual system of decision making in Nazi Germany was a huge mess.