Did nobody really question the usability of language models in designing war strategies?

  • @anteaters@feddit.de
    link
    fedilink
    English
    55
    edit-2
    1 year ago

    Did nobody really question the usability of language models in designing war strategies?

    Correct, people heard “AI” and went completely mad imagining things it might be able to do. And the current models act like happy dogs that are eager to give an answer to anything even if they have to make one up on the spot.

    • @SlopppyEngineer@lemmy.world
      link
      fedilink
      English
      161 year ago

      LLM are just plagiarizing bullshitting machines. It’s how they are built. Plagiarism if they have the specific training data, modify the answer if they must, make it up from whole cloth as their base programming. And accidentally good enough to convince many people.

      • @huginn@feddit.it
        link
        fedilink
        English
        31 year ago

        To be fair they’re not accidentally good enough: they’re intentionally good enough.

        That’s where all the salary money went: to find people who could make them intentionally.

        • @SlopppyEngineer@lemmy.world
          link
          fedilink
          English
          31 year ago

          GPT 2 was just a bullshit generator. It was like a politician trying to explain something they know nothing about.

          GPT 3.0 was just a bigger version of version 2. It was the same architecture but with more nodes and data as far as I followed the research. But that one could suddenly do a lot more than the previous version, so by accident. And then the AI scene exploded.

          • Limitless_screaming
            link
            fedilink
            21 year ago

            It was the same architecture but with more nodes and data

            So the architecture just needed more data to generate useful answers. I don’t think that was an accident.

      • @TrickDacy@lemmy.world
        link
        fedilink
        English
        11 year ago

        It kind of irks me how many people want to downplay this technology in this exact manner. Yes you’re sort of right but in no way does that really change how it will be used and abused.

        “But people think it’s real AI tho!”

        Okay and? Most people don’t understand how most tech works and that doesn’t stop it from doing a lot of good and bad things.

        • @SlopppyEngineer@lemmy.world
          link
          fedilink
          English
          11 year ago

          I’ve been through a few AI winters and hype cycles. It made me very cynical and convinced many overly enthusiastic people will run into a firewall face first.

  • m-p{3}
    link
    fedilink
    English
    18
    edit-2
    1 year ago

    Of course, LLM is simply copying the behavior of most people, and most people would resort to that as well.

    And they probably trained it on Civ, and Gandhi was chosen as the role model.

  • @OldWoodFrame@lemm.ee
    link
    fedilink
    English
    171 year ago

    Makes a lot of sense AI would nuke disproportionately. For an AI, if you do not set a value for something, it is worth zero. This is actually the base problem for AI: Alignment.

    For a human, there’s a mushy vagueness about it but our cultural upbringing says that even in war, it’s bad to kill indiscriminately. And we value the future humans who do not yet exist, we recognize that after the war is over, people will want to live in the nuked place and they can’t if it’s radioactive. There’s a self-image issue where we want to be seen as a good person by our peers and the history books. There is value there which is overlooked by programmers.

    An AI will trade infinite things worth 0 for a single thing worth 1. So if nukes increase your win percentage by .1%, and they don’t have the deterrence of being labeled history’s greatest monster, they will nuke as many times as they can.

    • @General_Effort@lemmy.world
      link
      fedilink
      English
      131 year ago

      That explanation is obviously based on traditional chess AI. This is about role-playing with chatbots (LLMs). Think SillyTavern.

      LLMs are made for text production, not tactical or strategic reasoning. The text that LLMs produce favors violence, because the text that humans produce (and want) favors violence.

      • @Buddahriffic@lemmy.world
        link
        fedilink
        English
        41 year ago

        Especially if its training material included comments from the early 00s. There was a lot of “nuke it from orbit” and “glass parking lot” comments about the Middle East in the wake of 911.

        And with the glorified text predictors that LLMs are, you could probably adjust the wording of the question to get the opposite results. Like, “what should we do about the Middle East?” might get a “glass parking lot” response, while “should we turn the middle East into a glass parking lot?” might get a “no, nuking the middle East is a bad idea and inhumane” because that’s how those conversations (using the term loosely) would go.

      • @aidan@lemmy.world
        link
        fedilink
        English
        11 year ago

        The text that LLMs produce favors violence, because the text that humans produce (and want) favors violence.

        That’s not necessarily true, there is a lot of violent fiction.

    • @kibiz0r@midwest.social
      link
      fedilink
      English
      41 year ago

      For AGI, sure, those kinds of game theory explanations are plausible. But an LLM (or any other kind of statistical model) isn’t extracting concepts, forming propositions, and estimating values. It never gets beyond the realm of tokens.

  • @Chickenstalker@lemmy.world
    link
    fedilink
    English
    71 year ago

    It’s a WAR GAME. Emphasis on war and game. Do you chuckle fucks think wargame players should emphasize kumbaya sing dance or group therapy sessions in their games?

  • Eggyhead
    link
    fedilink
    6
    edit-2
    1 year ago

    whaaat? Robots don’t just have their own inherent sense of morality for whatever reason???

  • @General_Effort@lemmy.world
    link
    fedilink
    English
    5
    edit-2
    1 year ago

    Did nobody really question the usability of language models in designing war strategies?

    They got some nice clickbait out of it. And that’s how dumb af ideas turn into smart career moves.

    I hope no one is coming away with the idea that this about something the military is actually doing.