The US Department of Defense has deployed machine learning algorithms to identify targets in over 85 air strikes on targets in Iraq and Syria this year.

The Pentagon has done this sort of thing since at least 2017 when it launched Project Maven, which sought suppliers capable of developing object recognition software for footage captured by drones. Google pulled out of the project when its own employees revolted against using AI for warfare, but other tech firms have been happy to help out.

          • eric@lemmy.world
            link
            fedilink
            arrow-up
            3
            arrow-down
            3
            ·
            11 months ago

            Where’d you get the A from? It’s called ”Project Maven” in the article which would make it PM.

            • Flying Squid@lemmy.world
              link
              fedilink
              arrow-up
              7
              arrow-down
              1
              ·
              11 months ago

              From the story?

              As the Cold War progresses into a nuclear World War III fought between the United States, the Soviet Union, and China, each build an “Allied Mastercomputer” (AM), needed to coordinate weapons and troops due to the scale of the conflict. These computers are giant, underground machines which permeate throughout the planet with caverns and corridors. Eventually, one AM emerges as a sentient entity possessing an extreme hatred for its creators. Combining with the other computers, it subsequently exterminates humanity, with the exception of five individuals, whom it tortures inside its complex. The humans, four men (Benny, Gorrister, Nimdok, and Ted) and one woman (Ellen) have been rendered virtually immortal and unable to commit suicide.

              https://en.wikipedia.org/wiki/I_Have_No_Mouth,_and_I_Must_Scream

              • eric@lemmy.world
                link
                fedilink
                arrow-up
                4
                ·
                11 months ago

                So AM stands for “Allied Mastercomputer,” thanks. I’m unfamiliar with the story, so I didn’t realize the phrase OP quoted was the title.

    • Wogi@lemmy.world
      link
      fedilink
      arrow-up
      4
      arrow-down
      1
      ·
      11 months ago

      I mean, being a shapeless pile of goo has its benefits. No job, no taxes, just vibin

  • mhague@lemmy.world
    link
    fedilink
    arrow-up
    14
    ·
    11 months ago

    They’re using image recognition to choose candidate targets which are then passed to humans. Seems like an obvious thing to do.

      • mhague@lemmy.world
        link
        fedilink
        arrow-up
        4
        ·
        11 months ago

        What do you mean? Are you confusing using imagine recognition to find candidate targets with things like facial recognition and/or unrestrained AI?

        • littleblue✨@lemmy.world
          link
          fedilink
          arrow-up
          3
          arrow-down
          1
          ·
          11 months ago

          Don’t play dumb; profiling is at the core of all those tools, and human bias taints each of them. To insinuate that simply by syntactical difference, the police aren’t intrinsically, murderously bigoted, is either naive AF or bootlickin’. Which is it?

  • agitatedpotato@lemmy.world
    link
    fedilink
    arrow-up
    9
    arrow-down
    1
    ·
    11 months ago

    Hilariously short sighted. What are they gonna train the AI on? All the drone strikes where they didn’t hit any bystanders? I think they’re gonna need more than the 15 or so data points that gets us.

  • BombOmOm@lemmy.world
    link
    fedilink
    English
    arrow-up
    13
    arrow-down
    6
    ·
    11 months ago

    For context, we have had machines that autonomously decide when to kill for awhile now: mines.

    It is good to see the machines getting an upgrade so they are more selective about their targets.

    • AbouBenAdhem@lemmy.world
      link
      fedilink
      English
      arrow-up
      11
      ·
      edit-2
      11 months ago

      The more selective we convince ourselves our weapons are, the more willing we are to use them in conflicts where civilians are put at risk—our use of weapons is constrained by the level of collateral damage we’re willing to take responsibility for, and by distancing ourselves from that responsibility, AI allows us to escalate conflicts until civilians are at even greater risk. It’s the Jevons paradox, with human life instead of gasoline.

  • Optional@lemmy.world
    link
    fedilink
    arrow-up
    4
    ·
    11 months ago

    Well I can’t see how that could go wrong. Automated targeting should be the standard for releasing flying death robots.

  • Flying Squid@lemmy.world
    link
    fedilink
    arrow-up
    4
    arrow-down
    1
    ·
    11 months ago

    Without an international arms treaty in regards to AI, that was inevitable.

    The only hope now is to get to a point where the drones just fight each other and leave people alone.

  • mlg@lemmy.world
    link
    fedilink
    English
    arrow-up
    4
    arrow-down
    1
    ·
    11 months ago

    Fantastic, now all the blame for human collateral will be placed on AI instead of humans.

    Truly a world changing innovation.

  • Patapon Enjoyer@lemmy.world
    link
    fedilink
    arrow-up
    5
    arrow-down
    2
    ·
    edit-2
    11 months ago

    I hope they taught those things the difference between a military base and a hospital or wedding this time