• @Ultraviolet
        link
        English
        91 year ago

        There should be some sort of law where if you want to offload decisions to AI, the person who decides to let the AI make those decisions needs to step up to take full civil and criminal liability for everything it does.

        • @orrk
          link
          English
          51 year ago

          so you only really need one martyr for the cause?

          • @[email protected]
            link
            fedilink
            English
            21 year ago

            Yes, one person we can pin all of humanity’s sins on, and then we just kill them. It’s almost like a religious ritual.

        • @ForgotAboutDre
          link
          English
          21 year ago

          No, every decision maker in the chain of command should be responsible. They should know what the intelligence is based on, if the people sharing the information are competent and should be validating the information.

          Using AI to perform these tasks requires gross negligence at several stages. However, it does appear killing civilians and children is the intended outcome so negligence about AI is likely just a cover.

  • Match!!
    link
    fedilink
    English
    311 year ago

    “as humans come to rely on these systems they become cogs in a mechanised process and lose the ability to consider the risk of civilian harm in a meaningful way.”

    that’s not even an AI problem, that’s a regular society problem

    • @SkyezOpen
      link
      English
      81 year ago

      And since “ai” isn’t actual artificial intelligence but rather a neural net trained on data from that society, it’s only going to reinforce existing issues. Remember the articles about crime ai being racist? That’s because they fed policing statistics as the training set, and uh… I’m sure you know the rest.

  • @[email protected]
    link
    fedilink
    English
    111 year ago

    This is fucking insane dystopian shit, it’s worse than I thought and has become real sooner than I thought it would, bloody hell.

  • @[email protected]
    link
    fedilink
    English
    101 year ago

    Yeah we’re getting really good at teaching computers to analyze satellite imagry and other forms of spatial data to find the spots we want. All we have to do is decide if we put green spaces, Walmarts or bombs in those spots.

    • @[email protected]
      link
      fedilink
      English
      8
      edit-2
      1 year ago

      Both this use and corporate use of AI isn’t really about making things better, it’s to avoid anyone having responsibility for anything. A human might have issues with picking out a target that kills 20 innocent people on the off chance a Hamas fighter might be there, and might hold back a little bit if they might be worried the ICC could come knocking, or a critical newspaper article could come out that calls you a merchant of death. AI will pop out coordinates all day and night based on the thinnest evidence, or even no evidence at all. Same with health insurers using AI to deny coverage, AI finding suspects based on grainy CCTV footage, etc, etc.

      Nobody’s responsible, because ‘the machine did it’, and we were just following its lead. In the same way that corporations really aren’t being held responsible for crimes a private person couldn’t get away with, AI is another layer of insulation between ‘externalities’ and anyone facing consequences for them.

    • @prime_number_314159
      link
      English
      71 year ago

      My favorite ML result was (details may be inaccurate, I’m trying to recall from memory) a model that analyzed scan images from MRI machines, that would have far more confidence of the problems it was detecting if the image was taken on a machine with an old manufacture date. The training data had very few negative results from older machines, so the assumption that an image taken on an old machine showed the issue fit the data.

      There was speculation about why that would happen in the training data, but the pattern noticing machine sure noticed the pattern.

  • @Zehzin
    link
    English
    41 year ago

    That explains a lot

  • AutoTL;DRB
    link
    fedilink
    English
    41 year ago

    This is the best summary I could come up with:


    As Israel resumes its offensive after a seven-day ceasefire, there are mounting concerns about the IDF’s targeting approach in a war against Hamas that, according to the health ministry in Hamas-run Gaza, has so far killed more than 15,000 people in the territory.

    The latest Israel-Hamas war has provided an unprecedented opportunity for the IDF to use such tools in a much wider theatre of operations and, in particular, to deploy an AI target-creation platform called “the Gospel”, which has significantly accelerated a lethal production line of targets that officials have compared to a “factory”.

    The Guardian can reveal new details about the Gospel and its central role in Israel’s war in Gaza, using interviews with intelligence sources and little-noticed statements made by the IDF and retired officials.

    This article also draws on testimonies published by the Israeli-Palestinian publication +972 Magazine and the Hebrew-language outlet Local Call, which have interviewed several current and former sources in Israel’s intelligence community who have knowledge of the Gospel platform.

    In the IDF’s brief statement about its target division, a senior official said the unit “produces precise attacks on infrastructure associated with Hamas while inflicting great damage to the enemy and minimal harm to non-combatants”.

    Multiple sources told the Guardian and +972/Local Call that when a strike was authorised on the private homes of individuals identified as Hamas or Islamic Jihad operatives, target researchers knew in advance the number of civilians expected to be killed.


    The original article contains 1,734 words, the summary contains 241 words. Saved 86%. I’m a bot and I’m open source!

  • Adub
    link
    English
    -61 year ago

    Can AI target what story might next outrage people on the political fringes? I’d like to not hear about Al Jazeera and the Daily Caller any more.

    • @bowcollector
      link
      English
      01 year ago

      you didn’t even bother to read the article, eh?

    • @[email protected]
      link
      fedilink
      English
      01 year ago

      It may shock you to hear that centrists and capitalists don’t have your best interests at heart.

      • Adub
        link
        English
        -11 year ago

        Oh my goodness unplug from the matrix. Live free kind internet soul.