• @mods_are_assholes
    link
    English
    24 months ago

    You cannot upset me more than the current common misunderstandings that everyone has about AI already does.

    I don’t think you understand the implications of undetectable AI to shift social conversation or the kind of world that those AI owners want to create.

    • @[email protected]
      link
      fedilink
      English
      04 months ago

      That might actually be the kind of thing where open source AI could help. At least I hope. To detect bias, lies or AI powered filtering / sorting of content.

      • @mods_are_assholes
        link
        English
        54 months ago

        Ok so this is one of the naive thoughts that makes me upset.

        The open source community can’t even make a distro of linux that is out of the box functional for everyday users and you think somehow they are going to be able to outcompete billion dollar companies that can afford the best gear and devs?

        Look, I bought in heavy to open source early on in the 90s, and have done my best to go open source for every tool I can, but the simple fact is that even the ‘best’ open source projects are severely lacking in aspects and YOU CAN’T TRUST DEVELOPMENT OF AI TO THAT.

        Compare The Gimp to Photoshop. It isn’t even close, why? Because Adobe has a fucktonne of cash to throw at their projects and they have clear direction and motivation.

        I don’t like it

        I’d prefer a fully open source world

        But it isn’t going to happen, and open source AI will always lag behind corporate AI, and considering how fast it has been developing, even being 3 months behind renders a tool useless as an AI detector.

        We aren’t prepared for this and 90% of what everyone on the internet says about AI is poorly informed and full of confabulation, and WORST of all, when you try and explain this to them they get antagonistic.

        We have already seen the threat AI can pose in 2016 with Cambridge Analytica helping to hand trumpty dumpty the election by using AI to focus target vulnerable facebook groups.

        AND THAT AI WAS A FUCKING INFANT compared to what we have now.

        It’s going to be so bad and almost none of you have the slightest clue.

        • @[email protected]
          link
          fedilink
          English
          14 months ago

          Honestly, most of what Cambridge analytica did was blackmail, illegal spending, and collusion between campaigns that were legally required to be separate.

          Much of the data processing/ml was intended as a smoke screen to distract from the big stuff that was known to work and consequently legislated against. The problem is that they were so incompetent that the distraction technique was also illegal.

          Maybe the machine learning also worked, but it’s really not clear.

        • @[email protected]
          link
          fedilink
          English
          14 months ago

          See, THIS is the criticism of AI I can actually empathize with, I might even agree with it somewhat