• @[email protected]
    link
    fedilink
    381 year ago

    Garbage headline. This isn’t “AI” doing this, it’s hiring managers and companies. It’s policy. If I put all my applicants into a Microsoft excel spreadsheet and use the sorting function to sort by race, then only hire ones of of a particular skin tone, is Excel keeping millions of qualified candidates out of the workforce? No, of course not. Neither is AI. Replace “AI” with “company policy” in every one of these articles and you get at what’s actually occurring.

    Same reason we don’t need to “regulate AI”. We need to regulate it’s deployment, just like we regulate whatever technology we used for it previously. In other words, we don’t need new rules, we just need enforcement of existing ones. You can’t have a hiring process that discriminatory. What tool you use to arrive at that end doesn’t matter.

    • @Poayjay
      link
      English
      161 year ago

      I completely disagree. It absolutely is AI doing this. The point the article is trying to make is that the data used to train the AI is full of exclusionary hiring practices. AI learns this and carries it forward.

      Using your metaphor, it would be like training AI on hundreds of excel spreadsheets that were sorted by race. The AI learns this and starts doing it too.

      This touches on one of the huge ethical questions with regulating AI. If you are discriminated against in a job hunt by an AI, who’s fault is that? The AI is just doing what it’s taught. The company is just doing what the AI said. The AI developers are just giving it previous hiring data. If the previous hiring data is racist or sexist or whatever you can’t retroactively correct that. This is exactly why we need to regulate AI not just its deployment.

      • @[email protected]
        link
        fedilink
        6
        edit-2
        1 year ago

        “This touches on one of the huge ethical questions with regulating AI. If you are discriminated against in a job hunt by an AI, who’s fault is that?” It is the fault of the company hiring practices, which are to blindly trust an AI without testing whether or not it is discriminatory. It is also the fault of the producer of that AI software (or service) sold to the company for screening candidates. No new laws are needed to hold either of them accountable, existing laws cover the ground well. That company selling the AI screening services could have just been called “crystal ball hiring” before AI and would be equally liable if they just discriminated in their hiring suggestions. The tool isn’t the thing that needs regulating, the actions people and companies take based on the tool is. And that is already well regulated.

        Make an AI in the privacy of your own home that does ____ literally anything? Fine. Collaborate making an OSS AI to do whatever with some of your friends? Also fine. Sell that AI as a employment screener app? Better make sure you’ve tested it to not have discriminatory outcomes. Use that AI to screen employees? Same deal.

      • @SheeEttin
        link
        English
        21 year ago

        It’s still the fault of the person using it. If you feed it racist data, or if it produces racist output and you use it, you are the person responsible.

        If an AI told you to jump off a bridge, would you do it?

        • @Poayjay
          link
          English
          01 year ago

          The person using the tool doesn’t know if it’s bias. If you say using previous hiring data and this stack of resumes, pick the best candidate, how are you responsible? You didn’t generate the training data. You didn’t write any code. You didn’t hire the previous employees. That’s the issue.

          Do you think that a company will manually review a thousand resumes to not be biased?

          • @virr
            link
            English
            11 year ago

            That’s literally the hiring manager job. Sort however as many resumes they get in an unbiased way to get good candidates.

            If they sort them in a racist way then the company is liable for those actions (sometimes individual can also be liable). Doesn’t matter how they sorted to get there, they’re still liable. Blaming AI is just a way to shift liability when they get caught. Trying to shift liability might because they are incompetent or they are racist. Hold them accountable in either case.

          • @SheeEttin
            link
            English
            11 year ago

            That doesn’t matter. Regardless of the tool you use, if it produces a racist result, you are responsible for that.

            To use a more extreme example, let’s say you’re an engineer. If you use an AI to design a bridge, then that bridge collapses and kills people, you don’t get to say “well the AI did it!”. No. You chose to use the tool, you chose to use its output.

            If you’re going to use any kind of automation, you are responsible for ensuring that it doesn’t run afoul of any laws or other regulations. If you can’t do that, don’t use it.

      • squiblet
        link
        fedilink
        11 year ago

        The “AI” is simply a program used to filter data. Whose fault is it if using it causes problems? The nitwits who choose to use these programs and trust the results without understanding the limitations.

        • donuts
          link
          fedilink
          21 year ago

          The “guns don’t kill people, people kill people” argument, eh?

          • squiblet
            link
            fedilink
            31 year ago

            Not really the most precise analogy, but sure. Individuals and companies are responsible for actions, not inanimate objects. However, that line of thinking is used as an argument against gun regulations. “AI” is a much broader field and more general tool than a firearm, and also newer and less well understood. People seem to be on average confused by the hype and unclear on what these systems can do well, or not, and what are appropriate uses. I’d hate to see legislators who know barely anything about technology write laws to restrict it prematurely, or more likely, adopt laws written directly by companies and industry groups who could have various hidden motives. For example we have seen both Sam Altman and Musk mention restricting AI research, and in both cases, the real motive seemed to be not safety but obtaining an advantage for their own companies.

            For now, I agree with the other poster who said existing laws are sufficient. If a company was to discriminate in hiring, it doesn’t matter whether they used a special program to make the decisions or not.

    • @[email protected]
      link
      fedilink
      41 year ago

      AI is the new scapegoat over immigrants on who to blame for jobs being lost (even though it’s always the capitalist owner that drive these decisions)