• kase@lemmy.world
    link
    fedilink
    arrow-up
    2
    ·
    8 months ago

    Ah ok. I don’t know much about it, but I’ve heard that AI could sometimes be negative toward commonly discriminated against groups because the data that it’s trained with is. (Side note: is that true? someone pls correct me if it’s not). I jumped to the conclusion that this was the same thing. My bad

    • SCB@lemmy.world
      link
      fedilink
      arrow-up
      3
      ·
      8 months ago

      That is both true and pivotal to this story

      It’s a major hurdle in some uses of AI

    • adrian783@lemmy.world
      link
      fedilink
      arrow-up
      3
      ·
      8 months ago

      what it did it expose just how much inherent bias there is in hiring. even just name and gender alone.

    • TAG@lemmy.world
      link
      fedilink
      arrow-up
      1
      ·
      8 months ago

      An AI is only as good as its training data. If the data is biased, then the AI will have the same bias. The fact that going to a women’s college was considered a negative (and not simply marked down as an education of unknown quality) is proof against the idea that many in the STEM field hold (myself included) that there is a lack of qualified female candidates but not an active bias against them.