- cross-posted to:
- technology@lemmy.ml
- cross-posted to:
- technology@lemmy.ml
Contrary to Silicon Valley wisdom, training AIs on larger data sets could worsen their tendency to replicate societal biases and racist stereotypes
Contrary to Silicon Valley wisdom, training AIs on larger data sets could worsen their tendency to replicate societal biases and racist stereotypes
The problem of current LLM implementations is that they learn from scratch, like taking a baby to a library and telling him “learn, I’ll wait out in the cafeteria”.
You need a lot of data to do so, just to learn how to write, gramma, styles, concepts, relationships without any guidance.
This strategy might change in the future, but the only solution we have now is to refine the model afterward, let’s say.
Tbf biases are integral part of literature and human artistic production. Eliminating biases means having “boring” texts. Which is fine for me, but a lot of people will complain that AI is dumb and boring
Does ai also need to learn about its grandpa?