- cross-posted to:
- technology@lemmy.ml
- cross-posted to:
- technology@lemmy.ml
Contrary to Silicon Valley wisdom, training AIs on larger data sets could worsen their tendency to replicate societal biases and racist stereotypes
The problem of current LLM implementations is that they learn from scratch, like taking a baby to a library and telling him “learn, I’ll wait out in the cafeteria”.
You need a lot of data to do so, just to learn how to write, gramma, styles, concepts, relationships without any guidance.
This strategy might change in the future, but the only solution we have now is to refine the model afterward, let’s say.
Tbf biases are integral part of literature and human artistic production. Eliminating biases means having “boring” texts. Which is fine for me, but a lot of people will complain that AI is dumb and boring
Does ai also need to learn about its grandpa?
Chatgpt seems pretty smart. Why couldn’t the ai detect and reject racism?
Because it’s not smart. It does not have intelligence. It simply attempts to imitate human language.
And it’s very good at pretending to be smart.