Some image generators produce more problematic stereotypes than others, but all fail at diversity
(algorithmwatch.org)
from JRepin@lemmy.ml to technology@lemmy.world on 05 Nov 2023 09:28
https://lemmy.ml/post/7524034
from JRepin@lemmy.ml to technology@lemmy.world on 05 Nov 2023 09:28
https://lemmy.ml/post/7524034
cross-posted from: lemmy.ml/post/7481270
Automated image generators are often accused of spreading harmful stereotypes, but studies usually only look at MidJourney. Other tools make serious efforts to increase diversity in their output, but effective remedies remain elusive.
threaded - newest
If you use google images to do basically the same searches you get the same diversity issues. It’s reflecting the training data, and the larger world by extension. Whatever they would have us do to fix that must be applied to reality before it can or should be artificially skewed in AI models. Because if you bias the model to compensate you will create a worse bias. One that was intentional.
Even if you don’t agree with that take, have a look at the Firefly example. they asked for a trucker named Paul, and they got a woman in the result set. Maybe somewhere out there exists a woman trucker named Paul, but it’s a clear reduction in accuracy and quality because Adobe attempted to inject artificial diversity.
Yes, but on the other hand biasing the models could be a way to influence reality.
Could be, maybe. Or maybe not. Not sure. But the thing for sure is that forcing the diversity reduces the quality of the model.
Yeah, AI generated images reflect various biases from training data
Who engages in activity X the most
And photographs themselves
And posts those photographs online
And labels them in a way that an AI might put correlate them
If most champaigne pictures are taken with selfie making white girls and eating waffles are black families (something I actually ran across earlier on), that’s going to be the bias in the AI images as well