Some image generators produce more problematic stereotypes than others, but all fail at diversity (algorithmwatch.org)
from JRepin@lemmy.ml to technology@lemmy.world on 05 Nov 2023 09:28
https://lemmy.ml/post/7524034

cross-posted from: lemmy.ml/post/7481270

Automated image generators are often accused of spreading harmful stereotypes, but studies usually only look at MidJourney. Other tools make serious efforts to increase diversity in their output, but effective remedies remain elusive.

#technology

threaded - newest

burliman@lemm.ee on 05 Nov 2023 09:54 next collapse

If you use google images to do basically the same searches you get the same diversity issues. It’s reflecting the training data, and the larger world by extension. Whatever they would have us do to fix that must be applied to reality before it can or should be artificially skewed in AI models. Because if you bias the model to compensate you will create a worse bias. One that was intentional.

Even if you don’t agree with that take, have a look at the Firefly example. they asked for a trucker named Paul, and they got a woman in the result set. Maybe somewhere out there exists a woman trucker named Paul, but it’s a clear reduction in accuracy and quality because Adobe attempted to inject artificial diversity.

jungle@lemmy.world on 05 Nov 2023 10:28 collapse

Yes, but on the other hand biasing the models could be a way to influence reality.

burliman@lemm.ee on 06 Nov 2023 01:22 collapse

Could be, maybe. Or maybe not. Not sure. But the thing for sure is that forcing the diversity reduces the quality of the model.

phx@lemmy.ca on 05 Nov 2023 17:37 collapse

Yeah, AI generated images reflect various biases from training data

  • Who engages in activity X the most

  • And photographs themselves

  • And posts those photographs online

  • And labels them in a way that an AI might put correlate them

If most champaigne pictures are taken with selfie making white girls and eating waffles are black families (something I actually ran across earlier on), that’s going to be the bias in the AI images as well