In a world where AI increasingly shapes how we see and represent ourselves, understanding how these systems interpret gender is crucial. Gender is deeply personal and much more than just anatomy, but AI systems often reduce this complexity to simple binary classifications.
To explore this topic, J. Rosenbaum, Melbourne based AI artist, researcher and lecturer shared key findings and artwork projects from their PhD focused on AI perceptions of gender. They share why gender is fundamentally misunderstood by machine learning systems, the implications for visual representation present in generative people images, and more.
Our key takeaways from J's talk, in particular around how we can fix AI's gender bias, are as follows:
Does the team building and training AI models all look like you? If yes, you might need to bring in other voices. Consider implementing participatory modelling approaches where affected communities are active partners in the AI development process, not just consulted after the fact. This helps ensure AI systems reflect diverse needs and values.
As with any data collection, consider:
When using AI to generate images of people, include diverse and intersectional representation across gender, race, ability, and socioeconomic backgrounds. Be sure to review your model outputs (especially outliers) to ensure all groups are represented appropriately.
View the slides from the talk and the full talk transcript.
Specific resources mentioned by J. during the talk are listed below:
Enjoyed watching J's Tech Leader Chat? Be sure to join our community for future talk updates.