Comment by crote
Let's say you are making a model to judge job applicants. You are aware that the training data is biased in favor of men, so you remove all explicit mentions of gender from their CVs and cover letters.
Upon evaluation, your model seems to accept everyone who mentions a "fraternity" and reject anyone who mentions a "sorority". Swapping out the words turns a strong reject into a strong accept, and vice versa.
But you removed any explicit mention of gender, so surely your model couldn't possibly be showing an anti-women bias, right?
I've never had any implication of my gender other than my name in any CV over the past decade.
Who are these people who make a career history doc include gender-implicating data? And if there are such CVs, they should be stripped of such data before processing.
The fraternity example is such a specific 1 in a 1000 case.