In machine learning, proxy (sensitive attributes) refers to variables that indirectly capture information about a sensitive attribute, such as race, gender, or age, which are often used in a model to make predictions or decisions. The use of proxy variables can inadvertently lead to biased outcomes or algorithmic discrimination, even when the original sensitive attribute is not explicitly used in the model. It is important for researchers and practitioners to be aware of potential proxy variables in their data, in order to mitigate the risk of unintended consequences stemming from their use.
Identifying proxy variables for sensitive attributes is a crucial step in mitigating bias in machine learning models. There are several methods that can be employed to identify potential proxy variables, including:
Once proxy variables have been identified, there are several techniques available to address the bias they may introduce, including:
Imagine you're trying to guess which of your friends got a high score on a video game, but you're not allowed to know their actual scores. Instead, you have some other information, like how long they've been playing or what their favorite game is. These pieces of information could help you make a good guess, even though they're not the actual scores. In machine learning, these pieces of information are called "proxy variables" for the thing you're not allowed to know, like someone's race, gender, or age. Even if you don't use the actual sensitive information, the proxy variables might still lead to unfair or biased decisions. To make sure the machine is fair, researchers try to figure out which pieces of information might be acting like proxy variables and then make adjustments to reduce their impact.