Measuring Algorithmic Bias In Automated Classification Flashcards

1
Q

How can discrimination in databases be uncovered?

A

Discrimination in databases can be uncovered by polling people for perceptions of discrimination, studying potential discriminators, and statistical analysis.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
2
Q

What is the Discrimination Discovery Task?

A

The Discrimination Discovery Task is to find discriminatory situations and practices given a database of historical decision records and a set of potentially discriminated groups.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
3
Q

What are the potentially discriminated or protected groups?

A

Protected groups are groups defined by socially salient characteristics that experience disadvantages, such as sex, sexual orientation, ethnicity, national origin, age, disability status, and intersections of these groups.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
4
Q

What are the challenges in discovering discrimination in databases?

A

contextualized non-discrimination requirements,

different conceptualizations of discrimination,

different metrics and criteria,

high dimensionality data,

the possibility of hidden indirect discrimination

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
5
Q

What is the relationship between data mining and discrimination?

A

Data mining can perpetuate discrimination if it is based on biased data or if it amplifies and reinforces existing societal biases.

For example, if a dataset used to make hiring decisions is biased against certain groups of people, such as women or people of color, data mining algorithms trained on that data will learn and perpetuate those biases.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
6
Q

What are the two examples of algorithmic fairness?

A

The two examples of algorithmic fairness are group fairness and individual fairness.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
7
Q

What is an example method for group fairness?

A

Classification Rules Mining is an example method for group fairness.

Identifying patterns in data that can be used to create fair decision-making rules that treat different groups equally.

For example, in the context of hiring decisions, Classification Rules Mining could be used to analyze the factors that lead to successful job performance and identify rules that prioritize these factors in the hiring process, while ensuring that these rules are applied equally to all demographic groups.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
8
Q

What are directly discriminatory rules?

A

Directly discriminatory rules are rules that discriminate against a protected group based on a database of past decisions.

Directly discriminatory rules are rules that explicitly use a protected attribute (such as race, gender, or age) to make a decision or classification about an individual or group, resulting in unfair treatment or exclusion based on that attribute.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
9
Q

What are indirectly discriminatory rules?

A

Indirectly discriminatory rules are rules that discriminate against a protected group based on an indirect relationship to a protected attribute.

A job requirement for a certain number of years of experience in a particular industry, which may disproportionately affect younger job applicants.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
10
Q

What are genuine occupational requirements?

A

A genuine occupational requirement (GOR) is a job requirement that is essential for a particular job and is justifiable on non-discriminatory grounds;

for example, only hiring female models to model women’s clothing in a fashion show would be a genuine occupational requirement.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
11
Q

What is the Discrimination Discovery Task limited to?

A

Limited to identifying the presence of bias and discrimination in machine learning models and does not address the root causes or solutions for these issues.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
12
Q

What is an example method for individual fairness?

A

We want to use KNN to predict which applicants are likely to be hired by a company, while ensuring individual fairness based on gender.

To achieve this, we can first pre-process the data by creating two separate datasets, one for male applicants and one for female applicants. Then, we can apply KNN separately to each dataset, using a common value for k.

Finally, we can combine the results of the two KNN models to obtain a final prediction that is individually fair with respect to gender.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
13
Q

What is situational testing?

A

Situational testing is an approach for creating controlled experiments in which matched pairs undergo the same situation.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
14
Q

What is k-NN as situation testing?

A

k-NN as situation testing is an algorithm that measures the degree of discrimination of the decision for r by looking at its k closest neighbors

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
15
Q

Question: What are some predictive tasks that can be performed with given data about a defendant in the criminal justice system?

A

Answer: Some predictive tasks that can be performed include the probability of general recidivism, probability of violent recidivism, probability of violence against others in prison, probability of self-harm in prison, and probability of breaking permits.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
16
Q

Question: What is the process of transitioning from human “clinical” judgment to actuarial systems in risk assessment?

A

Answer: The process involves transitioning from human “clinical” judgment to structured human judgment by guiding experts through items that are good predictors, which improves predictive power and increases agreement between experts. Then, the structured human judgment is translated to actuarial systems that use some sort of scoring, with or without models.

17
Q

What is SAVRY?

A

SAVRY is a structured risk assessment tool for juvenile offenders that considers 24 risk factors and 6 protective factors.

18
Q

How is the final assessment of risk level done with SAVRY?

A

The final assessment of risk level (low/medium/high) is done by an expert after seeing a score from an actuarial system.

19
Q

What is one potential issue with training labels for risk assessment?

A

Training labels might be biased due to differences in policing and conviction rates across communities.

20
Q

What is an example of how privilege and marginalization can affect risk assessment?

A

Privileged drug addicts may be seen as sick, while marginalized drug addicts may be seen as criminals.

21
Q

What is one reason that violent crime prediction is often studied in risk assessment?

A

Police often make arrests in cases of violent crime, giving them more data to work with for prediction.

22
Q

What is one consideration when evaluating the costs of false positives and false negatives in risk assessment?

A

The social and individual costs of detaining a low-risk individual (false positive) or releasing a high-risk individual (false negative) should be taken into account.

23
Q

What is the concept of independence and separation?
Independence and separation are two fairness criteria that cannot be achieved simultaneously.

A
24
Q

What is the proof that independence and separation cannot be achieved simultaneously?

A

Given (pa ≠ pb)∧(FNRa = FNRb) ⇒ FPRa ≠ FPRb.

25
Q

What is the definition of the generalized false positive rate?

A

The generalized false positive rate of a classifier is E[ R | Y=0 ].

26
Q

What is the definition of the generalized false negative rate?

A

The generalized false negative rate of a classifier is E[ 1-R | Y=1 ].

27
Q

What is the relationship between generalized false positive rate and generalized false negative rate of a calibrated classifier?

A

The generalized false positive rate and generalized false negative rates of a calibrated classifier depend linearly on each other.

28
Q

What is the incompatibility of calibration with separation?

A

The calibration can be manipulated to ensure separation but at the cost of fairness.

29
Q

What is the example of how calibration can be manipulated?

A

People can be grouped in a deceptive way (high risk and low risk together) to ensure calibration but also to ensure everybody is below a threshold of risk.

30
Q

What is the concept of infra-marginality?

A

The evaluation of the algorithm involves people who are unlikely to be affected by changes.

31
Q

What is the concept of infra-marginality?

A

The evaluation of the algorithm involves people who are unlikely to be affected by changes.

32
Q

What is the concept of observational criteria?

A

The observational criteria cannot help us distinguish between some fair and unfair situations.

33
Q

What are the three questions we should ask when considering race in machine learning?

A

(1) How was the software designed?
(2) Are the input data racialized?
(3) Is the system output racialized?

34
Q

What is the conclusion about many algorithmic fairness criteria?

A

Many algorithmic fairness criteria cannot be satisfied simultaneously.

35
Q

What is the conclusion about many algorithmic fairness criteria being easily manipulated?

A

Many algorithmic fairness criteria can be easily manipulated.

36
Q

What is the conclusion about observational criteria?

A

Observational criteria cannot help us distinguish between some fair and unfair situations.