AI Bias Flashcards
Skewed Sample
Occurs when the training data doesn’t fairly represent all groups or situations.
Ex: If an AI is mostly trained in English text from North America, it might not work well with content from other regions or cultures.
Limited Features/Sample Size Disparity
Occurs when some types of data are much more common than others in training set.
Ex: AI has access to millions of formal English, but not many examples of casual or spoken english.
Tainted Examples
Occurs when training data includes biased or incorrect info.
Ex: If AI learns from old texts with outdated views about certain groups of people, it may repeat those biased ideas.
Proxy Bias
Occurs when seemingly natural info in the data is acually linked to sensitive topics.
Ex: AI might learn to associate certain universities (which may be linked to races or social class) with job stability, even if race or class aren’t mentioned in the data.
Fairness Metric
Way to measure how fair an AI system is.
Protected Class
A general category of people who share a charcacteristic that is legally protected against discrimination.
Protected Feature
Specific instances or attributes within a protected class. These are the particular thins about a person that shouldn’t infuence the AI’s decisions unfairly.
Pre-processing
Changing or fixing the training data
In-processing
Charging how the AI learns during tarining to make it fair
Post-Processing
Changing AI’s outputs after it has made a decision, to make the results fairer.