Domain 1: foundations of AI Flashcards

1
Q

AI

A

Machines performing tasks normally thought to require human intelligence

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
2
Q

Turing Test

A

a machine is intelligent if it can fool humans into thinking is human

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
3
Q

Socio-technical systems

A

humans shape AI while AI shapes humans

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
4
Q

Machine Learning (ML)

A

sub-field of AI that uses algorithms to learn from input data, makes decisions/inferences/predictions, and performs tasks on new data without being explicitly programmed to do so

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
5
Q

Supervised learning overview

A
  • Requires large labeled datasets (5k+ for images)
  • Human labeled (knowingly or unknowingly)
How well did you know this?
1
Not at all
2
3
4
5
Perfectly
6
Q

Large-scale AI Open Network (LAION)

A

Has lots of large labeled datasets to use for training

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
7
Q

Supervised learning techniques (4)

A

Classification models (cat or dog)
Regression models (vehicle fuel efficiency)
Support Vector Machine (SVM): classification
Support Vector Regression (SVR): regression

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
8
Q

Unsupervised learning overview

A

extract features from data without labels, making the algorithm less predictable

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
9
Q

Unsupervised sub-cateogries

A

Clustering and association rule learning which doesn’t require the overhead of labeling data

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
10
Q

Supervised vs unsupervised

A
  • Neither is better
  • For financials:
    Supervised: find known types of fraud
    Unsupervised: find new patterns of behavior
How well did you know this?
1
Not at all
2
3
4
5
Perfectly
11
Q

Semi-supervised learning

A

Combines the benefits of both supervised and unsupervised learning to improve reliability and reduce cost of training.

Example: Large Language Models (LLMs)

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
12
Q

Reinforcement Learning

A

simulated motivation through rewards and punishment with no pre-labeled data

Uses a learning feedback loop where data is gathered through interaction with the environment

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
13
Q

Learning feedback loop

A

Action > feedback > repeat/adjust action
- Action > reward > continue/refine
- Action > punishment > change action

rewards need to be proportionate to the action

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
14
Q

Reinforcement Learning challenges

A

Works well for games

Real life is not as heavily bound as games

Time cannot be sped up

Failure may have serious consequences (self-driving cars)

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
15
Q

Discriminative model

A

classification technique to determine how data is grouped and create decision boundaries that cause the grouping

Example: label animals by type

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
16
Q

Generative model

A

generates requested data based on underlying known characteristics

Example: GenAI creating a picture of a dog

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
17
Q

Types of foundation models

A

Large Language Models (LLMs)
Vison models: interpret visual data
Scientific models: protein folding
Audio models: generate human speech, music

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
18
Q

Value of Foundational models

A
  • they save time and resources
  • generalized, adaptable, and scalable models
  • enable transfer learning, fine tuning
How well did you know this?
1
Not at all
2
3
4
5
Perfectly
19
Q

Categories of AI (3)

A

Artificial Narrow Intelligence (ANI)
Artificial General Intelligence (AGI)
Artificial Super Intelligence (ASI)

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
20
Q

Expert Systems components

A
  1. Knowledge base: organized facts for specific domains
  2. inference engine: rules-based system to locate facts for a prompt
  3. User interface: user inputs lead to system outputs
How well did you know this?
1
Not at all
2
3
4
5
Perfectly
21
Q

Fuzzy logic system (4 steps)

A
  1. Fuzzification: input converted to fuzzy data
  2. Rule eval: matches input to rule
  3. Aggregation: rule outputs combined
  4. Defuzzification: fuzzy outputs converted back to specific values

Example: automate vehicle breaking gets stronger as the car gets “closer”

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
22
Q

Common AI models (4 main groups, 3 sub groups)

A

Linear and statistical models
Decision trees
Robotics

Machine Learning > Neural networks >
- Computer Vision
- Speech recognition
- Language models

ML == hidden layers

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
23
Q

Types of neural networks (3)

A

Computer vision
Speech recognition
Natural Language Processing (NLP)

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
24
Q

AI use cases (Mnemonic)

A

Rabid = Recommendation
Raccoons = Recognition
Dance = Detection
Fancifully with = Forecasting
Giant = Goal-driven optimization
Indigo = Interaction support
Penguins = Personalization

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
25
Types of compute (3)
Serverless High-performance Trusted execution environments
26
Stages of data pipeline (4)
- Ingestion - Preparation (labeling) - Training - Output
27
When to fine tune (3)
- Adapt an existing model to a new domain - Improve task-specific performance - Customize output
28
How to fine tune (6 steps)
- start with a foundational model - gather task-specific data - pass data through the model and collect the outputs - calculate the difference with expected outputs - Tweak model parameters - Rinse and repeat until tuned
29
Types of Data transformation (6)
Data pre-processing Data post-processing Data labeling Data integrity Data drift Data observability
30
3 groups of data for AI
Training data Validation data Testing data
31
Those at risk of harm (5)
Individuals Groups Society Companies Ecosystems
32
Types of harms to individuals (4)
General (civil rights, safety, economic opportunity) Specific (employment, insurance, housing, ed) Privacy (secondary use, DSR) Economic (job displacement, bias)
33
Types of harms to groups (4)
Facial recognition Mass surveillance Civil rights Deepening of racial and socio-economic divides
34
Types of harms to Society (11)
Democratic process Trust in institutions Access to public services Employment Disinformation Misinformation Deepfakes Hallucination Echo chambers Safety Profiling
35
Types of harms to institutions (5)
Repetitional Cultural Economic Legal/regulatory Acceleration of risks
36
Types of harms to ecosystems (3)
Natural resource depletion Environment Supply chain
37
Types of Alignment (3)
Intended: desire of humans (human wants to get somewhere safe) Specified: explicitly programmed (shortest path) Emergent: actual objectives of the AI system (continuous recalculations)
38
Types of Misalignment (2)
Inner: mismatch between programed goals and what the system does (car doesn't avoid crash sitting at red light) Outer: disconnect between human intention and coded objectives (social media misinformation)
39
Types of bias (11)
Algorithmic (systematic) Computational (systemic errors) Cognitive (distorted thinking) Societal (prejudice, discrimination) Implicit (unconscious) Sampling (data sample is not representative) Temporal (model does not work consistently) Overfitting (works with training data, not real world) Underfitting (model doesn't work with any data) Edge cases / outliers (data that falls outside the boundaries of the training data) Noise (data that negatively impacts model)
40
Categories of Risks (4)
Security Operational Privacy Business KNOW THESE
41
Security risks for Generative AI (6)
Hallucination Deepfakes Adversarial attack Data poisoning Data leakage Filter bubble / Echo chamber
42
Security risks for General AI (4)
Power concentration erodes freedoms (market or gov) False sense of security (exceptionalism) Adversarial attacks System misuse (transfer learning, insecure environment)
43
Why harm taxonomies (4)
Anticipate risks Implement and monitor controls Understand legal and regulatory requirements Enhance empath for data subjects
44
MITRE PANOPTIC (harm taxonomy)
MITRE: non-profit org that manages federal R&D facilities in US PANOPTIC: Pattern and Action Nomenclature of Privacy Threats in Context - Privacy threat assessment and risk management framework - Contextual domain and privacy activities - considers benign and malicious intent
45
Ryan Calo (harm taxonomy)
subjective harms and objective harms Internal and external to the data subject
46
Danielle Citron, Daniel Solove (harm taxonomy) (9 harms)
Physical Repetitional Relationship Economic Discrimination Psychological Failure to inform Lack of control/agency Chilling effects
47
Operational risks of AI (5)
Hardware Storage High-speed network Expertise Environmental
48
Business risks of AI (7)
Reputation Culture Economic Acceleration Vendor related IP infringement Legal/Regulatory
49
Sociotechnical harms (5)
Representational Allocative Quality of service interpersonal Social systems
50
CSET AI harm taxonomy
Georgetown University Center for Security and Emerging Technology AI Incident Database (AIID) Four elements 1. entity that experienced 2. harm event or issue 3. directly linked to a consequence of behavior 4. AI system
51
Characteristics requiring Governance (8)
Complexity Interpretability Opacity Autonomy Speed and Scale Potential Harms and Misuse Data Dependency Probabilistic vs Deterministic outputs
52
Trustworthy AI principles (7)
Accountability Explainability Non-discrimination Privacy Safety Security Transparency
53
Trustworthy AI characteristics (4)
Human-centric: amplify human agency Accountable: outputs are traceable Transparent: explainable Acts in legal, fair manner
54
Developers responsiblities
- collect data required for use cases - document data provenance, lineage, and training - understand policy requirements, resources, and user experience feedback
55
Deployers responsibilities
- understand business objectives - provide users with transparency - deploy responsibly
56
Users responsiblities
- understand the AUP, business purpose - provide feedback and report incidents
57
Ethical AI: key principles (6)
Lawfulness People and planet Protection from bias and discrimination Choice over personal data Appropriate human intervention Accountability
58
Governance Framework building (Mnemonic)
Pigs = Principles Risk = Risk tolerance Injury = Industry Jumping = Jurisdiction About = Ability to implement Acrobatically = AI's purpose to the business
59
Governance Principles (7)
Pro-innovation mindset Consensus driven Outcome focused Sector agnostic Non-prescriptive Risk-centric End-to-end accountability
60
Standing up an AI Governance Body (5)
1. Understand the org 2. Get leadership buyin 3. Involve key stakeholders 4. Communicate collectively 5. Choose a governance model
61
Types of governance models
Centralized Decentralized Hybrid
62
MITRE AI Maturity Model
5 Levels: Initial, Engaged, Defined, Managed, Optimized 6 categories: Ethical, Strategy & Resources, Organization, Technology enablers, Data, Performance and Application
62
Focus of a training program
Content Characteristics (mandatory, measured) Audience
63
What is an AI Strategy
Comprehensive plan that integrates AI in support of the orgs mission, vision, and goals and aligns with the business objectives
64
AI Strategy components
- Understand business objectives - Assess data governance maturity - Develop ethical framework - Choose the right tech and tools - Prioritize AI skills development - Get leadership and employee buyin