AI Challenges and Responsibilities Flashcards
What is explainability in Responsible AI?
The understanding of the nature and behavior of an ML model, explaining outputs without knowing internal workings.
What does interpretability mean in Responsible AI?
A human can understand the cause of a model’s decision, answering ‘why’ and ‘how’.
What does privacy and security mean in Responsible AI?
Ensuring individuals control when and if their data is used by models.
What is transparency in Responsible AI?
Being open about how AI models work and how decisions are made.
What is veracity and robustness in Responsible AI?
The system remains reliable even in unexpected or adverse conditions.
What is AI governance?
Policies, processes, and tools that ensure AI is developed and used responsibly.
What is the aim of safety in Responsible AI?
To ensure algorithms are safe and beneficial to individuals and society.
What does controllability in Responsible AI refer to?
Aligning AI models to human values and intent.
What is Amazon Bedrock Guardrails used for?
Filtering content, redacting PII, enhancing safety, and blocking harmful content.
What can SageMaker Clarify evaluate?
Accuracy, robustness, toxicity, and bias in foundational models.
How does Data Wrangler help with bias?
Using Augment Data to balance datasets by generating new instances for underrepresented groups.
What does SageMaker Model Monitor do?
Performs quality analysis on models in production.
What is Amazon A2I (Augmented AI) used for?
Allows human review of ML predictions with low confidence.
What does SageMaker Role Manager help with?
Implements user-level security for model governance.
What is the purpose of Model Cards in AWS?
To document models including use cases, limitations, and metrics.
What are AWS AI Service Cards?
Responsible AI documentation for AWS services with use cases, limitations, and design choices.
What is a high-interpretability model example?
A decision tree – easy to interpret and visualize.
What is a trade-off in model interpretability and performance?
Higher interpretability often means lower performance and vice versa.
What is Human-Centered Design (HCD) in Responsible AI?
Designing AI systems to prioritize human needs.
What does amplified decision making focus on in HCD?
Designing for clarity, simplicity, and usability in high-pressure decisions.
What is unbiased decision making in HCD?
Recognizing and mitigating bias in datasets and decision processes.
What is cognitive apprenticeship in HCD?
AI learns from human experts (e.g., RLHF), and humans learn from AI with personalization.
What is user-centered design in Responsible AI?
Ensuring a wide range of users can access and benefit from AI systems.
What makes it difficult to regulate Generative AI?
Its complexity, black-box nature, and non-deterministic outputs make regulation challenging.
What are examples of social risks associated with Generative AI?
Spreading misinformation and contributing to toxic or biased content.
What is non-determinism in GenAI?
It means the same prompt can produce different outputs each time.
What is toxicity in the context of GenAI?
Generating offensive, inappropriate, or disturbing content.
What makes defining toxicity difficult?
It’s subjective and depends on cultural, contextual, and situational factors.
How can toxicity be mitigated?
By curating training data and using guardrails to filter outputs.
What are hallucinations in GenAI?
Factual errors where the model confidently generates false information.
What causes hallucinations in LLMs?
The probabilistic nature of next-word prediction, not fact-checking.
How can hallucinations be reduced?
User education, independent verification, and marking outputs as unverified.
What is plagiarism in the GenAI context?
Copying or producing content without attribution or original effort.
What is prompt misuse (poisoning)?
Introducing harmful or biased data to influence model behavior.
What is prompt injection?
Embedding hidden instructions in input prompts to manipulate model output.
What is an example of a harmful prompt injection?
Asking the model to generate malware or biased essays.
What is exposure in GenAI?
Unintended output of sensitive or private user data.
What is prompt leaking?
Revealing previously submitted prompts, potentially disclosing private information.
What is jailbreaking a GenAI model?
Bypassing built-in ethical or safety filters to unlock restricted behaviors.
What is few-shot jailbreaking?
Using a few crafted prompt-output pairs to trick the model into responding inappropriately.
What is many-shot jailbreaking?
Using a large number of prompt-output examples to weaken or bypass safety mechanisms.
Why is detecting AI-generated content important?
To prevent plagiarism, misinformation, and protect authenticity.
What’s a challenge with detecting AI-generated outputs?
It’s difficult to trace the source or verify accuracy of generated content.
What is an example of GenAI being misused in education?
Students using LLMs to write essays and cheat on assignments.
How do some tools detect AI-generated content?
By analyzing writing patterns, metadata, or embedded watermarks.
Which industries typically require an extra level of compliance when it comes to AI workloads?
Industries such as financial services, healthcare, and aerospace require stricter compliance measures.
What does it mean for a workload to be considered ‘regulated’ in terms of compliance?
A regulated workload must adhere to specific rules, audits, archival requirements, and security standards set by regulatory bodies.
What are some challenges of implementing compliance in AI systems?
Challenges include complexity and opacity, dynamism, emergent capabilities, algorithmic and human biases, and difficulty in auditing AI decision-making.
What is meant by ‘dynamism and adaptability’ in AI systems, and why does it present a compliance challenge?
AI systems continually evolve and adapt, meaning their decision-making processes may change over time, complicating audit and compliance efforts.
What are emergent capabilities in AI, and how can they impact compliance?
Emergent capabilities are unintended functionalities that develop in a system, potentially introducing risks beyond the original scope and complicating regulatory oversight.
How can algorithmic bias affect compliance in AI?
Algorithmic bias can perpetuate discrimination or unfair outcomes if the training data isn’t representative, which violates fairness and non-discrimination regulations.
What is a model card, and why is it important for AI compliance?
A model card is a standardized document that details key aspects of a machine learning model—its training data, performance metrics, biases, and limitations—to support transparency and audits.
How do AWS services help in achieving compliance for regulated workloads?
AWS provides over 140 security standards and compliance certifications (e.g., NIST, ISO, HIPAA, GDPR, PCI), as well as tools like SageMaker model cards and service cards to document AI systems.
What role do service cards play in compliance for AWS AI services?
AWS AI Service Cards provide documentation on intended use cases, limitations, design choices, and performance optimizations, aiding in transparency and regulatory compliance.
Why is documenting data sources and licensing important for AI compliance?
Documenting the origin, licenses, and potential biases in training data is crucial for demonstrating compliance, ensuring data quality, and identifying any risk factors.
What are some specific compliance frameworks mentioned in the lecture?
Examples include standards from NIST, the EU Agency for Cybersecurity, ISO, SOC, HIPAA, GDPR, and PCI DSS.
What is the connection between human bias and AI system compliance?
Human bias introduced during model development can affect outputs, and if unchecked, may result in unfair or non-compliant AI decisions.
What is the bottom line when it comes to implementing compliance on AI according to the lecture?
Understanding your regulatory obligations and leveraging appropriate tools and documentation on platforms like AWS is essential for compliance.
What are the main AI governance strategies?
Policies, Review Cadence, Review Strategy, Safety Checks, Decision-Making Framework, Transparency Standards, Stakeholder Feedback, Training and Certification
What are the key data governance strategies?
Responsible AI Framework, Data Governance Council, Data Sharing and Collaboration, Data Lifecycle Management, Data Logging, Data Monitoring, Data Analysis, Data Retention, Data Lineage, Data Cataloging
What are AWS tools that support AI governance and compliance?
AWS Config, Amazon Inspector, AWS Audit Manager, AWS Artifact, AWS CloudTrail, AWS Trusted Advisor
What are core policy areas for responsible AI?
Data Management, Model Training, Output Validation, Safety, Human Oversight, Bias Mitigation, Privacy Protection, Intellectual Property
What are types of AI reviews?
Technical Review, Non-Technical Review
What are key roles in AI/data governance?
Governance Board, Subject Matter Experts, Legal & Compliance Teams, Data Stewards, Data Owners, Data Custodians
What are data lifecycle stages in governance?
Collection, Processing, Storage, Consumption, Archival
What are key AI security strategies?
Threat Detection, Vulnerability Management, Infrastructure Protection, Prompt Injection Prevention, Data Encryption
What are key AI monitoring areas?
Model Performance Metrics, Infrastructure Monitoring, Bias and Fairness, Compliance and Responsible AI
What are important data security engineering practices?
Data Quality Assessment, Data Profiling and Monitoring, Data Lineage
What are privacy enhancing technologies?
Data Masking, Data Obfuscation, Encryption, Tokenization
What are key data access control strategies?
Role-Based Access Control, Fine-Grained Permissions, IAM Solutions, SSO and MFA, Access Logging and Review
What are core data integrity strategies?
Backup and Recovery, Audit Trails, Lineage Tracking, Integrity Monitoring and Testing
What is the main goal of MLOps?
To ensure models are deployed, monitored, and retrained systematically—not just developed once and forgotten.
What are the key principles of MLOps?
Version control, automation, CI/CD, continuous retraining, continuous monitoring
What should be version controlled in MLOps?
Data, code, and models
Why is automation important in MLOps?
To streamline data ingestion, preprocessing, training, evaluation, deployment, and monitoring
What is the purpose of continuous monitoring in MLOps?
To detect model drift, bias, and performance issues in production
What are the key stages of an ML pipeline?
Data preparation, model building, evaluation, selection, deployment, and monitoring
What types of repositories are used in MLOps?
Data repository, code repository, and model registry (with version control)
What is a benefit of automating MLOps processes?
Increased confidence in model development and deployment