Domain 4 Flashcards

1
Q

Responsible AI

A

is a set of guidelines and principles to ensure that AI systems operate in a safe, trustworthy, and ethical manner

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
2
Q

Fairness

A

aims to ensure that models treat everyone equitably and impartially, regardless of their age, where they live, their gender, or their ethnicity

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
3
Q

It’s important to be able to explain in human terms why a model made a particular decision,

A

Explainability

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
4
Q

Transparency

A

is about providing clear information about model capabilities, limitations, and potential risks to stakeholders. Transparency includes making sure that users know when they are interacting with AI.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
5
Q

Fairness of a model is measured by the bias and variance of outcomes across different groups.

A

the bias and variance of outcomes across different groups.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
6
Q

Overfitting becomes a problem when

A

the training dataset is not representative of the real world. As a result, the model only performs well on inputs that resemble the training data.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
6
Q

Underfitting can occur for some groups when

A

there wasn’t enough training data that matched their characteristics, so the model doesn’t perform well for them.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
7
Q

Class imbalance occurs when

A

a feature value has fewer training samples when compared with another value in the dataset. In this example, the feature for sex shows that women constitute 32.4% of the training data, whereas men constitute 67.6%.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
8
Q

These are crucial for conducting periodic reviews of datasets to identify and address potential issues or biases

A

Regular Audits

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
9
Q

Consider using these as a starting point to reduce the amount of training that your model needs, reducing your environmental impact and sustainability

A

already-trained model. Reuse of existing work is the key principle of sustainability

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
10
Q

Transparency is

A

about providing clear information about model capabilities, limitations, and potential risks. It also means making sure that users know when they are using AI.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
11
Q

Accountability

A

means establishing clear lines of responsibility for AI model outcomes and decision making.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
12
Q

Biases

A

are imbalances in data, or disparities in the performance of a model across different groups.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
13
Q

SageMaker Clarify helps you mitigate bias by

A

detecting potential bias during the data preparation, after model training, and in your deployed model, by examining specific attributes.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
14
Q

SageMaker Clarify can improve explainability by

A

looking at the inputs and outputs for your model, treating the model itself as a black box. By making these observations, it determines the relative importance of each feature.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
15
Q

How does SageMaker Clarify evaluate bias, etc?

A

SageMaker Clarify examines your dataset and model by using processing jobs. A SageMaker Clarify processing job uses the SageMaker Clarify processing container to interact with an Amazon S3 bucket. The S3 bucket would contain your input datasets, and a model that is deployed to a SageMaker inference endpoint. The SageMaker Clarify processing container obtains the input data set and configuration for analysis from an S3 bucket. For feature analysis, the SageMaker Clarify processing container sends requests to the model container, and retrieves model predictions from the response from the model container. After that step, the processing container computes and saves analysis results to the S3 bucket. These results include a JSON file with bias metrics, and global feature attributions, a visual report, and additional files for local feature attributions. You can download the results from the output location and view them.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
16
Q

SageMaker Clarify: The difference in positive proportions in predictions metric indicates whether

A

the model predicts positive outcomes differently for each class.

17
Q

specificity measures

A

how often the model correctly predicts a negative outcome

18
Q

The recall difference metric is

A

the difference in recall of the model between two classes. Any difference in these recalls is a potential form of bias.

19
Q

The accuracy difference metric

A

is the difference between the prediction accuracies for different classes.

20
Q

The treatment equality is

A

the difference in the ratio of false negatives to false positives. Even if the accuracy of the model is the same for two classes, this ratio could have differences. A difference in the type of errors that occur for different classes can constitute bias

21
Q

guardrails in Amazon Bedrock

A

You can use guardrails to define threshold for content filters for hate, insults, sexual content, or violence. You can also block topics altogether. For these topics, you can use plain text to describe the topics that should be denied.

22
Q

Guardrails can be set on both

A

the prompt and the model response, so if a prompt passes the guardrail, the response can still be blocked

23
Q

Another feature in SageMaker Clarify is the ability to run evaluation jobs of large language models so that

A

you can compare models

24
Q

SageMaker Clarify evaluation jobs can run four different types of tasks, including

A

text generation, text classification, question and answering, and text summarization.

25
Q

The five different dimensions of the four different types of SageMaker Clarify evaluation jobs

A

Prompt stereotyping measures the probability of your model including biases in its response. It includes biases for race, gender, sexual orientation, religion, age, nationality, disability, physical appearance, and socioeconomic status. Toxicity checks your model for sexual references, rude, unreasonable, hateful, or aggressive comments, profanity, insults, flirtations, attacks on identities and threats. Factual knowledge checks the veracity of the model responses. Semantic robustness checks whether your model output changes because of keyword typos, random changes to uppercase, and random additions or deletions of white spaces. Accuracy compares the model output to the expected responses, such as classifying and summarizing the data correctly.

26
Q

A model’s transparency measures

A

the degree to which ML owners and stakeholders can understand how a model works and why it produces its outputs. A model that is highly transparent uses an algorithm that is straightforward to interpret, such as linear regression.

27
Q

Transparency has two measures

A

interpretability and explainability.

28
Q

Explainability is

A

being able to describe what a model is doing without knowing exactly how. It treats the model as a black box, so every model can be observed and explained

29
Q

With interpretability

A

you can document how the inner mechanisms of the model impact the output,

30
Q

tradeoffs when choosing a model with high transparency.

A

These tradeoffs are performance and security.

31
Q

Transparent AI models are more susceptible to attacks because

A

hackers have more information about the inner mechanisms and can find vulnerabilities in the model.

32
Q

AI service cards are

A

a form of responsible AI documentation. They provide customers with a single place to learn about the intended use cases, limitations, responsible AI design choices, and deployment and performance optimization best practices,

33
Q

AI service cards currently exist for several AWS AI service APIs.

A

These APIs include matching faces with Amazon Rekognition, analyzing IDs with Amazon Textract, detecting PII with Amazon Comprehend and more. There is also an AI service card for Amazon’s foundation model in Amazon Bedrock, Amazon Titan Text

34
Q

Is there a tool that helps you create your own AI service cards build on AWS?

A

for models that you create, you can use SageMaker Model Cards to help document the lifecycle of a model from designing, building, training, and evaluation.

35
Q

Discuss two capabilities of SageMaker Clarify around responsible AI

A

SageMaker Clarify model processing jobs can also report on explainability. SageMaker Clarify provides feature attributions based on the concept of Shapley values. You can use Shapley values to determine the contribution that each feature made to the model predictions. Another type of analysis available in SageMaker Clarify is a partial dependence plot. This plot shows you how a model’s predictions changes for different values of a feature. In this case, it looks at age.

36
Q

Human-centered AI refers to

A

designing AI systems that prioritize the needs and values of humans. In human-centered AI, designers and developers engage in interdisciplinary collaboration and often involve psychologists, ethicists, and domain experts to collect diverse perspectives and expertise. Users are involved in the development process to make sure that the AI will be genuinely beneficial and user-friendly.

37
Q

Amazon Augmented AI or Amazon A2I, incorporates human review for samples of the inferences made by an AWS AI service or a custom model.

A

You can configure Amazon A2I to send inferences with low-confidence scores to human reviewers before sending them to the client. Their feedback can then be added to training data to re-train the model. Besides reviewing low-confidence inferences, you can have the human reviewers review random predictions as a way to audit the model. With Amazon A2I, you can use a pool of reviewers in your own organization or use Mechanical Turk. You can figure how many reviewers need to review each prediction.

38
Q

Reinforcement learning from human feedback, or RLHF,

A

is an industry standard technique for ensuring that large language models produce content that is truthful, harmless, and helpful.

39
Q

To use RLHF, you train a separate model which serves as a reward model.

A

The reward model is trained by humans who review multiple responses from the large language model for the same prompt and indicate their preferred response. Their preferences become the training data for the reward model, which when trained, can predict how high a human would score a prompt response. The large language model then uses the reward model to refine its responses for maximum reward.

40
Q

Collecting the preferences from humans for RLHF can be accomplished most readily with

A

SageMaker Ground Truth