Implement content moderation solutions (10–15%) Flashcards

1
Q

Which feature of Azure AI Content Safety helps protect large language models from document injection attacks?

A

Prompt Shields

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
2
Q

What is the purpose of the Groundedness detection feature in Azure AI Content Safety?

A

To verify AI-generated text is based on provided source materials.

Groundedness detection includes a reasoning option in the API response. This adds a reasoning field that explains any ungroundedness detection. However, reasoning increases processing time and costs.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
3
Q

Which social media issues does Azure AI Content Safety address?

A

The growth of inappropriate online content including bullying and hate speech.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
4
Q

How does Azure AI Content Safety help businesses to protect their brand image?

A

By moderating comments and messages from customers.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
5
Q

What does moderate text do?

A

scans text across four categories: violence, hate speech, sexual content, and self-harm. A severity level from 0 to 6 is returned for each category. This level helps to prioritize what needs immediate attention by people, and how urgently. You can also create a blocklist to scan for terms specific to your situation.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
6
Q

Define prompt shields

A

a unified API to identify and block jailbreak attacks from inputs to LLMs. It includes both user input and documents. These attacks are prompts to LLMs that attempt to bypass the model’s in-built safety features. User prompts are tested to ensure the input to the LLM is safe. Documents are tested to ensure they don’t contain unsafe instructions embedded within the text.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
7
Q

What does protected material do?

A

checks AI-generated text for protected text such as recipes, copyrighted song lyrics, or other original material.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
8
Q

What image content moderation options exist?

A
  1. moderate images
  2. Moderate multimodal content
How well did you know this?
1
Not at all
2
3
4
5
Perfectly
9
Q

What happens in moderate images?

A

scans for inappropriate content across four categories: violence, self-harm, sexual, and hate. A severity level is returned: safe, low, or high. You then set a threshold level of low, medium, or high. The combination of the severity and threshold level determines whether the image is allowed or blocked for each category.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
10
Q

How does Moderate multimodal content differ from moderate images?

A

scans both images and text, including text extracted from an image using optical character recognition (OCR). Content is analyzed across four categories: violence, hate speech, sexual content, and self-harm.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
11
Q

Describe custom categories

A

enables you to create your own categories by providing positive and negative examples, and training the model. Content can then be scanned according to your own category definitions.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
12
Q

What is the purpose of the safety system message?

A

Safety system message helps you to write effective prompts to guide an AI system’s behavior.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
13
Q

How should content moderation be used?

A

Azure AI Content Safety works best to support human moderators who can resolve cases of incorrect identification. When people add content to a site, they don’t expect posts to be removed without reason. Communicating with users about why content is removed or flagged as inappropriate helps everyone to understand what is permissible and what isn’t.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly