Questões gerais Flashcards

1
Q

Training data is best defined as a subset of data that is used to?

A. Enable a model to detect and learn patterns.
B. Fine-tune a model to improve accuracy and prevent overfitting.
C. Detect the initial sources of biases to mitigate prior to deployment.
D. Resemble the structure and statistical properties of production data.

A

A. Enable a model to detect and learn patterns.
Training data is used to enable a model to detect and learn patterns. During the training phase, the model learns from the labeled data, identifying patterns and relationships that it will later use to make predictions on new, unseen data. This process is fundamental in building an AI model’s capability to perform tasks accurately. Reference: AIGP Body of Knowledge on Model Training and Pattern Recognition.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
2
Q

Testing data is best defined as a subset of data that is used to?

A. Assess a model’s on-going performance in production.
B. Enable a model to discover and learn patterns.
C. Provide a robust evaluation of a final model.
D. Evaluate a model’s handling of randomized edge cases.

A

C. Provide a robust evaluation of a final model.

Testing data is a subset of data used to provide a robust evaluation of a final model. After training the model on training data, it is essential to test its performance on unseen data (testing data) to ensure it generalizes well to new, real-world scenarios. This step helps in assessing the model’s accuracy, reliability, and ability to handle various data inputs.
Reference: AIGP Body of Knowledge on Model Validation and Testing.

Explanation:
Training and testing data serve distinct purposes in the machine learning (ML) workflow, and testing data specifically is designed to evaluate the performance of a trained model.

Assess a model’s on-going performance in production (A):
This refers to monitoring in production environments, not testing during development. Testing data is used prior to deployment to validate the model’s accuracy and generalization, not for ongoing production evaluation.

Enable a model to discover and learn patterns (B):
This describes the purpose of training data, which is used during the training phase to allow the model to learn patterns and relationships in the data. Testing data, by contrast, is not used for learning.

Provide a robust evaluation of a final model (C):
Testing data is a reserved subset of the data used to evaluate the model’s performance after training. It helps measure how well the model generalizes to unseen data, ensuring it performs robustly on new or unknown cases.

Evaluate a model’s handling of randomized edge cases (D):
While testing data may include edge cases, its primary purpose is broader: to evaluate overall model performance. Edge-case testing is typically a more specific task within robustness testing or adversarial testing.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
3
Q

To maintain fairness in a deployed system, it is most important to?

A. Protect against loss of personal data in the model.
B. Monitor for data drift that may affect performance and accuracy.
C. Detect anomalies outside established metrics that require new training data.
D. Optimize computational resources and data to ensure efficiency and scalability

A

B. Monitor for data drift that may affect performance and accuracy.

To maintain fairness in a deployed system, it is crucial to monitor for data drift that may affect performance and accuracy. Data drift occurs when the statistical properties of the input data change over time, which can lead to a decline in model performance. Continuous monitoring and updating of the model with new data ensure that it remains fair and accurate, adapting to any changes in the data distribution. Reference: AIGP Body of Knowledge on Post-Deployment Monitoring and Model Maintenance.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
4
Q

When monitoring the functional performance of a model that has been deployed into production, all of the following are concerns EXCEPT?

A. Feature drift.
B. System cost.
C. Model drift.
D. Data loss.

A

Correct Answer: B system cost
When monitoring the functional performance of a model deployed into production, concerns typically include feature drift, model drift, and data loss. Feature drift refers to changes in the input features that can affect the model’s predictions. Model drift is when the model’s performance degrades over time due to changes in the data or environment. Data loss can impact the accuracy and reliability of the model. However, system cost, while important for budgeting and financial planning, is not a direct concern when monitoring the functional performance of a deployed model. Reference: AIGP Body of Knowledge on Model Monitoring and Maintenance.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
5
Q

After completing model testing and validation, which of the following is the most important step that an organization takes prior to deploying the model into production?

A Perform a readiness assessment.
B Define a model-validation methodology.
C Document maintenance teams and processes.
D Identify known edge cases to monitor post-deployment.

A

Correct Answer: A Perform a readiness assessment.
After completing model testing and validation, the most important step prior to deploying the model into production is to perform a readiness assessment. This assessment ensures that the model is fully prepared for deployment, addressing any potential issues related to infrastructure, performance, security, and compliance. It verifies that the model meets all necessary criteria for a successful launch. Other steps, such as defining a model-validation methodology, documenting maintenance teams and processes, and identifying known edge cases, are also important but come secondary to confirming overall readiness. Reference: AIGP Body of Knowledge on Deployment Readiness.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
6
Q

Which type of existing assessment could best be leveraged to create an Al impact assessment?

A. A safety impact assessment.
B. A privacy impact assessment.
C. A security impact assessment.
D. An environmental impact assessment.

A

Correct Answer: B. A privacy impact assessment.
A privacy impact assessment (PIA) can be effectively leveraged to create an AI impact assessment. A PIA evaluates the potential privacy risks associated with the use of personal data and helps in implementing measures to mitigate those risks. Since AI systems often involve processing large amounts of personal data, the principles and methodologies of a PIA are highly applicable and can be extended to assess broader impacts, including ethical, social, and legal implications of AI. Reference: AIGP Body of Knowledge on Impact Assessments.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
7
Q

What is the primary purpose of an Al impact assessment?

A. To define and evaluate the legal risks associated with developing an Al system.
B. Anticipate and manage the potential risks and harms of an Al system.
C. To define and document the roles and responsibilities of Al stakeholders.
D. To identify and measure the benefits of an Al system.

A

B. Anticipate and manage the potential risks and harms of an AI system.

Explanation:
The primary purpose of an AI impact assessment is to identify, evaluate, and manage the potential risks and harms associated with the deployment and use of an AI system. This process helps ensure that the AI system is developed and used in a way that minimizes negative consequences and aligns with ethical and legal standards.

Key aspects of an AI impact assessment include:

Identifying potential risks: Understanding how the AI system could cause harm to individuals, groups, or society.
Managing risks: Developing strategies to mitigate those risks and ensure that the AI system is safe, fair, and aligned with the organization’s values.
Considering broader impacts: Taking into account the social, ethical, and environmental implications of deploying the AI system.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
8
Q

Which of the following steps occurs in the design phase of the Al life cycle?

A. Data augmentation.
B. Model explainability.
C. Risk impact estimation.
D. Performance evaluation.

A

C. Risk impact estimation.

In the design phase, the focus is on planning and identifying potential risks and impacts of the AI system. Risk impact estimation involves assessing the potential consequences of deploying the model, including ethical, legal, and operational risks. The other steps typically occur in later stages of the AI life cycle:

A. Data augmentation happens during the data preparation phase.
B. Model explainability is often addressed during model development or validation.
D. Performance evaluation occurs after the model is trained, during testing and validation.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
9
Q

During the planning and design phases of the Al development life cycle, bias can be reduced by all of the following EXCEPT?

A. Stakeholder involvement.
B. Feature selection.
C. Human oversight.
D. Data collection.

A

B. Feature selection.

While feature selection is an important step in AI model development, it typically occurs during the modeling phase, not the planning or design phases. Bias can be reduced during planning and design through A. Stakeholder involvement, C. Human oversight, and D. Data collection, which ensure that diverse perspectives and appropriate data are considered early on. Feature selection focuses more on refining the model’s inputs and is not directly related to bias reduction at the planning and design stages.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
10
Q

Which of the following use cases would be best served by a non-AI solution?

A. A non-profit wants to develop a social media presence.
B. An e-commerce provider wants to make personalized recommendations.
C. A business analyst wants to forecast future cost overruns and underruns.
D. A customer service agency wants automate answers to common questions.

A

A. A non-profit wants to develop a social media presence.

Building a social media presence typically involves content creation, scheduling posts, and engagement strategies, which can be handled effectively with standard tools and human effort rather than requiring AI. The other use cases—such as personalized recommendations, forecasting, and automating customer service—are more suited to AI-driven solutions that can leverage data and machine learning models.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
11
Q

All of the following are elements of establishing a global Al governance infrastructure EXCEPT?

A. Providing training to foster a culture that promotes ethical behavior.
B. Creating policies and procedures to manage third-party risk.
C. Understanding differences in norms across countries.
D. Publicly disclosing ethical principles.

A

Answer : D Publicly disclosing ethical principles.

Establishing a global AI governance infrastructure involves several key elements, including providing training to foster a culture that promotes ethical behavior, creating policies and procedures to manage third-party risk, and understanding differences in norms across countries. While publicly disclosing ethical principles can enhance transparency and trust, it is not a core element necessary for the establishment of a governance infrastructure. The focus is more on internal processes and structures rather than public disclosure. Reference: AIGP Body of Knowledge on AI Governance and Infrastructure.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
12
Q

Which of the following would be the least likely step for an organization to take when designing an integrated compliance strategy for responsible Al?

A. Conducting an assessment of existing compliance programs to determine overlaps and integration points.
B. Employing a new software platform to modernize existing compliance processes across the organization.
C. Consulting experts to consider the ethical principles underpinning the use of Al within the organization.
D. Launching a survey to understand the concerns and interests of potentially impacted stakeholders.

A

Answer : B. Employing a new software platform to modernize existing compliance processes across the organization.

When designing an integrated compliance strategy for responsible AI, the least likely step would be employing a new software platform to modernize existing compliance processes. While modernizing compliance processes is beneficial, it is not as directly related to the strategic integration of ethical principles and stakeholder concerns. More critical steps include conducting assessments of existing compliance programs to identify overlaps and integration points, consulting experts on ethical principles, and launching surveys to understand stakeholder concerns. These steps ensure that the compliance strategy is comprehensive and aligned with responsible AI principles. Reference

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
13
Q

A company initially intended to use a large data set containing personal information to train an Al model. After consideration, the company determined that it can derive enough value from the data set without any personal information and permanently obfuscated all personal data elements before training the model.

This is an example of applying which privacy-enhancing technique (PET)?

A Anonymization.
B Pseudonymization.
C Differential privacy.
D Federated learning.

A

A. Anonymization.

Justification:
Definition of Anonymization:

Anonymization is the process of irreversibly transforming personal data so that individuals can no longer be identified, directly or indirectly. In this case, the company permanently obfuscated all personal data elements, ensuring that the data set no longer contains any personally identifiable information (PII).
Key Characteristics of Anonymization:

The process is irreversible.
The data set cannot be used to identify individuals, even when combined with other data sets.
It ensures compliance with privacy laws like GDPR, which treats anonymized data as no longer subject to data protection regulations.
Why not the other options?
B. Pseudonymization:

Pseudonymization replaces personal identifiers with pseudonyms (e.g., a unique ID) but does not make the data irreversible. Pseudonymized data can still be linked back to individuals with additional information, unlike anonymization.
C. Differential privacy:

Differential privacy involves adding statistical noise to the data to protect individual privacy while allowing insights at an aggregate level. It does not obfuscate or remove personal data entirely.
D. Federated learning:

Federated learning trains machine learning models across multiple decentralized data sets without sharing raw data. It does not involve obfuscating or removing personal data in a single data set.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
14
Q

The planning phase of the Al life cycle articulates all of the following EXCEPT the?

A Objective of the model.
B Approach to governance.
C Choice of the architecture.
D Context in which the model will operate.

A

Answer : B Approach to governance.

The planning phase of the AI life cycle typically includes defining the objective of the model, choosing the appropriate architecture, and understanding the context in which the model will operate. However, the approach to governance is usually established as part of the overall AI governance framework, not specifically within the planning phase. Governance encompasses broader organizational policies and procedures that ensure AI development and deployment align with legal, ethical, and operational standards

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
15
Q

What is the best reason for a company adopt a policy that prohibits the use of generative Al?

A. Avoid using technology that cannot be monetized.
B. Avoid needing to identify and hire qualified resources.
C. Avoid the time necessary to train employees on acceptable use.
D. Avoid accidental disclosure to its confidential and proprietary information.

A

Correct Answer: D Avoid accidental disclosure to its confidential and proprietary information.

The primary concern for a company adopting a policy prohibiting the use of generative AI is the risk of accidental disclosure of confidential and proprietary information. Generative AI tools can inadvertently leak sensitive data during the creation process or through data sharing. This risk outweighs the other reasons listed, as protecting sensitive information is critical to maintaining the company’s competitive edge and legal compliance. This rationale is discussed in the sections on risk management and data privacy in the IAPP AIGP Body of Knowledge.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
16
Q

Which of the following is an example of a high-risk application under the EU Al Act?

A. A resume scanning tool that ranks applicants.
B. An Al-enabled inventory management tool.
C. A government-run social scoring tool.
D. A customer service chatbot tool.

A

Correct Answer: C A government-run social scoring tool.
The EU AI Act categorizes certain applications of AI as high-risk due to their potential impact on fundamental rights and safety. High-risk applications include those used in critical areas such as employment, education, and essential public services. A government-run social scoring tool, which assesses individuals based on their social behavior or perceived trustworthiness, falls under this category because of its profound implications for privacy, fairness, and individual rights. This contrasts with other AI applications like resume scanning tools or customer service chatbots, which are generally not classified as high-risk under the EU AI Act.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
17
Q

All of the following are penalties and enforcements outlined in the EU Al Act EXCEPT?

A. Fines for SMEs and startups will be proportionally capped.
B. Rules on General Purpose Al will apply after 6 months as a specific provision.
C. The Al Pact will act as a transitional bridge until the Regulations are fully enacted.
D. Fines for violations of banned Al applications will be €35 million or 7% global annual turnover (whichever is higher).

A

C. The AI Pact will act as a transitional bridge until the Regulations are fully enacted.

The EU AI Act outlines specific penalties and enforcement mechanisms to ensure compliance with its regulations. Among these, fines for violations of banned AI applications can be as high as €35 million or 7% of the global annual turnover of the offending organization, whichever is higher. Proportional caps on fines are applied to SMEs and startups to ensure fairness. General Purpose AI rules are to apply after a 6-month period as a specific provision to ensure that stakeholders have adequate time to comply. However, there is no provision for an “AI Pact” acting as a transitional bridge until the regulations are fully enacted, making option C the correct answer.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
18
Q

According to the EU Al Act, providers of what kind of machine learning systems will be required to register with an EU oversight agency before placing their systems in the EU market?

A. Al systems that are harmful based on a legal risk-utility calculation.
B. Al systems that are “strong” general intelligence.
C. Al systems trained on sensitive personal data.
D. Al systems that are high-risk.

A

D. AI systems that are high-risk.

Explanation:
The EU AI Act introduces a regulatory framework aimed at ensuring the safe and responsible deployment of AI systems in the European Union. A key provision of the Act is the classification of AI systems into risk categories: unacceptable risk, high risk, limited risk, and minimal risk.

High-Risk AI Systems:
Definition of High-Risk AI Systems:
AI systems are considered high-risk if they:

Affect fundamental rights, health, safety, or access to opportunities.
Are used in critical areas such as healthcare, law enforcement, education, employment, and biometric identification.
Registration Requirement:
Providers of high-risk AI systems must:

Register their systems in an EU database managed by an oversight agency before placing them on the EU market.
Demonstrate compliance with strict requirements, including risk management, data governance, transparency, and human oversight.
Why the Other Options Are Incorrect:
AI systems that are harmful based on a legal risk-utility calculation (A):
While harm is a consideration, the Act focuses on predefined risk categories rather than requiring a general risk-utility calculation. “High-risk” classification depends on the system’s application and sector.

AI systems that are “strong” general intelligence (B):
The Act does not specifically regulate systems with “strong” or “general” intelligence. Current regulations are focused on specific use cases and risks rather than theoretical advancements in AI.

AI systems trained on sensitive personal data (C):
The Act regulates how personal data is processed within AI systems but does not require registration solely based on the type of training data. Compliance with the GDPR governs data protection aspects.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
19
Q

All of the following may be permissible uses of an AI system under the EU AI Act EXCEPT?

A. To detect an individual’s intent for law enforcement purposes.
B. To promote equitable distribution of welfare benefits.
C. To implement social scoring.
D. To manage border control.

A

C. To implement social scoring.

Justification:
Prohibition of Social Scoring:

The EU AI Act explicitly prohibits the use of AI systems for social scoring, especially when it involves evaluating individuals based on behavior, predicted personality traits, or social circumstances in ways that result in discriminatory or unfair treatment.
Permissible Uses Under the EU AI Act:

A. Law enforcement purposes: AI can be used under strict regulations for specific law enforcement purposes, such as detecting intent, provided it complies with safeguards.
B. Welfare distribution: AI may assist in ensuring equitable welfare distribution by analyzing eligibility or managing resources.
D. Border control: AI systems can be deployed for border management tasks like verifying identities or analyzing risks, subject to safeguards against misuse.
Why Social Scoring is the Exception:

Social scoring, often associated with surveillance and discriminatory practices (e.g., the “credit score” systems used in some regions), is inconsistent with EU principles of fairness, privacy, and non-discrimination.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
20
Q

What is the best method to proactively train an LLM so that there is mathematical proof that no specific piece of training data has more than a negligible effect on the model or its output?

A Clustering.
B Transfer learning.
C Differential privacy.
D Data compartmentalization.

A

C. Differential privacy.

Explanation:
Differential privacy is the best method to ensure that no specific piece of training data has a significant effect on the model or its output. This technique involves adding noise to the data or the training process in a controlled manner, such that it becomes mathematically provable that the model’s output does not change significantly due to the inclusion or exclusion of any single data point.

Key reasons why differential privacy is suitable:

It provides mathematical guarantees that the contribution of individual data points is limited.
It helps ensure data privacy because the model cannot be used to infer whether any specific data point was present in the training set.
Here’s why the other options are less suitable:

A. Clustering: Clustering is a method for grouping similar data points together but does not inherently protect individual data points’ influence on the model or provide mathematical guarantees about privacy.

B. Transfer learning: Transfer learning involves using a pre-trained model and fine-tuning it on new data, but it does not focus on ensuring that individual data points have a minimal impact on the overall model output.

D. Data compartmentalization: This is a method for organizing and isolating data into segments but does not directly address controlling the influence of specific data points on the model.

Differential privacy is specifically designed for scenarios where it is important to ensure that the presence or absence of any single piece of data cannot be detected or inferred from the model, making it the best choice for this purpose.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
21
Q

Machine learning is best described as a type of algorithm by which?

A. Systems can mimic human intelligence with the goal of replacing humans.
B. Systems can automatically improve from experience through predictive patterns.
C. Statistical inferences are drawn from a sample with the goal of predicting human intelligence.
D. Previously unknown properties are discovered in data and used to predict and make improvements in the data.

A

B. Systems can automatically improve from experience through predictive patterns.

Explanation:
Machine learning (ML) is a branch of artificial intelligence (AI) that focuses on building algorithms and systems that can learn from data and improve their performance over time without being explicitly programmed for each specific task. ML algorithms learn from past data (experience) to identify patterns and make predictions or decisions.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
22
Q

You asked a generative Al tool to recommend new restaurants to explore in Boston, Massachusetts that have a specialty Italian dish made in a traditional fashion without spinach and wine. The generative Al tool recommended five restaurants for you to visit.
After looking up the restaurants, you discovered one restaurant did not exist and two others did not have the dish.

This information provided by the generative Al tool is an example of what is commonly called?
A. Prompt injection.
B. Model collapse.
C. Hallucination.
D. Overfitting.

A

C. Hallucination.

Explanation:
In the context of AI, hallucination refers to instances where a generative AI model produces information that is false, inaccurate, or fabricated. This means the model might generate responses that seem plausible or detailed but are not grounded in reality.

In your case, the generative AI tool recommended a restaurant that does not exist and suggested dishes that were not actually available at the other restaurants. This is a classic example of hallucination, where the model produces responses based on patterns it has learned, even though those responses do not correspond to real-world facts.

Here’s why the other options are incorrect:

A. Prompt injection: This occurs when a user manipulates the prompt to alter or exploit the AI’s behavior. It’s not relevant here, as the issue is about the AI providing inaccurate information, not about how the prompt influenced it.

B. Model collapse: This refers to a situation where a model’s performance deteriorates over time, often due to training issues. It’s not related to the generation of incorrect information.

D. Overfitting: Overfitting happens when a model learns too closely from its training data, resulting in poor performance on new, unseen data. It is not related to the generation of false information like recommending non-existent restaurants.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
23
Q

Each of the following actors are typically engaged in the Al development life cycle EXCEPT?

A. Data architects.
B. Government regulators.
C. Socio-cultural and technical experts.
D. Legal and privacy governance experts.

A

B. Government regulators.

Explanation:
In the context of the AI development life cycle, various stakeholders are typically involved, such as:

A. Data architects: They play a crucial role in designing the data infrastructure, preparing and structuring data, and ensuring it is suitable for training and testing AI models.

C. Socio-cultural and technical experts: These experts help ensure that the AI system is developed with consideration for its social and cultural impact and that it aligns with technical best practices and societal values.

D. Legal and privacy governance experts: These professionals ensure that the AI system complies with laws and regulations regarding data privacy, security, and ethical considerations throughout its development.

B. Government regulators, however, are generally not directly involved in the AI development process itself. Instead, they play a role in setting standards, creating regulations, and ensuring compliance after the AI system is deployed or during audits. They might interact with organizations to ensure adherence to laws, but they are not typically part of the internal development process.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
24
Q

A company is working to develop a self-driving car that can independently decide the appropriate route to take the driver after the driver provides an address.

If they want to make this self-driving car “strong” Al, as opposed to “weak,” the engineers would also need to ensure?

A. That the Al has full human cognitive abilities that can independently decide where to take the driver.
B. That they have obtained appropriate intellectual property (IP) licenses to use data for training the Al.
C. That the Al has strong cybersecurity to prevent malicious actors from taking control of the car.
D. That the Al can differentiate among ethnic backgrounds of pedestrians.

A

A. That the AI has full human cognitive abilities that can independently decide where to take the driver.

Explanation:
The distinction between “strong” AI (also known as Artificial General Intelligence, or AGI) and “weak” AI (also known as narrow AI) lies in the scope of cognitive abilities.

Weak AI is designed to perform a specific task or set of tasks, such as driving a car or playing chess. It does not possess general understanding or reasoning beyond its designated functions.

Strong AI, or AGI, would have the ability to understand, learn, and reason across a wide range of topics, similar to a human. It would be capable of making decisions autonomously in a manner that reflects broad human-like understanding.

In the context of a self-driving car, making the car “strong” AI would require it to have the capability to independently decide where to take the driver even without a specific address, reflecting human-like judgment and understanding of complex situations.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
25
Q

Which of the following is NOT a common type of machine learning?
A. Deep learning.
B. Cognitive learning.
C. Unsupervised learning.
D. Reinforcement learning

A

B. Cognitive learning.

Explanation:
Cognitive learning is not a standard term used to describe a type of machine learning. It generally refers to human learning processes, such as understanding, applying knowledge, and thinking. It is not specifically related to machine learning algorithms or methods.

The other options are common types of machine learning:

A. Deep learning: A subset of machine learning that uses neural networks with many layers (deep neural networks) to learn from large amounts of data. It is particularly effective in complex tasks like image and speech recognition.

C. Unsupervised learning: A type of machine learning where the model is trained on data without labeled outcomes. It is used to find patterns or groupings within the data, such as clustering and association.

D. Reinforcement learning: A type of machine learning where an agent learns to make decisions by taking actions in an environment to maximize cumulative rewards. It is commonly used in robotics, game playing, and autonomous systems.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
26
Q

An EU bank intends to launch a multi-modal Al platform for customer engagement and automated decision-making assist with the opening of bank accounts. The platform has been subject to thorough risk assessments and testing, where it proves to be effective in not discriminating against any individual on the basis of a protected class.

What additional obligations must the bank fulfill prior to deployment?

A. The bank must obtain explicit consent from users under the privacy Directive.
B. The bank must disclose how the Al system works under the Ell Digital Services Act.
C. The bank must subject the Al system an adequacy decision and publish its appropriate safeguards.
D. The bank must disclose the use of the Al system and implement suitable measures for users to contest

A

D. The bank must disclose the use of the AI system and implement suitable measures for users to contest.

Explanation:
Under the EU AI Act and other relevant EU regulations, when deploying an AI system that is used in high-stakes contexts like customer engagement and automated decision-making for opening bank accounts, the bank has certain obligations:

Transparency: The bank is required to disclose to customers that an AI system is being used in the decision-making process. This ensures that users are aware that decisions affecting them are partially or wholly automated.

User Rights: The bank must also implement mechanisms for users to contest decisions made by the AI system. This means that if a customer disagrees with a decision made by the AI (e.g., rejection of a bank account application), they should have a way to seek a human review or appeal the decision.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
27
Q

Random forest algorithms are in what type of machine learning model?

A. Symbolic.
B. Generative.
C. Discriminative.
D. Natural language processing.

A

C. Discriminative.

Explanation:
Random forest algorithms fall under the category of discriminative models in machine learning. Discriminative models are designed to classify or predict a target outcome by learning the boundary between different classes based on the features in the data.

Here’s why the other options are not correct:

A. Symbolic: Symbolic AI involves rule-based systems where knowledge is encoded in symbols and rules. Random forests do not follow this approach; they are based on data-driven learning of decision trees.

B. Generative: Generative models focus on modeling the joint probability of the input features and the output labels, allowing them to generate new data instances. Random forests do not attempt to model the joint probability; instead, they learn to differentiate between classes based on input features.

D. Natural language processing: Natural Language Processing (NLP) is a field of AI focused on interactions between computers and human language. Random forest is a type of algorithm that can be applied to NLP tasks, but it is not a category of machine learning itself.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
28
Q

Under the NIST Al Risk Management Framework, all of the following are defined as characteristics of trustworthy Al EXCEPT?

A. Tested and Effective.
B. Secure and Resilient.
C. Explainable and Interpretable.
D. Accountable and Transparent.

A

A. Tested and Effective.

Explanation:
Under the NIST AI Risk Management Framework (NIST AI RMF), the focus is on ensuring that AI systems are developed and deployed in a way that makes them trustworthy. Trustworthiness is defined through several key characteristics, including:

B. Secure and Resilient: Ensuring that AI systems are protected against adversarial attacks, vulnerabilities, and can recover from unexpected events is a key aspect of trustworthiness.

C. Explainable and Interpretable: It is important for AI systems to provide outputs that can be understood and explained to human users, especially in high-stakes environments. This ensures that stakeholders understand how decisions are made.

D. Accountable and Transparent: Trustworthy AI systems require clear accountability structures and transparency around how decisions are made, ensuring that stakeholders can understand and hold the AI system accountable.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
29
Q

Pursuant to the White House Executive Order of November 2023, who is responsible for creating guidelines to conduct red-teaming tests of Al systems?

A. National Institute of Standards and Technology (NIST).
B. National Science and Technology Council (NSTC).
C. Office of Science and Technology Policy (OSTP).
D. Department of Homeland Security (DHS).

A

A. National Institute of Standards and Technology (NIST).

Explanation:
According to the White House Executive Order on AI issued in November 2023, the National Institute of Standards and Technology (NIST) is tasked with developing guidelines for conducting red-teaming tests of AI systems. These guidelines are intended to provide a framework for testing and evaluating the robustness, security, and trustworthiness of AI systems, particularly to identify vulnerabilities and risks associated with their deployment.

Red-teaming involves subjecting AI models to rigorous testing, often simulating adversarial conditions, to assess their performance under various challenging scenarios. NIST’s role is to ensure that these guidelines are comprehensive and aligned with standards that promote the safe and responsible use of AI.

The other options are less relevant for this specific responsibility:

B. National Science and Technology Council (NSTC): This body coordinates science and technology policy across federal agencies but is not specifically tasked with creating guidelines for red-teaming.

C. Office of Science and Technology Policy (OSTP): The OSTP plays a role in setting overall policy direction and priorities for AI but does not directly create testing guidelines like those developed by NIST.

D. Department of Homeland Security (DHS): The DHS is involved in matters of national security and could be concerned with the implications of AI in that context but is not responsible for creating technical testing guidelines for AI systems.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
30
Q

According to November 2023 White House Executive Order, which of the following best describes the guidance given to governmental agencies on the use of generative AI as a workplace tool?

A. Limit access to specific uses of generative AI.
B. Impose a general ban on the use of generative AI.
C. Limit access of generative AI to engineers and developers.
D. Impose a ban on the use of generative AI in agencies that protect national security.

A

A. Limit access to specific uses of generative AI.

Justification:
White House Executive Order on AI Guidance:

The November 2023 White House Executive Order emphasizes responsible use of generative AI, focusing on limiting its use to specific applications that align with governmental priorities and ensuring its deployment is ethical, secure, and fair.
Context of Guidance:

Rather than implementing a blanket ban, the guidance seeks to control specific use cases to minimize risks such as misuse, security breaches, or ethical concerns.
Why Not Other Options:

B. Impose a general ban: The order does not call for a general ban but promotes responsible and controlled use.
C. Limit access to engineers and developers: The focus is on use-case restrictions, not limiting it to specific roles.
D. Impose a ban on use in agencies that protect national security: Instead of banning use in certain agencies, the order likely includes additional security protocols for high-risk environments.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
31
Q

The White House Executive Order from November 2023 requires companies that develop dual-use foundation models to provide reports to the federal government about all of the following EXCEPT?

A. Any current training or development of dual-use foundation models.
B. The results of red-team testing of each dual-use foundation model.
C. Any environmental impact study for each dual-use foundation model.
D. The physical and cybersecurity protection measures of their dual-use foundation models.

A

C. Any environmental impact study for each dual-use foundation model.

Explanation:
The Executive Order issued by the White House in October 2023 mandates that companies developing dual-use foundation models provide reports to the federal government on several aspects of their AI systems. These requirements include:

Ongoing or Planned Activities: Companies must disclose any ongoing or planned activities related to the training, development, or production of dual-use foundation models.
ARNOLD & PORTER

Red-Team Testing Results: Developers are required to report the outcomes of red-team testing—structured efforts to identify flaws and vulnerabilities in AI systems—based on guidelines developed by the National Institute of Standards and Technology (NIST).
MOFO

Physical and Cybersecurity Measures: Companies must detail the physical and cybersecurity protections implemented to safeguard the integrity of the training process against potential threats.
DECHERT

However, the Executive Order does not require companies to provide reports on environmental impact studies for each dual-use foundation model. While environmental considerations are important, they are not specified as a reporting requirement in this particular Executive Order.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
32
Q

What is the primary purpose of conducting ethical red-teaming on an Al system?

A. To improve the model’s accuracy.
B. To simulate model risk scenarios.
C. To identify security vulnerabilities.
D. To ensure compliance with applicable law.

A

B. To simulate model risk scenarios.

Justification:
Definition of Ethical Red-Teaming:

Ethical red-teaming involves testing an AI system by simulating adversarial or high-risk scenarios to identify potential weaknesses, risks, or unintended consequences. This process helps evaluate how the AI system behaves under different stress conditions, edge cases, or malicious misuse.
Primary Purpose:

The primary objective is to simulate risk scenarios to uncover issues related to fairness, robustness, security, and ethical implications. These scenarios might include bias exploitation, adversarial attacks, or inappropriate outputs in sensitive contexts.
Proactive Risk Mitigation:

By identifying potential risks early, ethical red-teaming helps organizations address vulnerabilities before the AI system is deployed or scaled, thus ensuring safer and more responsible AI usage.
Why not the other options?
A. To improve the model’s accuracy:

While ethical red-teaming can indirectly lead to improved performance, its primary goal is not to enhance accuracy but to uncover risks and vulnerabilities.
C. To identify security vulnerabilities:

Security vulnerabilities are one aspect of ethical red-teaming, but the process is broader and also focuses on ethical, fairness, and societal risks.
D. To ensure compliance with applicable law:

Compliance may be an outcome of addressing risks uncovered through red-teaming, but it is not the primary focus. Ethical red-teaming is about proactively identifying risks, which may or may not be directly tied to legal compliance.

33
Q

What is the main purpose of accountability structures under the Govern function of the NIST Al Risk Management Framework?

A. To empower and train appropriate cross-functional teams.
B. To establish diverse, equitable and inclusive processes.
C. To determine responsibility for allocating budgetary resources.
D. To enable and encourage participation by external stakeholders.

A

A. To empower and train appropriate cross-functional teams.

Explanation:
Under the Govern function of the NIST AI Risk Management Framework (AI RMF), accountability structures are designed to ensure that there are clear roles, responsibilities, and processes in place for managing AI risks. This includes empowering and training cross-functional teams who are responsible for overseeing the AI system’s lifecycle, from development to deployment and monitoring.

Key aspects of accountability structures include:

Empowering teams: Ensuring that the teams responsible for AI systems have the authority and resources they need to carry out their responsibilities effectively.
Training: Providing the necessary training to cross-functional teams, including data scientists, legal experts, and compliance officers, so they understand the risks and responsibilities associated with the AI system.
Clarity of roles: Defining who is responsible for different aspects of the AI system, such as data management, ethics, and compliance.

34
Q

Which of the following most encourages accountability over Al systems?

A. Determining the business objective and success criteria for the Al project.
B. Performing due diligence on third-party Al training and testing data.
C. Defining the roles and responsibilities of Al stakeholders.
D. Understanding Al legal and regulatory requirements.

A

C. Defining the roles and responsibilities of AI stakeholders.

Explanation:
Accountability in the context of AI systems means ensuring that individuals or groups are responsible for the various aspects of the AI system, including its design, deployment, and impact. Defining roles and responsibilities of stakeholders is crucial for creating clear lines of accountability. This ensures that everyone involved knows their duties and obligations regarding the AI system’s performance, monitoring, and ethical considerations. It also helps ensure that if something goes wrong, there are clear points of contact for taking corrective action.

35
Q

All of the following are common optimization techniques in deep learning to determine weights that represent the strength of the connection between artificial neurons EXCEPT?

A. Gradient descent, which initially sets weights arbitrary values, and then at each step changes them.
B. Momentum, which improves the convergence speed and stability of neural network training.
C. Autoregression, which analyzes and makes predictions about time-series data.
D. Backpropagation, which starts from the last layer working backwards.

A

C. Autoregression, which analyzes and makes predictions about time-series data.

Explanation:
Autoregression is not a common optimization technique used to determine weights in deep learning models. It is a statistical method often used for time-series analysis, where future values are predicted based on past values. While autoregression can be used in time-series forecasting, it is not a technique used for optimizing weights in neural networks.

36
Q

What is the technique to remove the effects of improperly used data from an ML system?

A. Data cleansing.
B. Model inversion.
C. Data de-duplication.
D. Model disgorgement.

A

D. Model disgorgement.

Explanation:
Model disgorgement refers to the process of removing or undoing the influence of improperly used or compromised data on a trained machine learning model. This might be necessary when a model has been trained using data that was obtained or used inappropriately (e.g., data that violates privacy laws or was collected without proper consent).

In such cases, the organization might be required to retrain the model from scratch without the tainted data or modify the existing model to remove the impact of the improperly used data. Model disgorgement is a technique often discussed in regulatory and compliance contexts, especially when data privacy violations have occurred.

Here’s why the other options are not suitable:

A. Data cleansing: This involves correcting or removing errors or inconsistencies from a dataset before it is used for training a model. It does not address the issue of a model that has already been influenced by improperly used data.

B. Model inversion: This is a technique that attempts to reconstruct input data from a trained model, often used as a privacy attack technique. It is not related to removing the effects of bad data from a model.

C. Data de-duplication: This is the process of removing duplicate records from a dataset. It helps in improving data quality but does not address the issue of data that was improperly used in training.

37
Q

What is the term for an algorithm that focuses on making the best choice achieve an immediate objective at a particular step or decision point, based on the available information and without regard for the longer-term best solutions?

A. Single-lane.
B. Optimized.
C. Efficient.
D. Greedy.

A

D. Greedy.

Explanation:
A greedy algorithm is one that makes the best immediate choice at each step or decision point, aiming to optimize the solution for that specific moment. It focuses on achieving the local optimum at each step without considering the broader, long-term consequences or whether that immediate decision leads to the overall best solution.

Greedy algorithms are often used in optimization problems where the goal is to find a solution quickly, though they may not always guarantee the global optimal solution for the entire problem.

38
Q

All of the following are reasons to deploy a challenger Al model in addition a champion Al model EXCEPT to?

A. Provide a framework to consider alternatives to the champion model.
B. Automate real-time monitoring of the champion model.
C. Perform testing on the champion model.
D. Retrain the champion model.

A

D. Retrain the champion model.

Explanation:
Deploying a challenger model alongside a champion model is typically done for the following reasons:

A. Provide a framework to consider alternatives to the champion model: This is a core purpose of having a challenger model. It allows for testing new models against the champion to see if they offer improved performance.

B. Automate real-time monitoring of the champion model: While not the main reason for deploying a challenger model, the comparison between the challenger and champion can inform monitoring decisions. However, direct real-time monitoring would be done through other tools, not solely through the presence of a challenger model.

C. Perform testing on the champion model: A challenger model helps test and validate the performance of the champion model by providing a benchmark for comparison. This ensures that the champion model remains effective over time.

D. Retrain the champion model is not a reason to deploy a challenger model. Retraining is a process where the champion model is updated or improved based on new data, not something that requires the presence of a challenger model. The purpose of a challenger model is to provide an alternative for evaluation, not to directly trigger or facilitate the retraining of the existing champion model.

39
Q

You are part of your organization’s ML engineering team and notice that the accuracy of a model that was recently deployed into production is deteriorating.

What is the best first step address this?

A. Replace the model with a previous version.
B. Conduct champion/challenger testing.
C. Perform an audit of the model.
D. Run red-teaming exercises.

A

B. Conduct champion/challenger testing.

When the accuracy of a model deteriorates, the best first step is to conduct champion/challenger testing. This involves deploying a new model (challenger) alongside the current model (champion) to compare their performance. This method helps identify if the new model can perform better under current conditions without immediately discarding the existing model. It provides a controlled environment to test improvements and understand the reasons behind the deterioration. This approach is preferable to directly replacing the model, performing audits, or running red-teaming exercises, which may be subsequent steps based on the findings from the champion/challenger testing.

40
Q

All of the following are included within the scope of post-deployment Al maintenance EXCEPT?

A. Ensuring that all model components are subject a control framework.
B. Dedicating experts to continually monitor the model output.
C. Evaluating the need for an audit under certain standards.
D. Defining thresholds to conduct new impact assessments.

A

D. Defining thresholds to conduct new impact assessments.

Post-deployment AI maintenance typically includes ensuring that all model components are subject to a control framework, dedicating experts to continually monitor the model output, and evaluating the need for audits under certain standards. However, defining thresholds to conduct new impact assessments is usually part of the initial deployment and ongoing governance processes rather than a maintenance activity. Maintenance focuses more on the operational aspects of the AI system rather than setting new thresholds for impact assessments.

(Nota: o chatgpt escolheu a solução A inicialmente)

41
Q

In the machine learning context, feature engineering is the process of?

A. Converting raw data into clean data.
B. Creating learning schema for a model apply.
C. Developing guidelines to train and test a model.
D. Extracting attributes and variables from raw data.

A

D. Extracting attributes and variables from raw data.

Explanation:
D. Extracting attributes and variables from raw data accurately describes feature engineering. In the context of machine learning, feature engineering involves transforming raw data into meaningful features (attributes or variables) that can be used by machine learning models to improve their accuracy and performance. This process often includes selecting, creating, or modifying variables that best capture the patterns in the data.
Why Not the Other Options?
A. Converting raw data into clean data refers more to data cleaning or preprocessing, which involves handling missing values, removing duplicates, and correcting inconsistencies in the data.

B. Creating learning schema for a model apply does not specifically define feature engineering. This would be more related to defining the structure or configuration of a machine learning model rather than the extraction of features.

C. Developing guidelines to train and test a model is more about model training protocols and best practices, not about creating or selecting features from data.

42
Q

Which of the following best defines an “AI model”?

A. A system of controls that is used to govern an AI algorithm.
B. A program that has been trained on a set of data to find patterns within the data.
C. A system that applies defines rules to execute tasks.
D. A corpus of data which an AI algorithm analyzes to make predictions.

A

B. A program that has been trained on a set of data to find patterns within the data.

Justification:
Definition of an AI Model:

An AI model is the result of training an algorithm on a dataset. Through this process, the model learns to recognize patterns and relationships in the data, enabling it to make predictions, classifications, or decisions based on new inputs.
Role of Training Data:

The training phase involves feeding the model a dataset, which it uses to generalize and identify patterns. Once trained, the model can apply these learned patterns to new data.
Distinction from Algorithms and Data:

An AI model is distinct from the algorithm (the method or process used to learn) and the data (the raw information it is trained on). The model is the practical implementation of the learned relationships.
Why not the other options?
A. A system of controls that is used to govern an AI algorithm:

This describes governance mechanisms for AI, not the model itself.
C. A system that applies defined rules to execute tasks:

This describes rule-based systems, not AI models. AI models typically operate on learned patterns, not pre-defined rules.
D. A corpus of data which an AI algorithm analyzes to make predictions:

The corpus of data refers to the training dataset, not the model. The model is the result of the algorithm analyzing the data.

43
Q

According to the Singapore Model AI Governance Framework, all of the following are recommended measures to promote the responsible use of AI EXCEPT?

A. Employing human-over-the-loop protocols for high-risk systems.
B. Determining the level of human involvement in algorithmic decision-making.
C. Adapting the existing governance structure algorithmic decision-making.
D. Establishing communications and collaboration among stakeholders.

A

A. Employing human-over-the-loop protocols for high-risk systems.

Explanation:
The Singapore Model AI Governance Framework provides guidance on the responsible use of AI, emphasizing the importance of human involvement in AI-augmented decision-making. It outlines three models of human involvement:

Human-in-the-loop (HITL): Humans are directly involved in decision-making, with AI systems providing recommendations or analyses.

Human-on-the-loop (HOTL): Humans monitor the AI system’s operations and can intervene if necessary.

Human-in-command (HIC): Humans have overarching control over the AI system, including the ability to design, deploy, and deactivate it.

The framework does not specifically mention “human-over-the-loop” protocols. Therefore, option A is not a recommended measure within this framework.

The other options are recommended measures in the framework:

Determining the level of human involvement in algorithmic decision-making (B): The framework advises organizations to assess and decide the appropriate level of human involvement based on the potential impact and risks associated with AI decisions.

Adapting the existing governance structure for algorithmic decision-making (C): Organizations are encouraged to modify their governance structures to effectively oversee and manage AI systems, ensuring alignment with ethical principles and regulatory requirements.

Establishing communications and collaboration among stakeholders (D): The framework highlights the importance of engaging with various stakeholders, including employees, customers, and regulators, to build trust and ensure the responsible use of AI.

44
Q

What is the primary reason the EU is considering updates to its Product Liability Directive?

A. To increase the minimum warranty level for defective goods.
B. To define new liability exemptions for defective products.
C. Address digital services and connected products.
D. Address free and open-source software.

A

C. Address digital services and connected products.

Justification:
Modernization of the Directive:

The EU’s primary objective in updating the Product Liability Directive is to ensure it remains relevant in the digital age. This involves addressing issues related to digital services and connected products, which were not fully covered in the original directive.
Adapting to Emerging Technologies:

With the rise of IoT (Internet of Things), smart devices, and software-driven products, liability laws need to reflect the complexities of these technologies and their potential defects.
Focus Areas:

While other options (e.g., warranties or open-source software) may be part of the broader discussion, the central driver is the integration of digital and connected technologies into the liability framework.

45
Q

A Canadian company is developing an AI solution to evaluate candidates in the course of job interviews.
Before offering the AI solution in the EU market, the company must take all of the following steps EXCEPT?

A. Register the AI solution in a public EU database.
B. Establish a risk and quality management system.
C. Engage a third-party auditor to perform a bias audit.
D. Draw up technical documentation and instructions for use.

A

A. Register the AI solution in a public EU database.

Justification:
EU AI Act Requirements:

The EU AI Act outlines specific requirements for deploying AI systems, especially high-risk AI like those used for recruitment. These requirements include establishing a risk management system, performing bias audits, and preparing technical documentation to ensure transparency and compliance.
No Requirement for Public Database Registration:

While the Act requires conformity assessments and documentation, it does not mandate that AI systems be registered in a public EU database at this time. Registration requirements generally apply to specific certifications or regulatory bodies but not necessarily as a public listing.
Other Options:

B. Risk and quality management and D. Technical documentation are explicit requirements under the EU AI Act.
C. Bias audits are essential to ensure fairness and compliance with anti-discrimination principles, especially for high-risk AI systems.

46
Q

Which of the following is a subcategory of AI and machine learning that uses labeled datasets to train algorithms?

A. Segmentation.
B. Generative AI.
C. Expert systems.
D. Supervised learning.

A

D. Supervised learning.

Justification:
Definition of Supervised Learning:

Supervised learning is a subcategory of machine learning that uses labeled datasets to train algorithms. The model learns from the input-output mapping provided in the labeled data to make predictions or decisions.
Why Not Other Options?:

A. Segmentation: Segmentation refers to dividing data into meaningful groups and is not specifically related to labeled datasets.
B. Generative AI: Generative AI involves creating new content (text, images, etc.) and does not necessarily require labeled datasets for training.
C. Expert systems: Expert systems rely on predefined rules and knowledge bases, not learning from labeled datasets.
Alignment with the Question:

The use of labeled datasets specifically defines supervised learning, making it the correct answer.

47
Q

The framework set forth in the White House Blueprint for an AI Bill of Rights addresses all of the following EXCEPT?

A. Human alternatives, consideration and fallback.
B. High-risk mitigation standards.
C. Safe and effective systems.
D. Data privacy.

A

B. High-risk mitigation standards.

Justification:
White House Blueprint for an AI Bill of Rights:

The framework focuses on principles such as human alternatives and fallback options, safe and effective systems, and data privacy to protect individuals from potential harm caused by AI systems.
High-Risk Mitigation Standards:

While important in general AI governance, high-risk mitigation standards are not explicitly mentioned in the AI Bill of Rights framework. This concept is more relevant to the EU AI Act or other regulatory frameworks focusing on risk classification and mitigation.
Other Options:

A. Human alternatives, consideration, and fallback: This ensures humans can intervene or provide alternatives when AI systems fail.
C. Safe and effective systems: A core principle of the framework to ensure AI does no harm.
D. Data privacy: The framework emphasizes privacy protections to safeguard individual rights.

48
Q

Which of the following disclosures is NOT required for an EU organization that developed and deployed a high-risk AI system?

A. The human oversight measures employed.
B. How an individual may contest a decision.
C. The location(s) where data is stored.
D. The fact that an AI system is being used.

A

C. The location(s) where data is stored.

Justification:
EU AI Act and Transparency Requirements:

The EU AI Act mandates that high-risk AI systems include transparency disclosures such as:
The human oversight measures employed.
How individuals can contest decisions made by the AI system.
Disclosure that an AI system is being used to ensure informed consent.
Data Storage Location:

While organizations must ensure data protection and comply with laws like GDPR, disclosing the location(s) where data is stored is not specifically required under the AI Act’s transparency obligations for high-risk AI systems. GDPR may indirectly address data localization, but it is not a mandated disclosure under the AI Act.
Other Options:

A. Human oversight measures: Required to ensure accountability and mitigate risks associated with high-risk AI.
B. Contesting decisions: A key principle of fairness, allowing individuals to challenge AI decisions.
D. Disclosure of AI use: Ensures individuals are aware that an AI system is involved in decision-making.

49
Q

Which of the following use cases would be best served by a non-AI solution?

A. A non-profit wants to develop a social media presence.
B. An e-commerce provider wants to make personalized recommendations.
C. A business analyst wants to forecast future cost overruns and underruns.
D. A customer service agency wants to automate answers to common questions.

A

A. A non-profit wants to develop a social media presence.

Justification:
Developing a Social Media Presence:

Building a social media presence primarily involves creating engaging content, interacting with followers, and managing platforms—tasks that rely on creativity and human intuition rather than AI-driven automation.
Why Not the Other Options?:

B. Personalized recommendations: AI excels at analyzing user data and making personalized recommendations, making this task best suited for an AI solution.
C. Forecasting cost overruns and underruns: Predictive analytics, a core function of AI, is highly effective in identifying trends and forecasting financial risks.
D. Automating answers to common questions: AI-driven chatbots are widely used for automating customer service tasks, making this a suitable AI application.

50
Q

If it is possible to provide a rationale for a specific output of an AI system, that system can best be described as?

A. Accountable.
B. Transparent.
C. Explainable.
D. Reliable.

A

C. Explainable.

Justification:
Definition of Explainability:

A system is considered explainable if it is possible to provide a rationale or reasoning behind its specific outputs. Explainability ensures that users or stakeholders understand why the AI system produced a particular result.
Why Not Other Options?:

A. Accountable: Accountability refers to assigning responsibility for the AI system’s actions or decisions, which is different from providing explanations for specific outputs.
B. Transparent: Transparency relates to how open and clear the system is about its design, data, or processes but does not guarantee the ability to explain individual outputs.
D. Reliable: Reliability refers to the consistency and accuracy of the system’s performance, not its ability to explain its outputs.
Alignment with the Question:

The question specifically asks about the ability to provide a rationale for a specific output, which directly corresponds to the concept of explainability.

51
Q

Which of the following deployments of generative AI best respects intellectual property rights?

A. The system produces content that is modified to closely resemble copyrighted work.
B. The system categorizes and applies filters to content based on licensing terms.
C. The system provides attribution to creators of publicly available information.
D. The system produces content that includes trademarks and copyrights.

A

B. The system categorizes and applies filters to content based on licensing terms.

Justification:
Respecting Intellectual Property:

Categorizing and applying filters based on licensing terms ensures that the AI system only uses content in ways that comply with copyright and licensing requirements. This approach respects intellectual property rights by adhering to the terms set by creators.
Why Not Other Options?:

A. Modified content resembling copyrighted work: Producing content that closely resembles copyrighted work can still infringe on intellectual property rights, even if modifications are made.
C. Attribution to creators: Providing attribution is important but does not always satisfy intellectual property laws. Licensing terms must also be adhered to, especially for commercial or restricted content.
D. Producing content with trademarks and copyrights: Including trademarks or copyrighted material without proper authorization or licensing violates intellectual property rights.
Key Principle:

Intellectual property laws require systems to honor usage rights, licensing terms, and restrictions, which is achieved through categorization and filtering.

52
Q

A company developed AI technology that can analyze text, video, images and sound to tag content, including the names of animals, humans and objects.
What type of AI is this technology classified as?

A. Deductive inference.
B. Multi-modal model.
C. Transformative AI.
D. Expert system.

A

B. Multi-modal model.

Justification:
Definition of Multi-Modal Model:

A multi-modal model is an AI system capable of processing and analyzing data from multiple modalities, such as text, video, images, and sound, to perform tasks like content tagging or classification.
Relevance to the Scenario:

The described AI system combines different types of data (text, video, images, and sound) and processes them together, which is the hallmark of a multi-modal AI system.
Why Not the Other Options?:

A. Deductive inference: Refers to reasoning based on general principles to reach specific conclusions, which is unrelated to analyzing multiple data modalities.
C. Transformative AI: Typically refers to AI with the potential to create large-scale societal change, not specific to multi-modal capabilities.
D. Expert system: Refers to AI systems based on rule-based logic or knowledge bases, typically limited to a single domain or task.
Key Characteristic:

Multi-modal models enhance the ability to make connections across different data types, which matches the described system’s capabilities.

53
Q

A US company has developed an AI system, CrimeBuster 9619, that collects information about incarcerated individuals to help parole boards predict whether someone is likely to commit another crime if released from prison.

When considering expanding to the EU market, this type of technology would?

A. Require the company to register the tool with the EU database.
B. Be subject approval by the relevant EU authority.
C. Require a detailed conformity assessment.
D. Be banned under the EU AI Act.

A

C. Require a detailed conformity assessment.

Justification:
EU AI Act Classification:

Predictive AI systems used in sensitive areas like law enforcement or criminal justice are classified as high-risk under the EU AI Act.
High-risk systems are subject to stringent requirements, including a detailed conformity assessment to ensure compliance with safety, fairness, and transparency standards.
Why Not Other Options?:

A. Require registration: Registration alone is insufficient for high-risk AI systems, as they must also undergo conformity assessments.
B. Approval by EU authority: The EU AI Act does not prescribe a blanket approval process by authorities but instead requires adherence to technical and regulatory standards via assessments.
D. Be banned: Predictive systems in criminal justice are not outright banned under the EU AI Act. They are allowed if they meet high-risk requirements.
Key Requirement:

A conformity assessment evaluates whether the system meets the Act’s criteria for fairness, transparency, and accountability before being deployed in the EU market.

54
Q

A company plans on procuring a tool from an Al provider for its employees to use for certain business purposes.
Which contractual provision would best protect the company’s intellectual property in the tool, including training and testing data?

A. The provider will give privacy notice to individuals before using their personal data to train or test the tool.
B. The provider will defend and indemnify the company against infringement claims.
C. The provider will obtain and maintain insurance to cover potential claims.
D. The provider will warrant that the tool will work as intended.

A

B. The provider will defend and indemnify the company against infringement claims.

Justification:
Protecting Intellectual Property:

The provision to defend and indemnify against infringement claims ensures that the AI provider is responsible for addressing any legal disputes related to the use of intellectual property within the tool, such as improperly sourced training or testing data.
Why Not Other Options?:

A. Privacy notice: While important for compliance, a privacy notice does not address the protection of intellectual property.
C. Insurance to cover claims: Insurance is useful for mitigating financial losses but does not specifically protect intellectual property rights.
D. Warranty for functionality: A warranty ensures the tool works as intended but does not address intellectual property concerns or legal liabilities.
Key Requirement:

Indemnification provisions are critical for protecting a company from legal risks associated with third-party intellectual property violations.

55
Q

Which of the following AI uses is best described as human-centric?

A. Pattern recognition algorithms are used to improve the accuracy of weather predictions, which benefits many industries and everyday life.
B. Autonomous robots are used to move products within a warehouse, allowing human workers to reduce physical strain and alleviate monotony.
C. Machine learning is used for demand forecasting and inventory management, ensuring that consumers can find products they want when they want them.
D. Virtual assistants are used to adapt educational content and teaching methods to individuals, offering personalized recommendations based on ability and needs.

A

D. Virtual assistants are used to adapt educational content and teaching methods to individuals, offering personalized recommendations based on ability and needs.

Justification:
Human-Centric Definition:

Human-centric AI focuses on improving the experience, outcomes, or capabilities of individuals by addressing their specific needs and personalizing interactions.
Why Option D?:

Virtual assistants that adapt educational content and methods to individual abilities and needs demonstrate a direct and personalized focus on human well-being, which aligns perfectly with the definition of human-centric AI.
Why Not Other Options?:

A. Pattern recognition for weather predictions: This benefits industries and society but lacks the direct, individualized focus characteristic of human-centric AI.
B. Autonomous robots in warehouses: While they alleviate physical strain for workers, their primary focus is on efficiency, not personalization or individual needs.
C. Demand forecasting and inventory management: These improve supply chains for consumer benefits but are business-centric rather than human-centric.

56
Q

Which risk management framework/guide/standard focuses on value-based engineering methodology?

A. ISO/IEC Guide 51 (Safety).
B. ISO 31000 Guidelines (Risk Management).
C. IEEE 7000-2021 Standard Model Process for Addressing Ethical Concerns during System Design.
D. Council of Europe Human Rights, Democracy, and the Rule of Law Assurance Framework (HUDERIA) for AI Systems.

A

C. IEEE 7000-2021 Standard Model Process for Addressing Ethical Concerns during System Design.

Justification:
Value-Based Engineering Methodology:

The IEEE 7000-2021 Standard explicitly addresses the integration of ethical values into the engineering process. It provides a framework for identifying and incorporating human and societal values into system design, making it directly aligned with value-based engineering.
Why Not the Other Options?:

A. ISO/IEC Guide 51: Focuses on safety but does not address value-based methodologies.
B. ISO 31000 Guidelines: Provides a general risk management framework but does not explicitly focus on value-based engineering or ethical considerations.
D. Council of Europe HUDERIA Framework: Focuses on ensuring AI systems comply with human rights, democracy, and rule of law principles but is not specifically tied to engineering methodologies.
Alignment with the Question:

The IEEE 7000-2021 Standard is designed to guide engineers in aligning technology development with ethical values, making it the best fit for value-based engineering methodology.

57
Q

The OECD’s Ethical AI Governance Framework is a self-regulation model that proposes to prevent societal harms by?

A. Establishing explainability criteria to responsibly source and use data to train AI systems.
B. Defining requirements specific to each industry sector and high-risk AI domain.
C. Focusing on AI technical design and post-deployment monitoring.
D. Balancing AI innovation with ethical considerations.

A

D. Balancing AI innovation with ethical considerations.

Justification:
Core Principle of the OECD Framework:

The OECD Ethical AI Governance Framework emphasizes a balance between fostering AI innovation and ensuring that AI systems align with ethical principles to prevent societal harms. This balance is central to its recommendations for responsible AI use.
Why Not the Other Options?:

A. Establishing explainability criteria: While explainability is important, the OECD framework is broader and focuses on balancing innovation and ethics rather than specific technical aspects like explainability alone.
B. Defining industry-specific requirements: The OECD framework provides general ethical guidelines rather than detailed requirements tailored to specific industries.
C. Focusing on technical design and monitoring: The framework covers more than technical aspects, advocating for a comprehensive approach that includes ethical considerations.
Alignment with the Question:

The OECD framework aims to prevent societal harms through a balanced approach that integrates innovation with ethical safeguards, making D the best answer.

58
Q

What is the 1956 Dartmouth summer research project on AI best known as?

A. A meeting focused on the impacts of the launch of the first mass-produced computer.
B. A research project on the impacts of technology on society.
C. A research project to create a test for machine intelligence.
D. A meeting focused on the founding of the AI field.

A

D. A meeting focused on the founding of the AI field.

Justification:
Historical Context:

The 1956 Dartmouth Summer Research Project is widely regarded as the birthplace of the field of Artificial Intelligence (AI). This event introduced the term “Artificial Intelligence” and set the foundation for the field.
Significance:

It brought together prominent researchers and established a framework for exploring AI concepts, marking the beginning of AI as an academic discipline.
Why Not the Other Options?:

A. Impacts of the first mass-produced computer: The meeting was not related to hardware advancements but focused on conceptualizing AI as a field.
B. Impacts of technology on society: The project was focused on the technical and conceptual development of AI, not its societal impacts.
C. A test for machine intelligence: Although discussions about intelligence were central, the meeting’s goal was much broader than creating a test.

59
Q

According to the GDPR, an individual has the right to have a human confirm or replace an automated decision unless that automated decision?

A. Is authorized with the data subject’s explicit consent.
B. Is authorized by applicable EU law and includes suitable safeguards.
C. Is deemed to solely benefit the individual and includes documented legitimate interests.
D. Is necessary for entering into or performing under a contract between the data subject and data controller.

A

A. Is authorized with the data subject’s explicit consent.

Justification:
GDPR Article 22:

Article 22 of the GDPR provides individuals the right not to be subject to decisions based solely on automated processing, including profiling, unless specific conditions are met.
Conditions Allowing Automated Decision-Making:

Automated decisions are allowed if:
Explicit consent has been obtained from the data subject.
The decision is authorized by EU law and includes safeguards.
It is necessary for the performance of a contract.
Why Not Other Options?:

B. Authorized by applicable EU law and includes safeguards: This is correct but only one of the possible exceptions.
C. Solely benefits the individual and includes documented legitimate interests: This is not a recognized condition under GDPR Article 22.
D. Necessary for entering into or performing under a contract: While this is also a valid exception, explicit consent is the most direct basis for automated decision-making.
Key Principle:

Explicit consent from the data subject provides the clearest legal basis for bypassing the prohibition on fully automated decision-making.

60
Q

According to the GDPR, what is an effective control to prevent a determination based solely on automated decision-making?

A. Provide a just-in-time notice about the automated decision-making logic.
B. Define suitable measures to safeguard personal data.
C. Provide a right to review automated decision.
D. Establish a human-in-the-loop procedure.

A

D. Establish a human-in-the-loop procedure.

Explanation:
Under the General Data Protection Regulation (GDPR), automated decision-making, including profiling, is specifically addressed in Article 22. This provision states that individuals have the right not to be subject to decisions based solely on automated processing if such decisions produce legal effects or significantly affect them.

To comply with GDPR and prevent fully automated decision-making:

Establish a human-in-the-loop procedure (Correct):
Including human involvement in the decision-making process ensures that decisions are not made solely by automated systems. This is an effective safeguard under Article 22(3) and helps mitigate risks of bias, error, or unfair treatment in automated decisions.

Provide a right to review automated decision (Partially relevant):
While GDPR grants data subjects the right to obtain human intervention, express their point of view, or contest a decision, this is a response mechanism after the decision is made, not a preventive control.

Provide a just-in-time notice about the automated decision-making logic (Insufficient):
Informing data subjects about the use of automated decision-making and its logic is required under Articles 13 and 14 (transparency obligations). However, providing notice alone does not prevent decisions based solely on automated processing.

Define suitable measures to safeguard personal data (Broad and not specific):
Implementing technical and organizational safeguards is essential under GDPR, but these measures focus on protecting personal data integrity and security. They are not directly aimed at preventing solely automated decision-making.

61
Q

According to the GDPR’s transparency principle, when an Al system processes personal data in automated decision-making, controllers are required to provide data subjects specific information on?

A. The existence of automated decision-making and meaningful information on its logic and consequences.
B. The personal data used during processing, including inferences drawn by the Al system about the data.
C. The data protection impact assessments carried out on the Al system and legal bases for processing.
D. The contact details of the data protection officer and the data protection national authority.

A

A. The existence of automated decision-making and meaningful information on its logic and consequences.

Explanation:
Under the General Data Protection Regulation (GDPR), particularly Articles 13(2)(f), 14(2)(g), and 22, the transparency principle requires data controllers to provide specific information to data subjects when their personal data is subject to automated decision-making, including profiling, that has legal or similarly significant effects. The required information includes:

The existence of automated decision-making, including profiling:
Data subjects must be informed if their personal data is being processed using automated systems to make decisions.

Meaningful information about the logic involved:
This means providing an understandable explanation of how the decision-making process works, without requiring technical details that might confuse the subject.

The significance and potential consequences of the processing:
Data subjects must understand how the decision-making might impact them, such as in terms of legal rights, opportunities, or access to services.

Why the Other Options Are Incorrect:
B. The personal data used during processing, including inferences drawn by the AI system about the data:
While data subjects have the right to access their personal data under Articles 15(1)(h) and 22, this is part of broader rights like data access, not specifically required under the transparency principle for automated decision-making.

C. The data protection impact assessments carried out on the AI system and legal bases for processing:
Data protection impact assessments (DPIAs) are a tool for internal compliance and risk assessment, not information that must be provided to data subjects. The legal basis for processing must be disclosed, but this is not uniquely tied to automated decision-making.

D. The contact details of the data protection officer and the data protection national authority:
While controllers must provide this information under Articles 13 and 14, it is general information required for all data processing, not specific to automated decision-making.

62
Q

A company is creating a mobile app to enable individuals to upload images and videos, and analyze this data using ML to provide lifestyle improvement recommendations. The signup form has the following data fields:
1.First name
2.Last name
3.Mobile number
4.Email ID
5.New password
6.Date of birth
7.Gender
In addition, the app obtains a device’s IP address and location information while in use.

What GDPR privacy principles does this violate?

A. Purpose Limitation and Data Minimization.
B. Accountability and Lawfulness.
C. Transparency and Accuracy.
D. Integrity and Confidentiality.

A

A. Purpose Limitation and Data Minimization.

Explanation:
The GDPR (General Data Protection Regulation) establishes several privacy principles that organizations must adhere to when processing personal data. Two of these principles are particularly relevant in this scenario:

Purpose Limitation: This principle requires that personal data be collected only for specified, explicit, and legitimate purposes and not further processed in a way that is incompatible with those purposes. The company must clearly define why each piece of personal data is being collected (e.g., why the app needs date of birth and gender for lifestyle recommendations).

Data Minimization: This principle mandates that the data collected should be adequate, relevant, and limited to what is necessary in relation to the purposes for which it is processed. If the app collects data that is not strictly needed for providing lifestyle recommendations or delivering core functionalities, it may violate this principle. For example, if the app can function without collecting a mobile number or precise location information, then collecting this data might be considered excessive.

63
Q

A company has trained an ML model primarily using synthetic data, and now intends to use live personal data to test the model.
Which of the following is NOT a best practice apply during the testing?

A. The test data should be representative of the expected operational data.
B. Testing should minimize human involvement to the extent practicable.
C. The test data should be anonymized to the extent practicable.
D. Testing should be performed specific to the intended uses.

A

B. Testing should minimize human involvement to the extent practicable.

Justification:
Testing ML Models with Live Personal Data:

Testing an ML model with live personal data requires adherence to best practices, such as ensuring the data reflects operational conditions, protecting personal information, and aligning tests with intended use cases.
Why Minimizing Human Involvement is NOT Best Practice:

While automation is useful, testing with live personal data should involve human oversight to ensure compliance with data protection regulations, ethical considerations, and the model’s proper functioning.
Human involvement is critical during testing to validate the results, monitor unexpected behavior, and address potential biases or issues in the model.
Why the Other Options Are Best Practices:

A. Representative of operational data: Ensures the model performs reliably in real-world scenarios.
C. Anonymized data: Protects the privacy of individuals during testing, aligning with data protection laws like GDPR.
D. Specific to intended uses: Ensures that testing aligns with the model’s deployment objectives, minimizing risks in real-world applications.

64
Q

You are an engineer that developed an Al-based ad recommendation tool.
Which of the following should be monitored to evaluate the tool’s effectiveness?

A. Output data, assess the delta between the prediction and actual ad clicks.
B. Algorithmic patterns, to show the model has a high degree of accuracy.
C. Input data, to ensure the ads are reaching the target audience.
D. GPU performance, to evaluate the tool’s robustness.

A

A. Output data, assess the delta between the prediction and actual ad clicks.

Justification:
Effectiveness of an Ad Recommendation Tool:

The effectiveness of an AI-based ad recommendation tool is primarily evaluated based on how well its predictions (e.g., ad clicks) match the actual user behavior.
Monitoring the output data to measure the difference (delta) between predictions and actual ad clicks directly reflects the tool’s performance.
Why Not the Other Options?:

B. Algorithmic patterns: While accuracy is important, monitoring algorithmic patterns alone does not directly measure the tool’s effectiveness in achieving its goal (ad engagement).
C. Input data: Ensuring the quality of input data is necessary for model training but does not evaluate the tool’s effectiveness in production.
D. GPU performance: GPU performance relates to computational efficiency, not the tool’s ability to deliver accurate ad recommendations.
Key Metric:

The critical metric for evaluating an ad recommendation tool is how well it predicts user engagement, making A the most relevant choice.

65
Q

An artist has been using an AI tool to create digital art and would like to ensure that it has copyright protection in the United States.
Which of the following is most likely to enable the artist to receive copyright protection?

A. Ensure the tool was trained using publicly available content.
B. Obtain a representation from the AI provider on how the tool works.
C. Provide a log of the prompts the artist used to generate the images.
D. Update the images in a creative way to demonstrate that it is the artist’s.

A

D. Update the images in a creative way to demonstrate that it is the artist’s.

Justification:
Copyright Protection in the United States:

Under U.S. copyright law, for works to be eligible for protection, they must exhibit a modicum of creativity and be created by a human. AI-generated works may not qualify unless the human creator significantly contributes creative elements.
Why Option D?:

By updating the images in a creative way, the artist ensures that their personal input adds originality, making the work eligible for copyright protection under U.S. law. This demonstrates human authorship.
Why Not the Other Options?:

A. Ensure the tool was trained using publicly available content:
The training data’s source is irrelevant to the artist’s copyright claim.
B. Obtain a representation from the AI provider on how the tool works:
Knowing how the AI tool works does not establish human authorship or originality.
C. Provide a log of the prompts the artist used to generate the images:
While prompts are part of the creation process, they may not sufficiently demonstrate human creativity.
Key Requirement:

U.S. copyright law prioritizes human creativity, so the artist must show their contribution to the work beyond what the AI generates.

66
Q

Which of the following is the least relevant consideration in assessing whether users should be given the right to opt out from an Al system?
A. Feasibility.
B. Risk to users.
C. Industry practice.
D. Cost of alternative mechanisms.

A

D. Cost of alternative mechanisms.

Revised Justification:
The explanation highlights that the primary considerations in deciding whether users should have the right to opt out focus on:

Feasibility – Whether it is practical to implement an opt-out mechanism.
Risk to users – The potential harm or benefits users might face without the option.
Industry practice – The norms and standards within the industry, which may influence decisions but are tied to ethical AI practices.
However, the cost of alternative mechanisms is noted as less relevant to the ethical question of whether users should be given the right to opt out. While cost matters in the broader implementation context, it does not directly affect the ethical principle of protecting user rights and ensuring ethical practices.

67
Q

You are the chief privacy officer of a medical research company that would like to collect and use sensitive data about cancer patients, such as their names, addresses, race and ethnic origin, medical histories, insurance claims, pharmaceutical prescriptions, eating and drinking habits and physical activity.
The company will use this sensitive data to build an Al algorithm that will spot common attributes that will help predict if seemingly healthy people are more likely to get cancer. However, the company is unable to obtain consent from enough patients to sufficiently collect the minimum data to train its model.

Which of the following solutions would most efficiently balance privacy concerns with the lack of available data during the testing phase?
A. Utilize synthetic data to offset the lack of patient data.
B. Deploy the current model and recalibrate it over time with more data.
C. Extend the model to multi-modal ingestion with text and images.
D. Refocus the algorithm to patients without cancer.

A

Correct Answer: A. Utilize synthetic data to offset the lack of patient data.

Justification:
Balancing Privacy Concerns and Data Needs:

Synthetic data provides a solution that mimics the statistical properties of real data without including personally identifiable information. This allows for the protection of sensitive data while addressing the lack of available data for model training.
Why Synthetic Data?:

Synthetic data reduces privacy risks since it is not linked to real individuals.
It enables testing and refining the AI algorithm efficiently while avoiding potential ethical and legal violations related to the use of insufficiently consented data.
Why Not the Other Options?:

B. Deploy the current model and recalibrate it over time with more data:
Deploying an insufficiently trained model may lead to unreliable results, harming patients and undermining trust in the system.
C. Extend the model to multi-modal ingestion with text and images:
While multi-modal models are valuable for improving insights, they do not address the fundamental issue of a lack of sufficient training data.
D. Refocus the algorithm to patients without cancer:
Changing the focus of the algorithm does not solve the core problem of training data availability and privacy concerns. It may also deviate from the company’s primary goal of identifying cancer predictors.
Efficient and Ethical Solution:

Using synthetic data aligns with privacy principles and ensures that the AI algorithm can be tested and refined without compromising sensitive patient information.

68
Q

What type of organizational risk is associated with Al’s resource-intensive computing demands?
A. Third-party risk.
B. Environmental risk.
C. People risk.
D. Security risk.

A

Correct Answer: B. Environmental risk.

Justification:
AI’s Resource-Intensive Computing Demands:

AI systems, especially those involving large-scale machine learning models, require significant computational power. This leads to high energy consumption, which has an environmental impact, such as increased carbon emissions from data centers.
Why Environmental Risk?:

The resource-intensive nature of AI contributes to energy use and environmental degradation, which falls under the category of environmental risks. Organizations must consider the carbon footprint of their AI infrastructure and the sustainability of their operations.
Why Not the Other Options?:

A. Third-party risk:
This refers to risks associated with external vendors or partners, such as cloud providers, but does not specifically relate to computing demands.
C. People risk:
People risk involves issues related to employees, such as skill gaps or resistance to change, and is not directly tied to computing resources.
D. Security risk:
Security risk involves threats like data breaches or cyberattacks, which are separate concerns from the environmental impact of resource-intensive computing.
Key Consideration:

Organizations need to address environmental risks by optimizing energy efficiency, using renewable energy sources, or implementing carbon-offset strategies.

69
Q

A U.S. mortgage company developed an Al platform that was trained using anonymized details from mortgage applications, including the applicant’s education, employment and demographic information, as well as from subsequent payment or default information. The Al
platform will be used automatically grant or deny new mortgage applications, depending on whether the platform views an applicant as presenting a likely risk of default.

Which of the following laws is NOT relevant to this use case?
A. Fair Housing Act.
B. Fair Credit Reporting Act.
C. Equal Credit Opportunity Act.
D. Title VII of the Civil Rights Act of 1964.

A

D. Title VII of the Civil Rights Act of 1964.

Explanation:
Here’s why Title VII of the Civil Rights Act of 1964 is not relevant, while the other laws are:

Fair Housing Act (FHA):
The FHA is relevant because it prohibits discrimination in housing-related transactions (like mortgage applications) based on protected characteristics, such as race, color, national origin, religion, sex, familial status, or disability. The use of demographic information in the AI platform could raise concerns about discriminatory impacts under this act.

Fair Credit Reporting Act (FCRA):
The FCRA governs how consumer credit information can be used and ensures accuracy, fairness, and privacy. Since the AI platform uses payment and default history—potentially sourced from credit data—this act would apply.

Equal Credit Opportunity Act (ECOA):
The ECOA prohibits discrimination in credit applications based on factors such as race, color, religion, national origin, sex, marital status, or age. The AI platform’s reliance on applicant demographic information could lead to discriminatory lending practices, making this law applicable.

Title VII of the Civil Rights Act of 1964:
Title VII specifically addresses discrimination in employment, not housing or credit transactions. Therefore, it is not relevant to this use case, which involves mortgage applications.

70
Q

The most important factor in ensuring fairness when training an Al system is?

A. The architecture and model selection.
B. The data labeling and classification.
C. The data attributes and variability.
D. The model accuracy and scale.

A

C. The data attributes and variability.

Explanation:
Fairness in AI systems primarily depends on the quality and diversity of the data used for training. Here’s why each option is relevant but why C is the most critical:

The architecture and model selection (A):
While the choice of model architecture can influence how well an AI system performs or generalizes, it does not inherently address fairness. Fairness issues typically arise from biases in the training data or how the data is processed, not from the model architecture itself.

The data labeling and classification (B):
Accurate and unbiased data labeling is crucial for a well-performing model, but fairness depends more broadly on the representativeness and variability of the data. Even correctly labeled data can lead to unfair outcomes if it lacks diversity or reflects societal biases.

The data attributes and variability (C):
This is the most important factor for fairness. Ensuring the dataset includes a diverse range of attributes (e.g., race, gender, age, socioeconomic status) and sufficient variability prevents the model from disproportionately favoring or disadvantaging certain groups. Bias in data attributes or a lack of variability often leads to unfair outcomes in AI systems.

The model accuracy and scale (D):
While high accuracy and scalability are important for model performance, they do not inherently guarantee fairness. A model can be highly accurate yet biased if the training data is not fair or representative.

71
Q

During the development of semi-autonomous vehicles, various failures occurred as a result of the sensors misinterpreting environmental surroundings, such as sunlight.

These failures are an example of?
A. Hallucination.
B. Brittleness.
C. Uncertainty.
D. Forgetting.

A

B. Brittleness.

Explanation:
In the context of AI and machine learning systems, brittleness refers to a system’s inability to handle unexpected or edge-case scenarios effectively. Brittleness often occurs when the system has been trained or designed for specific, controlled conditions and struggles to generalize or adapt to new, variable, or challenging environments.

Here’s why the other options are incorrect:

Hallucination (A):
Hallucination in AI typically refers to generating or interpreting information that does not exist, such as an AI model producing fabricated data or incorrect outputs. This is more common in generative AI systems, like language or image models, not sensor failures in vehicles.

Brittleness (B):
This is the correct answer. The sensor failures in semi-autonomous vehicles due to environmental challenges (e.g., sunlight) highlight the system’s lack of robustness and inability to handle real-world variability, which are classic symptoms of brittleness.

Uncertainty (C):
Uncertainty refers to situations where the system lacks confidence in its predictions or decisions. While this can contribute to failures, the example described points to a failure to adapt to specific conditions rather than uncertainty about the decision.

Forgetting (D):
Forgetting, particularly in the context of AI, often relates to catastrophic forgetting, where a model loses previously learned information when trained on new data. This is not relevant to the described scenario of environmental sensor misinterpretation.

72
Q

You are a privacy program manager at a large e-commerce company that uses an Al tool to deliver personalized product recommendations based on visitors’ personal information that
has been collected from the company website, the chatbot and public data the company has scraped from social media.
A user submits a data access request under an applicable U.S. state privacy law, specifically seeking a copy of their personal data, including information used to create their profile for
product recommendations.

What is the most challenging aspect of managing this request?
A. Some of the visitor’s data is synthetic data that the company does not have to provide to the data subject.
B. The data subject’s data is structured data that can be searched, compiled and reviewed only by an automated tool.
C. The data subject is not entitled to receive a copy of their data because some of it was scraped from public sources.
D. Some of the data subject’s data is unstructured data and you cannot untangle it from the other data, including information about other individuals.

A

D. Some of the data subject’s data is unstructured data and you cannot untangle it from the other data, including information about other individuals.

Explanation:
Managing a data access request in this scenario involves significant challenges, particularly when dealing with unstructured data. Here’s why each option is relevant but D is the most challenging:

Some of the visitor’s data is synthetic data that the company does not have to provide to the data subject (A):
Synthetic data is not real personal data but rather artificially generated data based on patterns or characteristics of actual data. While it may not need to be disclosed under privacy laws, the challenge is minor compared to handling unstructured data intertwined with others’ information.

The data subject’s data is structured data that can be searched, compiled and reviewed only by an automated tool (B):
Structured data is relatively easier to manage since it is organized and searchable. While automation might be necessary to process large amounts of data, it does not pose as significant a challenge as dealing with unstructured or intertwined data.

The data subject is not entitled to receive a copy of their data because some of it was scraped from public sources (C):
This is incorrect. U.S. privacy laws often grant individuals rights to access their personal data, regardless of whether it was obtained from public sources. Scraped data would still qualify as personal data if it is linked to the individual.

Some of the data subject’s data is unstructured data and you cannot untangle it from the other data, including information about other individuals (D):
This is the most challenging aspect. Unstructured data, such as free text from chat logs or social media content, is harder to process and separate. Additionally, privacy laws may require you to ensure that other individuals’ data is not inadvertently disclosed when fulfilling the request. This requires significant effort to disentangle and anonymize data, making it a substantial challenge.

73
Q

All of the following types of testing can help evaluate the performance of a responsible Al system EXCEPT?

A. Risk probability/severity.
B. Adversarial robustness.
C. Statistical sampling.
D. Decision analysis.

A

A. Risk probability/severity.

Risk probability/severity testing is not typically used to evaluate the performance of an AI system. While important for risk management, it does not directly assess an AI system’s operational performance. Adversarial robustness, statistical sampling, and decision analysis
are all methods that can help evaluate the performance of a responsible AI system by testing its resilience, accuracy, and decision-making processes under various conditions. Reference:
AIGP Body of Knowledge on AI Performance Evaluation and Testing.

The question asks which type of testing does not help evaluate the performance of a responsible AI system.

Reassessing Each Option:
Risk probability/severity (A):
This type of testing focuses on assessing the likelihood and impact of risks associated with the AI system, including potential harm to users or failures in operation. While it is critical for risk management and ensuring responsible AI, it is less about evaluating performance metrics like accuracy, robustness, or fairness.

Adversarial robustness (B):
This directly measures the system’s ability to handle adversarial attacks or edge cases, which is a key component of performance evaluation for responsible AI. It ensures the system operates correctly under challenging conditions.

Statistical sampling (C):
Statistical sampling is a method to evaluate how well the system performs across diverse subsets of data, often used to test fairness, accuracy, and representativeness. This is a direct measure of performance.

Decision analysis (D):
Decision analysis is a strategic tool used to assess decision-making under uncertainty. It is not a direct method for testing the technical performance of an AI system, making it less relevant in this context.

A. Risk probability/severity is less about performance testing and more about broader risk management in AI systems. While still critical for responsible AI, it does not directly evaluate the performance of the system itself.

The most accurate answer is therefore A. Risk probability/severity.

74
Q

Under the Canadian Artificial Intelligence and Data Act, when must the Minister of Innovation, Science and Industry be notified about a high-impact Al system?

A. When use of the system causes or is likely to cause material harm.
B. When the algorithmic impact assessment has been completed.
C. Upon release of a new version of the system.
D. Upon initial deployment of the system.

A

ChatGPT: A. When use of the system causes or is likely to cause material harm.
(NOTA: o resumo diz que é a resposta D. Upon initial deployment of the system, de acordo com o AIGP Body of Knowledge, domain on AI laws and standards).

Under the proposed Canadian Artificial Intelligence and Data Act (AIDA), entities responsible for high-impact AI systems are required to notify the Minister of Innovation, Science, and Industry in specific circumstances. According to the Act, notification is mandated when the use of a high-impact AI system results in, or is likely to result in, material harm. This requirement is outlined in Section 12 of the Act, which states that if a high-impact system causes or is likely to cause material harm, the responsible person must notify the Minister as soon as feasible.

75
Q

What is the key feature of Graphical Processing Units (GPUs) that makes them well-suited to running Al applications?

A. GPUs run many tasks concurrently, resulting in faster processing.
B. GPUs can access memory quickly, resulting in lower latency than CPUs.
C. GPUs can run every task on a computer, making them more robust than CPUs.
D. The number of transistors on GPUs doubles every two years, making the chips smaller and lighter.

A

A. GPUs run many tasks concurrently, resulting in faster processing.

Explanation:
The key feature of Graphical Processing Units (GPUs) that makes them ideal for AI applications lies in their ability to process multiple tasks in parallel. This is critical for AI workloads, especially for training and inference in machine learning models, which often require performing a massive number of matrix operations and computations simultaneously.

Why Each Option Matters:
GPUs run many tasks concurrently, resulting in faster processing (A):
GPUs are designed with a high number of cores optimized for parallel processing, enabling them to handle thousands of operations simultaneously. This makes them particularly well-suited for AI tasks like deep learning, where large amounts of data need to be processed in parallel.

GPUs can access memory quickly, resulting in lower latency than CPUs (B):
While GPUs often have high memory bandwidth, this is not their defining feature for AI applications. CPUs generally have lower latency for accessing memory but lack the parallel processing capabilities of GPUs.

GPUs can run every task on a computer, making them more robust than CPUs (C):
GPUs are specialized processors optimized for parallel tasks and cannot run every task efficiently. CPUs remain better suited for general-purpose computing and sequential task execution.

The number of transistors on GPUs doubles every two years, making the chips smaller and lighter (D):
This statement references Moore’s Law, which applies to all integrated circuits, not specifically GPUs. Additionally, while miniaturization is a factor in improved performance, it is not the defining feature that makes GPUs ideal for AI.

76
Q

An Al system that maintains its level of performance within defined acceptable limits despite real world or adversarial conditions would be described as?

A. Robust.
B. Reliable.
C. Resilient.
D. Reinforced.

A

Answer: C. Resilient.
Explanation:
An AI system that maintains its level of performance within defined acceptable limits despite real-world or adversarial conditions is described as resilient. Resilience in AI refers to the system’s ability to withstand and recover from unexpected challenges, such as cyber-attacks, hardware failures, or unusual input data. This characteristic ensures that the AI system can continue to function effectively and reliably in various conditions, maintaining performance and integrity. Robustness, on the other hand, focuses on the system’s strength against errors, while reliability ensures consistent performance over time. Resilience combines these aspects with the capacity to adapt and recover.

Analysis of Terms:
Resilient (C):
Resilience combines robustness (withstanding challenges) and recovery/adaptation after disruptions. In this explanation, resilience is positioned as the overarching characteristic that ensures the system maintains functionality and integrity under various conditions.

Robust (A):
While robustness emphasizes resistance to disruptions, it does not explicitly include the ability to recover or adapt, which is part of resilience.

Reliable (B):
Reliability is about consistent performance over time but does not specifically address how the system handles adversarial or unexpected conditions.

Reinforced (D):
This term refers to reinforcement learning techniques and is not applicable here.

77
Q

Which of the following elements of feature engineering is most important to mitigate the potential bias in an Al system?

A. Feature selection.
B. Feature validation.
C. Feature transformation.
D. Feature importance analysis

A

A. Feature selection.

Explanation:
Feature selection is the most critical element of feature engineering for mitigating bias in an AI system. This process involves choosing which input features (attributes or variables) should be used by the machine learning model. Carefully selecting features helps reduce the risk of introducing bias and ensures that the model is trained on data that is representative and relevant to the problem at hand.

Analysis of Each Option:
Feature selection (A):

Selecting the appropriate features is crucial to avoid including variables that may reflect societal biases or irrelevant patterns. For example, using race or gender as a feature in a hiring algorithm could lead to discriminatory outcomes.
Proper feature selection ensures that only fair, unbiased, and necessary variables are used, mitigating potential bias at its source.
Feature validation (B):

Feature validation ensures that the chosen features meet quality standards, such as being complete and correctly formatted. While this is important for ensuring data accuracy, it does not directly address bias in the features.
Feature transformation (C):

Feature transformation involves converting data into a format that is more suitable for machine learning (e.g., normalizing values or creating categorical encodings). While transformation can impact how features are interpreted, it does not inherently mitigate bias.
Feature importance analysis (D):

Analyzing feature importance helps identify which features the model considers most influential in making predictions. While this can reveal biases after training, it is a diagnostic tool rather than a proactive mitigation strategy.

78
Q

Retraining an LLM can be necessary for all of the following reasons EXCEPT?

A. To minimize degradation in prediction accuracy due to changes in data.
B. Adjust the model’s hyper parameters specific use case.
C. Account for new interpretations of the same data.
D. To ensure interpretability of the model’s predictions.

A

D. To ensure interpretability of the model’s predictions.

Explanation:
Retraining a large language model (LLM) is typically necessary for several practical reasons:

A. To minimize degradation in prediction accuracy due to changes in data:

Over time, data distributions can change (data drift), and retraining ensures the model adapts to new patterns or trends to maintain accuracy.
B. Adjust the model’s hyperparameters for a specific use case:

While hyperparameters themselves are not directly part of retraining, optimization and fine-tuning for a use case might necessitate retraining on adjusted configurations or data.
C. Account for new interpretations of the same data:

Retraining allows the model to learn updated patterns, nuances, or labeling changes in the data.
D. To ensure interpretability of the model’s predictions:

Interpretability relates to explaining or understanding the model’s decisions, which is more about choosing interpretable algorithms, creating tools for explanation, or post-hoc analysis rather than retraining. Retraining does not inherently improve or ensure interpretability.