Questions Flashcards
1
Q
Fiona leads the data science team at a media company. She is developing a new machine learning model to
improve its recommender systems. She has a large volume of input data around user preferences, but not
all users actively confirmed their likes and dislikes. What would be the most suitable ML approach for Fiona
to use?
A. Supervised learning.
B. Unsupervised learning.
C. Reinforcement learning.
D. Semi-supervised learning.
A
- The correct answer is D. The use of semi-supervised learning is the most suitable approach in this situation
given the combination of labelled and unlabeled data in the full dataset.
Body of Knowledge Domain I, Competency B
2
Q
- Which of the following is a method that helps to ensure AI integrity and that data is representative,
accurate and unbiased?
A. Tracking data lineage.
B. Assuming the data is accurate.
C. Purchasing data from a vendor.
D. Reversioning the data internally.
A
- The correct answer is A. Data lineage tracks data over time from the source to any other intervening
programs or uses, and ultimately to the AI program or process utilizing it. Knowing where and how data has
been used and manipulated before it is incorporated into an AI program or process helps ensure the data
being used is accurate and appropriate.
Body of Knowledge Domain VI, Competency D
3
Q
- You are tasked with designing an AI system to predict customer attrition for a telecommunications
company. During the design phase, you are focusing on developing the data strategy. Which of the
following statements best describes the critical importance of data gathering in this phase?
A. It primarily aims to reduce data storage costs, prioritizing low-cost repositories over data quality or
relevance.
B. It ensures the AI system has access to high-quality and relevant data, which is the fuel for training
effective models.
C. It focuses on designing attractive user interfaces for data input, catering to user experience over data
completeness or accuracy.
D. It is only important for regulatory complianc
A
- The correct answer is B. Implementing a data strategy during the AI system design phase is crucial because
it ensures access to high-quality and relevant data, which is fundamental for training effective AI models.
The quality and relevance of data directly impacts the performance and accuracy of AI systems. The other
options are irrelevant aspects of an AI system’s design phase.
Body of Knowledge Domain V, Competency B
4
Q
- Which of the following is included in the Canadian Artificial Intelligence and Data Act’’s risk-based approach?
A. Source of the artificial intelligence technology.
B. Elimination of data privacy impact assessments.
C. Consideration of the nature and severity of harms.
D. Equal application to companies operating in and outside of Canada.
A
- The correct answer is C. Bill C-27 bases the risk evaluation on:
a. Nature and severity of harms.
b. Scale of use of the system.
c. Extent to which individuals can opt out or control interactions.
d. Imbalance of economic and social status of the individual interacting with the system.
Body of Knowledge Domain IV, Competency B
5
Q
- In the context of workforce readiness for AI, what is an essential consideration for businesses to fully
leverage AI benefits responsibly?
A. Focusing on training a select IT task force in advanced AI techniques.
B. Avoiding employee training on AI to prevent over-reliance on technology.
C. Implementing comprehensive training programs across various departments.
D. Outsourcing AI-related tasks to third-party specialists to reduce training needs.
A
- The correct answer is C. To maximize the benefits of AI, businesses should implement comprehensive
training programs that encompass not just technical AI skills but also ethical considerations of AI usage. This
approach ensures a broad-based understanding of AI across the organization, enabling more effective and
responsible use of AI technologies.
Body of Knowledge Domain VII, Competency B
6
Q
- The Human Rights, Democracy and Rule of Law Impact Assessment builds on a history of impact
assessments used extensively in other domains; such as, for example, data protection regulation. What is
one of the main objectives of this type of impact assessment?
A. To produce a document with mandatory requirements and enforcement mechanisms for harmful
design choices.
B. To provide a process for assessing and grading the likelihood of risks associated with potentially
harmful outcomes.
C. To expand on the proposals and guidelines for specific sectors and AI applications decided solely by the
Council of Europe.
D. To voluntarily endorse business practices, technology or policies after they have been deployed or
when most convenient to developers.
A
- The correct answer is B. The Human Rights, Democracy and the Rule of Law Impact Assessment provides an
opportunity for project teams and engaged stakeholders to come together to produce detailed evaluations
of the potential and actual harmful impacts that the design, development and use of an AI system could
have on human rights, fundamental freedoms, elements of democracy and the rule of law.
Body of Knowledge Domain IV, Competency C
7
Q
- The manager of a call center decides to use an AI system to record and conduct sentiment analysis on
interactions between customers and call center agents. Prior to using this data to evaluate the
performance of the agents, what step should be implemented to minimize error in the evaluations?
A. A security review of the AI system.
B. A manual review of the recordings.
C. An estimate of call rates to the center.
D. A retention policy for aggregate reports.
A
- The correct answer is B. Human review or supervision of the recorded call should be done to ensure the
sentiment analysis of the AI system is accurate prior to using the data to evaluate the agent’s performance.
Retention of data is important, but aggregate report data will not create risk for an individual. An estimate of
call rates may help in staffing or workforce planning but won’t directly affect an individual’s performance
rating. A security review of the AI will ensure security of the system, but also does not directly affect the
individual or how their performance is evaluated.
Body of Knowledge Domain III, Competency B
8
Q
- Consider you are a privacy consultant with a tech company developing an AI-driven health-monitoring
application. The application collects user data to provide personalized health insights. The company is
dedicated to ensuring user privacy and seeks your advice. In the context of developing a privacy-enhanced
AI health-monitoring application, which of the following measures would be most effective in aligning with
privacy standards?
A. Implementing user consent mechanisms and anonymizing collected data to protect user identities.
B. Utilizing social media integration to enrich user profiles with additional lifestyle information and offers.
C. Sharing raw user data with various pharmaceutical companies to facilitate targeted drug development.
D. Storing user health data on a publicly accessible server for easy collaboration with health care
researchers.
A
- The correct answer is A. Obtaining consent from users empowers individuals to make informed decisions
about the use of their health data. Furthermore, anonymizing involves removing personally identifiable
information from the collected data, such as names, addresses and contact details.
Body of Knowledge Domain II, Competency B
9
Q
- Which of the following is an important consideration for mitigating AI system errors?
A. Excluding a review of previous incidents.
B. Focusing exclusively on legal compliance.
C. Adopting a uniform approach for all errors.
D. Understanding the AI use case and the data type.
A
- The correct answer is D. The AI use case and the data type involved in system errors will guide the method
of mitigation needed. This allows for an appropriate response and helps to guide any updates or changes
needed to avoid similar issues in the future.
Body of Knowledge Domain VI, Competency E
10
Q
- Large language models are considered massive neural networks that can generate human-like text (e.g.,
emails, poetry, stories, etc.). These models also have the ability to predict the text that is likely to come
next. How is machine learning related to LLMs?
A. LLMs do not use ML but rely on a fixed answer database.
B. ML and LLMs are unrelated; LLMs use manual programming, not ML.
C. LLMs learn from text data and use ML algorithms to generate responses.
D. ML and LLMs are separate; ML is for numerical analysis, and LLMs are for graphics.
A
- The correct answer is C. Large language models cannot be designed without the training of the datasets
required for the model to learn, make decisions and predict. LLMs are built using a machine learning
algorithm trained to make classifications or predictions.
Body of Knowledge Domain I, Competency A
11
Q
- In the design phase of AI development, privacy-enhancing technologies play a critical role in addressing
data privacy concerns. Among these technologies, differential privacy and federated learning are
significant. Given this context, which of the following best illustrates the application of differential privacy in
the AI design phase?
A. An AI model is trained in a centralized dataset in which individual data points are anonymized before
being added to the dataset.
B. An AI system enables multiple models to learn from separate datasets, then combines them to improve
the overall model without sharing individual data points.
C. An AI application uses encryption to secure data in transit between the client and server, ensuring that
intercepted data cannot be read by unauthorized parties.
D. An AI system adds random noise to the raw data it collects, ensuring that any output reflects general
trends in the dataset but does not compromise individual data privacy.
A
- The correct answer is D. This option directly aligns with the principles of differential privacy. Differential
privacy is a technique used to ensure the privacy of individual data points in a dataset by adding statistical
noise. This method allows the dataset to be used for analysis or AI model training while maintaining the
confidentiality of individual data entries. The key aspect of differential privacy is that it provides insights into
the general patterns or trends of the data without revealing sensitive information about any specific
individual.
Body of Knowledge Domain V, Competency B
12
Q
- When implementing responsible AI governance and risk management, what is a critical consideration for
testing and evaluation in alignment with a specific use case?
A. Include potentially malicious data in the test.
B. There is no need for repeatability assessments for novel cases.
C. Use the same testing and evaluation approach for all algorithms.
D. Exclude cases that the AI has not encountered during its training.
A
- The correct answer is A. Incorporating potentially malicious data into the testing process would enable the
assessment of the AI’s robustness and resilience against adversarial inputs and ensure that it behaves safely
in real-world scenarios. The other options are not aligned with the best practices for responsible AI testing.
Body of Knowledge Domain VI, Competency E
13
Q
- The Digital Services Act aims to regulate the digital marketplace in the EU. Which of the following best
describes how the DSA will empower users?
A. Online marketplaces will have to comply with special obligations to combat the sale of illegal products
and services.
B. Very large online search engines will be held accountable for their role in disseminating illegal and
harmful content.
C. Very large online platforms will have to be more transparent in how they select the companies allowed
to advertise on their platform.
D. Platforms will have to provide clear information on why content is shown and give users the right to opt
out of profiling-based recommendations.
A
- The correct answer is D. Platforms will be required to disclose to users why the user is being shown specific
content, as well as provide users with the right to opt out of being shown content that was derived through
profiling. This is the only option that provides users with control over their content.
Body of Knowledge Domain III, Competency A
14
Q
- When assessing the success of an AI system meeting its objectives, which of the following approaches best
aligns with the requirement to ensure a comprehensive evaluation while avoiding automation bias?
A. Evaluate and align predefined benchmarks that will provide evidence of having achieved your system
goals.
B. Rely on the AI system’s output to determine if the goals were achieved, as this will provide a reliable
and objective measure.
C. Conduct a review that includes the AI system’s output, human interpretation, and potential secondary
or unintended outputs.
D. Focus on user feedback about the AI system’s performance to directly measure the system’s
effectiveness in meeting the objectives
A
- The correct answer is C. This is a comprehensive approach as it includes different testing initiatives. By
integrating the AI system’s output, human interpretation and potential secondary or unintended outputs,
you will avoid automation bias (when only relying on the output), as well as the limitation of having only
human feedback that might not fully capture AI specificities. Additionally, this process avoids a narrow
focus, which would be the result of using only a benchmark as a comparison item.
Body of Knowledge Domain VI, Competency E
15
Q
- Which of the following pairs provides examples of potential group harms associated with AI technologies?
A. Safety and deep fakes.
B. Acceleration and reputational.
C. Mass surveillance and facial recognition.
D. Carbon dioxide emission and agricultural.
A
- The correct answer is C. Mass surveillance and facial recognition can lead to group harm given the potential
use of AI technology to target and/or discriminate against specific groups/demographics. Safety and deep
fakes are examples of societal harm, which is different than group harms in that they are not targeted
against a specific group or demographic. Acceleration refers to the potential harm caused by AI advancing
beyond the ability to safeguard it properly. Reputational harm refers to the potential risk to organizations or
individuals from poorly managed AI or competitors creating deep fakes. Carbon dioxide emissions and
agricultural harm refer to the possibility of increased computer reliance affecting the environment.
Body of Knowledge Domain II, Competency A
16
Q
- What are the initial steps in the AI system planning phase?
A. (1). Plan the team; (2). Determine specific use cases the organization wants the AI system to solve; (3).
Identify gaps to achieve the use cases; (4). Identify necessary data for the AI system.
B. (1). Gain stakeholder buy in; (2). Determine specific use cases the organization wants the AI system to
solve; (3). Identify gaps to achieve the use cases; (4). Identify necessary data for the AI system.
C. (1). Determine the budget; (2). Determine specific use cases the organization wants the AI system to
solve; (3). Identify gaps to achieve the use cases; (4). Identify necessary data for the AI system.
D. (1). Determine the business problem to be solved by the AI system; (2). Determine the specific use
cases; (3). Identify gaps to achieve the use cases; (4). Identify necessary data for the AI system.
A
- The correct answer is D. Identifying the business problem to be solved is always the first step in any
significant technology investment. The steps listed in other answers may be relevant but are not initial
considerations.
Body of Knowledge Domain V, Competency A
17
Q
- During the AI winter of the 1970s, funding declined partially due to?
A. Failing to meet expectations from initial hype and promises.
B. Government agencies reallocating budgets to other sciences.
C. Hardware restrictions on complex neural network simulations.
D. Mathematical proofs showing capabilities were overestimated.
A
- The correct answer is A. Disillusion set in when early hopes did not match the actual capability delivered.
Mathematical theories did not yet fully characterize limitations.
Body of Knowledge Domain I, Competency D
18
Q
- Which of the following is NOT true about the relationship of explainability and transparency as they relate
to AI?
A. They are synonymous.
B. They support trustworthy AI.
C. They are distinct characteristics that support each other.
D. They enable users of the system to understand the outcomes.
A
- The correct answer is A. Both are types of documentation that help users understand how the system
produces its outputs. Both support trustworthy AI as defined by NIST, OECD and others. “Explainability and
transparency are distinct characteristics that support each other. Transparency can answer the question of ‘what happened’ in the system. Explainability can answer the question of ‘how’ a decision was made in the
system” (NIST AI RMF, 3.5 Explainable and Interpretable). They are not synonymous. The words do not have
nearly the same meaning but can be presented in ways that may make this appear to be the case.
Body of Knowledge Domain II, Competency B
19
Q
- A regulator has ruled that the AI model used by some features of your product should not be used in its
jurisdiction. What is the most robust way of reacting to this situation?
A. Shut down the service in the jurisdiction to pressure the regulator to change its position.
B. Create a second, separate version of the application and deploy it to users in the jurisdiction.
C. Implement a feature flag to enable or disable the features based on a user’s country of origin.
D. Retrain and deploy your AI model immediately to prove you can react quickly to regulatory demands.
A
- The correct answer is C. Feature flags can be toggled without major engineering efforts, and keeping the
model offline during regulatory negotiations avoids unnecessary confrontations, such as requiring a
redeploy.
Body of Knowledge Domain VI, Competency F
20
Q
- Which of the following is often used as a defense to a claim for copyright infringement with respect to AI?
A. Fair use.
B. Ethical use.
C. Patentability of AI models.
D. AI models made by the government.
A
- The correct answer is A. The fair use doctrine states that excerpts of copyright material may, under certain
circumstances, be copied verbatim for purposes such as criticism, news reporting, teaching and research
without the need for permission from or payment to the copyright holder. Many AI manufacturers rely on
fair use as a copyright infringement defense, alleging that the use of copyrighted material for training an AI
model is ”transformative” (does not merely replicate the material) and does not harm the commercial
market for the copyright holder’s works.
Body of Knowledge Domain III, Competency A
21
Q
- What does the term “generalization” mean regarding an AI model?
A. The model is a good data classifier.
B. The model is effective with new data.
C. The model can solve a variety of business problems.
D. The model can produce good outputs from few inputs.
A
- The correct answer is B. A model’s generalization is its ability to perform well on new data, in addition to
training and initial test data.
Body of Knowledge Domain V, Competency C
22
Q
- How should an organization allocate resources to effectively address AI risks?
A. Allocate resources to highest risk areas.
B. Allocate resources to highest risk factors.
C. Allocate resources to highest risk tolerance.
D. Allocate resources to highest risk outcomes.
A
- The correct answer is A. After performing a risk assessment to determine where AI risk resides in an
organization, priority should be given to the highest risk areas in terms of resources allocated. Risk
tolerance relates to organizational risk appetite, risk factors are inputs to determining a risk level and risk
outcomes relate to adverse scenarios that might transpire. They are related concepts, but when it comes to
prioritizing and allocating resources, best practice suggests it is done by risk area.
Body of Knowledge Domain VI, Competency A
23
Q
- Why should an AI developer separate data into training and test sets?
A. To create models that can be trained quickly.
B. To reduce the inherent biases in stochastic gradient descent.
C. To avoid overfitting to the specific characteristics of the training data.
D. To improve the test validation scores by presenting multiple datasets.
A
- The correct answer is C. It is best practice to separate training and test sets to ensure any random biases in
the training set do not carry over into the final model. In machine learning, overfitting occurs when an
algorithm fits too closely or even exactly to its training data, resulting in a model that can’t make accurate
predictions or conclusions from any data other than the training data.
Body of Knowledge Domain VI, Competency F
24
Q
- When considering the implementation of AI systems, how should companies educate users to best address
their concerns?
A. Abort the AI system to avoid addressing user concerns.
B. Assume that users will self-educate through online resources.
C. Provide highly technical details relevant only to AI professionals.
D. Offer comprehensive information on AI functionalities and limitations.
A
- The correct answer is D. To effectively address user concerns, companies should provide comprehensive
and accessible information regarding the AI system. This includes educating users on the capabilities and
limitations of AI to ensure that they have a balanced and realistic understanding of the AI technology in
place.
Body of Knowledge Domain VII, Competency B
25
Q
- Which of the following is TRUE of a “weak” AI system?
A. It closely mimics the operation of the human brain.
B. It boosts productivity by enabling smarter decision-making.
C. It outperforms humans across a comprehensive range of tasks.
D. It can perform complex tasks and achieve goals in different contexts.
A
- The correct answer is B. Weak AI systems boost productivity by enabling smarter decision making. The other
responses are examples of strong AI systems. “Strong” AI refers to artificial general intelligence, which does
not yet exist. All current AI systems today are considered to be “weak” AI that is suitable for supporting
human decision-making and automating repetitive tasks but does not meet the standard of being
equivalent to human intelligence in terms of understanding context and meaning.
Body of Knowledge Domain I, Competency B
26
Q
- Alex is working for a public transit agency that is building out its AI program. He is collaborating with a local
university to work on an AI project that involves what the NIST AI Risk Management Framework defines as
social responsibility. Which proposal below best aligns with the concepts of social responsibility?
A. Using AI-assisted software to analyze anonymized ridership data to fulfill government reporting
requirements.
B. Attaching sensors on buses to determine and assess heavy traffic periods on specific routes to provide
accurate travel time to passengers.
C. Analyzing video from station surveillance cameras to determine when office trash cans need to be
emptied to save staff from unnecessary trips.
D. Developing an AI-based mobile application that provides pseudonymized assistance to disabled
customers booking point-to-point ride services.
A
- The correct answer is D. NIST refers to the organization’s social responsibility as considering the “impacts of
its decisions and activities on society and the environment through transparent and ethical behavior.” While
each of these are important or useful in one way or another, providing ride assistance to disabled customers meets the goal of ethical social behavior. By pseudonymizing the information and the service
itself, the company is providing a socially responsible service that protects the individual.
Body of Knowledge Domain IV, Competency C
27
Q
- The engineering team is working on training an AI model with a large, new dataset culled from social
media. The team has built out the model to predict trends in car buying for the coming year based on social
media posts. The AI model has a clear bias toward one particular automotive model. It turns out the recent
data input was pulled when the car was featured in a blockbuster Hollywood movie. This is an example of
which type of AI bias?
A. Human-cognitive.
B. Systemic and directed.
C. Computational and statistical.
D. Psychological and sociological.
A
- The correct answer is C. Computational and statistical biases may stem from errors due to non-
representative samples. In this example, the data is skewed based on the movie release.
Body of Knowledge Domain VI, Competency E
28
Q
- What is the AI Liability Directive proposed by the European Commission?
A. A directive that will regulate organizations’ implementation of AI systems.
B. A directive that will provide for legal responsibility by organizations using AI systems.
C. A directive that will minimize and regulate the development and adoption of AI systems.
D. A directive that will reduce organizational legal risks and costs associated with AI systems.
A
- The correct answer is B. The AI liability directive is a proposal by the European Commission for a directive on
civil liability for AI systems that complements and modernizes the EU liability framework to introduce new
rules specific to damages caused by AI systems.
Body of Knowledge Domain III, Competency C
29
Q
- George is developing an advertising campaign for Splendide Company’s new product. George wants to
incorporate colorful, eye-catching modern art as the backdrop for the product in print and digital ads. He
uses a generative AI application to create the campaign and plans to use the prompt, “colorful abstract
paintings in the style of contemporary artist Alexandra Artiste.” George determines that, because he wants
to reference a specific artist, he should obtain a copyright license from the artist. To protect Splendide from
liability, he wants to include terms in the license that provide exceptions to copyright infringement
indemnification, as is commonly done with copyright infringement licensing contracts. Which of the
following is a common exception to copyright infringement indemnification that does NOT work well in an
AI model?
A. Identifying the licensee as the owner of the resulting work.
B. Requiring the licensee to use the artwork to train the model.
C. Identifying the output from the AI model as an original work.
D. Combining the licensed artwork with other artwork in the AI model.
A
- The correct answer is D. Typical exceptions to intellectual property infringement indemnification in IP
license agreements include exceptions for modifications to the licensed work, unauthorized combinations of
the licensed work and other works, and use of the licensed work beyond the scope authorized in the license
agreement. Because AI models modify, combine and use the works on which they are trained in other
contexts, these exceptions to infringement indemnification do not work.
Body of Knowledge Domain VII, Competency A
30
Q
- Nolan is in the IT procurement department of Happy Customer Inc., a provider of call center solutions.
Other companies are using Happy Customer Inc. services to help their clients with easy, first-level support
requests. Happy Customer Inc. wants to offer its B2B clients a new text-based chatbot solution that offers
the right answers to a given set of questions. What type of AI model should Nolan look for?
A. A robotics model.
B. A statistical model.
C. A decision tree model.
D. A speech recognition model.
A
- The correct answer is C. A decision tree model predicts an outcome based on a flowchart of questions and
answers. A statistical model is used to model the relationships between two variables. A speech recognition
model is used to analyze speech; e.g., for voice assistants. A robotics model is based on a multidisciplinary
field that encompasses the design, construction, operation and programming of robots and allows AI
systems and software to interact with the physical world without human intervention.
Body of Knowledge Domain I, Competency C
31
Q
- Which of the following best describes feature engineering?
A. Transforming or manipulating features to enhance the performance of the model.
B. Consulting with subject matter experts about features to include in the training data.
C. Anonymizing raw data so personal information is not included in the features for the model.
D. Annotating the raw data to create features with categorical labels that align with internal terminology.
A
- The correct answer is A. While the other options are all actions that are performed in the AI development life
cycle, feature engineering refers specifically to the transformation of “features” used as inputs for the
model.
Body of Knowledge Domain V, Competency C