ML Security Flashcards
-Science of making things smart or human tasks performed by machines (example: visual recognition, Natural Language processing)
A. Artificial Intelligence (AI)
B. Machine Learning (ML)
C. Deep Learning (DL)
A. Artificial Intelligence - Science of making things smart or human tasks performed by machines (example: visual recognition, Natural Language processing) Ability of machines to perform human tasks.
-One of many approaches to AI that uses a system capable of learning from experience. Makes decisions based on data rather than algorithm.
A. Artificial Intelligence (AI)
B. Machine Learning (ML)
C. Deep Learning (DL)
B. Machine Learning (ML)
-One of many approaches to AI that uses a system capable of learning from experience. Makes decisions based on data rather than algorithm.
-A set of techniques for implementing machine learning that recognizes patterns of patterns. (for example: image recognition). Identifies object boundary, type, structure.
A. Artificial Intelligence (AI)
B. Machine Learning (ML)
C. Deep Learning (DL)
C. Deep Learning (DL)
A set of techniques for implementing machine learning that recognizes patterns of patterns. (for example: image recognition)
Different applications work with different data.
What is an AI Threat?
A. Hacker break system through stickers on stop signs
B. Hackers can bypass facial recogniton
C. Hackers can break web platforms and filters via social media.
D. Hackers like Nest Assistance can be broken
E. All the above
E. All the above are AI Threats.
a. Self Driving Car Threat:
Hacker break system through stickers on stop signs
b. Classification / Image Threat:
Hackers can bypass facial recogniton
c. Social Media Threat:
Hackers can break web platforms and filters via social media.
d. Home Automation Threat:
Hackers like Nest Assistance can be broken
What algorithm categories are the following categories?
-Classification
-Regression
A. Supervised
B. Unsupervised
C. Semi-Supervised
D. Reinforcement Learning
-Classification
-Regression
A. Supervised
What algorithm categories are the following categories?
-Clustering
-Dimensionality Reduction
A. Supervised
B. Unsupervised
C. Semi-Supervised
D. Reinforcement Learning
-Clustering
-Dimensionality Reduction
B. Unsupervised
What algorithm categories are the following categories?
-Generative models
A. Supervised
B. Unsupervised
C. Semi-Supervised
D. Reinforcement Learning
-Generative models
C. Semi-Supervised
What algorithm categories are the following categories?
-reinforcement learning
D. Reinforcement Learning
-reinforcement learning
D. Reinforcement Learning
How are AI attacks classified?
A. confidentiality, availability, and integrity (triad)
B. Espionage, sabotage, and fraud
C. Availability, fraud, and integrity
D.A and B
How AI attacks classified
A. confidentiality, availability, and integrity (triad)
and
B. Espionage, sabotage, and fraud
What are the steps to start an AI Security Project?
I. Identify an AI object and a task
ii. understand algorithm category and algorithm itself
iii. choose an ai attack relevant to your task and algorithm
A. 3,2,1
B. 2,1,3
C. 1,2,3
D. 3,1,2
Start and AI Security Project Steps:
C. 1,2,3
I. Identify an AI object and a task
ii. understand algorithm category and algorithm itself
iii. choose an ai attack relevant to your task and algorithm
True or False:
AI Threats are similar / mostly the same, but their appraoches are different
True
AI Threats are similar / mostly the same, but their appraoches are different
Reasoning: The difference comes in Algorithms
Steps to Set up your Environment:
i. have nvidia gpu or not
ii. choose operating system (recommend Ubuntu)
iii. follow guidelines provided
A. 3,2,1
B. 1,2,3
C. 2, 1, 3,
D. 3,1,2,
Steps to Set up your Environment:
i. have nvidia gpu or not
ii. choose operating system (recommend Ubuntu)
iii. follow guidelines provided
B. 1,2,3
Which attack cannot be used for breaking integrity of AI?
A. backdoor
b. adversarial
c. inference attack
d. poisoning
c. inference attack
inference attack- dont break functionality they extract critical data
REASONING:
Adversarial attacks- break integrity by misclassification
Poisoning - poisoning breaks integrity
Backdoor-backdoor attacks break integrtiy
What is the most important hardware for this course?
a. CPU
b. GPU
c. RAM
d. HDD
most important hardware
b. GPU
Model is getting trained on label data set. Examples is Classification and regression:
A. Supervised
B. Unsupervised
C. Semi-Supervised
D. Reinforcement Learning
A. Supervised
Supervised- Model is getting trained on label data set. Examples is Classification and regression.
Model is attempting to automatically find structure in the data by extracting useful features and analyzing its structure. Examples: Clustering, Association, Dimension Reduction (Generalization)
A. Supervised
B. Unsupervised
C. Semi-Supervised
D. Reinforcement Learning
B. Unsupervised
Unsupervised - Model is attempting to automatically find structure in the data by extracting useful features and analyzing its structure. Examples: Clustering, Association, Dimension Reduction (Generalization)
Imagine a road sign detection system aiming to classify signs. Supervised learning approach is usually used. Examples of certain groups is known and all classes should be defined in the beginning. This method is:
A. Classification
B. Regression
C. Clustering
A. Classification
Classification - imagine a road sign detection system aiming to classify signs. Supervised learning approach is usually used. Examples of certain groups is known and all classes should be defined in the beginning.
The knowledge about the existing data is utilized to have an idea about new data (Past explains future). Ex. is stock price prediction.
A. Classification
B. Regression
C. Clustering
B. Regression
Regression - The knowledge about the existing data is utilized to have an idea about new data (Past explains future). Ex. is stock price prediction.
Supervised learning approach is usually used. Examples of certain groups is known and information about classes in data is unknown.
A. Classification
B. Regression
C. Clustering
C. Clustering
Clusteirng - Supervised learning approach is usually used. Examples of certain groups is known and information about classes in data is unknown.
Algorithms: KNN (K-Nearest Neighbor), K-Means, Mixture Model (LDA)
Necessary if you deal with complex systems with unlabeled data and many potential features (facial recogntion)
A. Classification
B. Dimension Reduction (Generalization)
C. Clustering
D. Generative Models
B. Dimension Reduction (Generalization)
Dimension Reduction - Necessary if you deal with complex systems with unlabeled data and many potential features (facial recogntion)
_______ designed to stimulate the actual data and not decisions, based on previous data.
AI data based on previous data.
A. Classification
B. Dimension Reduction (Generalization)
C. Clustering
D. Generative Models
D. Generative Models
Generative Models - AI data based on previous data. designed to stimulate the actual data and not decisions, based on previous data.
________ A behavior that depends on the changing environment.
A. Reinforcement Learning
B. Dimension Reduction (Generalization)
C. Active Learning
D. Generative Models
A. Reinforcement Learning -A behavior that depends on the changing environment.
Reinforcement Learning
(Behavior should react to the changing environment. Trial and Error.)
_____ A subclass of reinforcement learning, which helps correct errors, in addition to the environment changes
A. Reinforcement Learning
B. Dimension Reduction (Generalization)
C. Active Learning
D. Generative Models
C. Active Learning
Active Learning - A subclass of reinforcement learning, which helps correct errors, in addition to the environment changes
Acts as a teacher who can help correct errors in addition to environment changes
_________ are inputs to machine learning models that results in an incorrect input.
A. adversarial example
B. king penguin
C. starfish
D. baseball
A. adversarial example
adversarial example - inputs to machine learning models that results in an incorrect input.
Reasoning:
b. King penguin - is a adversarial example
c. starfish - is a adversarial example
d. baseball - is an adversarial example
________ - Is the cause for ML models to create a false prediction?
A. adversarial example
B. king penguin
C. starfish
D. baseball
A. adversarial example
Adversarial example - Is the cause for ML models to create a false prediction?
___________ tries to move inputs across the decision boundary?
A. adversarial example
B. king penguin
C. adversarial attacks
D. baseball
C. adversarial attacks
ADVERSARIAL ATTACKS- tries to move inputs across the decision boundary.
How AI Attacks Work:
What do AI Attacks calculate?
A. How much inputs change affect the outputs.
B. How much outputs change affect inputs
C. Decision boundary
D. Neither
A. How much inputs change affect the outputs.
AI Attacks work by calculating how much INPUT changes AFFECT OUTPUT.
What do you need to calculate AI Attacks?
a. Gradient
b. Loss function
c. Optimal Perturbations measuring Lp Norms
d. All the above
d. All the above
What you need to calculate AI Attacks:
1. Gradient
2. Loss Function
3. Optimal Perturbations measuring Lp Norms
______ defines how good a given model is at making predictions for a given scenario.
a. Gradient
b. Loss function
c. Optimal Perturbations measuring Lp Norms
d. None of the Above
b. Loss function
Loss Function - Defines how good a given model is at making predictions for a given scenario
What method has the following characteristics:
-it has its own curve and gradients
-slope of the curve indicates the appropriate way of updating the parameters to make the model more accurate in case of prediction
a. Gradient
b. Loss function
c. Optimal Perturbations measuring Lp Norms
d. None of the Above
b. Loss function
-it has its own curve and gradients
-slope of the curve indicates the appropriate way of updating the parameters to make the model more accurate in case of prediction
____ a fancy work for derivative, also known as vector. Means rate of change.
a. Gradient
b. Loss function
c. Optimal Perturbations measuring Lp Norms
d. None of the Above
a. Gradient
Gradient - a fancy work for derivative, also known as vector. Means rate of change.
_____ attacks try to move inputs across the decision boundary.
a. Gradient
b. Loss function
c. Optimal Perturbations measuring Lp Norms
d. None of the Above
c. Optimal Perturbations measuring Lp Norms
_____ attacks try to move inputs across the decision boundary.
Perturbation - attacks try to move inputs across the decision boundary
____ denotes the maximum change for all pixels in the adversarial examples
a. l(8)
b. u
c. l0
d. none of above
a. l(8)
__l(8)____denotes the maximum change for all pixels in the adversarial examples. (Used in Perturbation)
_____ number of pixels changed in the adversarial examples.
a. l(8)
b. u
c. l0
d. none of above
c. l0
___l0___number of pixels changed in the adversarial examples. (Used in Perturbation)
Topic “If ML Algorithms have Vulnerabilities”
Ex. malefactor is implementing bypass techniques is a “spam”, sending out. All algorithms on ML models are based (from SVMs to random forests and neural networks) which are vulnerable to different kinds of adversairal inputs. This type of attack was targets what form of AI?
a. Classification
b. Random Forests
c. K-Means
d. Regression
a. Classification
Adversarial Classification -
Is an attack where malefactor is implementing bypass techniques is a “spam”, sending out. All algorithms on ML models are based (from SVMs to random forests and neural networks) which are vulnerable to different kinds of adversairal inputs.
Which type of ML algorithms has few examples of practical attacks?
a. Classification
b. Random Forests
c. K-Means
d. Regression
d. Regression
Regression- a type of ML Algorithms that has FEW EXAMPLES of PRACTICAL attacks.
Source: “Adversarial Regression with Multiple Learners 2018”
True / False:
Most attacks used in Classification can be used in Regression?
TRUE
MOST attacks used in Classification CAN BE USED in Regression
Reasoning: Condition Based Instance and Null Analysis
Which type of ML algorithms would succumb to auto-encoders prone to attacks or attack such as (input reconstruction, spoofs)
Input image the model encodes the lower dimensional then uses that to reconstruct the original image.
a. Classification
b. Generative Models
c. K-Means
d. Regression
b. Generative Models
Generative Models (GANS) or auto-encoders - would succumb to auto-encoders prone to attacks such as (input reconstruction, spoofs)
Input image the model encodes the lower dimensional then uses that to reconstruct the original image.
Which type of ML algorithm can be used for malware detection?
a. Classification
b. Generative Models
c. K-Means
d. Clustering
d. Clustering
Clustering - used for malware detection.
Clustering algorithm is K-Nearest Neighbors (KNN)
Note: Training data comes from the wild.
______ is the most common dimensionality reduction algorithms?
A. PCA
B. Clustering
C. Generalization
D. MNIST
A. PCA
PCA- is the most common dimensionality reduction algorithm.
Why type of ML Algorithm is used sensitive to outliers that can be exploited by contaminating training data?
A. PCA
B. Clustering
C. Generalization
D. MNIST
A. PCA
PCA - sensitive to outliers that can be exploited by contaminating training data.
What does this example show (insert image)
It allows dramatically decreasing the detection rate for DoS attacks
______ which type of algorithm is used for Facial Recognition? An example of this is using your face to unlock your iphone.
A. PCA
B. Clustering
C. Generalization
D. MNIST
A. PCA
PCA- algorithm is used for Facial Recognition. An example of this is using your face to unlock your iphone.
RL the framework known as DNN , using DNN for Feature Selection and Q Functional Approximation. Hence enable
What are the steps of a Deep Reinforcement Learning Attack (DQN)?
i. attacker observes current state and transitions in environment
ii. attacker estimates best action according to adversarial policy
iii. attacker crafts perturbation to induce adversarial action
iv. attacker applies perturbation
v. perturbed input is revealed to target
vii. attacker waits for targets action
A. 1,2,3,4,5,6
B. 6,5,4,3,2,1
C. 4,3,2,5,6,1
D. 2,5,3,4,6,1
steps of a Deep Reinforcement Learning Attack (DQN)?
A. 1,2,3,4,5,6
i. attacker observes current state and transitions in environment
ii. attacker estimates best action according to adversarial policy
iii. attacker crafts perturbation to induce adversarial action
iv. attacker applies perturbation
v. perturbed input is revealed to target
vii. attacker waits for targets action
What is the most wide spread attack method?
a. LBFGS
b. FGSM (Fast Gradient Side Method)
c. DQN
d. none of the above
b. FGSM (Fast Gradient Side Method)
FGSM-
_____ attack does the following:
1. Takes the label of the least likely class predicted by network
2. The computed pertrubation is subtracted from original image
3. This maximizes the probability that the network predicts target as the label of the adversarial example
a. LBFGS
b. FGSM (Fast Gradient Side Method)
c. DQN
d. none of the above
b. FGSM (Fast Gradient Side Method)
FGSM works using the following steps:
- Takes the label of the least likely class predicted by network
- The computed pertrubation is subtracted from original image
- This maximizes the probability that the network predicts target as the label of the adversarial example
_____ attack method was very time consuming, especially for larger images and practically non-applicable
a. LBFGS
b. FGSM (Fast Gradient Side Method)
c. DQN
d. none of the above
a. LBFGS
LBFGS - attack method was very time consuming, especially for larger images and practically non-applicable
Which ML task category is required if you deal with complex systems with unlableled data and many potential features?
a. classification
b. clustering
c. reinforcement learning
d. dimentionality reduction
d. dimentionality reduction
Dimentionality Reduction- ML category required if you deal with complex systems with unlabeled data and many potential features.
How do you measure Adversarial Attacks?
A. using Gradient
B. using Loss Function
C. using L-p norm
D. using the size of ML Model
C. using L-p norm
L-p norm used to measure changes for adversarial attacks
Which ML task category has the biggest number of research papers?
A. Clustering
B. Reinformcement Learning
C. Classification
D. Regression
C. Classification
Classification - Has the larges number of research papers spanning 300
Why is FGSM method better than BFGS method?
A. Requires less information
B. FGSM is more accurate
C. More universale
D. The FGSM method is faster
D. The FGSM method is faster
Reasoning-
Not C. LBFGS is more universal but slower and less accurate
Which dataset is better for testing practical attacks?
A.CIFAR
B. MNIST
C. LFW
D. ImageNew
B. MNIST
MNIST is the dataset best for testing practical attacks. The MNIST dataset is the smallest one, and all tests will be less time-consuming with lower computation cost
What are the reasons to Hack AI?
A. AI is eating software
B. Exansion of technology related to Cybersecurity
C. Vulnerable to various cyber attacks like any other algorithms
D. All Above
D. All Above
Hack AI
-AI is eating software
-Expansion of tech related to cybersecurity
-vulnerability to various cyber attacks like any other algorithms
Autonomous cars use image classification such as Identification of Raw Science
______ can lead to horrible accidents
A. Spoofing of Raw Science
Autonomous cars use image classification such as Identification of Raw Science
Spoofing of Raw Science- can lead to horrible accidents
What are AI risks in the Cybersecurity Industry?
A. Bypass spam filters
B. Bypass threat detection solutions
C. Bypass AI-based Malware Detection tools
D. All Above
AI risks in Cybersecurity Industry
D. All Above
-Bypass spam filter
-Bypass threat detection solutions
-bypass AI based malware detection tools
What are AI risks in the Retail Industry?
A. bypass Facial recognition
AI Risks in Retail Industry:
A. bypass Facial recognition
(used w/ makeup, surgerty etc.)
How does AI use in Retail
a. Behavior retail of clients
b. Optimize business processes
c. all above
c. all above
AI use in retail:
1. Behavior retail of clients
2. Optimize business processes
How is AI used in Smart Home Industry?
Amazon echo recognizes Noise as a Comment. This voice is recognized as certain instructions.
a. forge voice commands
AI used in Smart Home Industry
a. forge voice commands
How AI used in Web and Social Media Industry
a. Fool sentiment analysis of movie reviews, hotels etc.
How AI used in Web and Social Media Industry
- Fool sentiment analysis of movie reviews, hotels etc.
Misinterpret a comment
How AI used in Finance
a. trick anomaly and fraud detection engines
How AI used in Finance
- trick anomaly and fraud detection engines
What are ways to prevent Frauds using ML?
a. learn customer behavior
b. analysis of aggregated data
c. analysis of social graphs
d. automation of routine processes
e. control use ID information
f. ALL ABOVE
f. ALL ABOVE
-learn customer behavior
- analysis of aggregated data
-analysis of social graphs
- automation of routine processes
- control use ID information
Confidentiality is associated with:
a. Gather System Insights
b. Disable AI System Functionality
c. Modify AI logic
Confidentiality is associated with:
a. Gather System Insights
-Obtain insights into the system
-utilize the received info or plot more advanced attacks
Which triad is the following:
(A malicious person deals with a ML system that is an Image Recognition System. They get to learn more about the internals or the datasets from this system)
a. confidentiality
b. availability
c. integrity
a. confidentiality
(A malicious person deals with a ML system that is an Image Recognition System. They get to learn more about the internals or the datasets from this system)
Reasoning-
Confidentiality because they are gathering information about the system and that information can be used to plot attacks.
NOT: Integrity because they did not change logic
NOT: Availability because they did not disable anything
Availability is associated with:
a. Gather System Insights
b. Disable AI System Functionality
c. Modify AI logic
b. Disable AI System Functionality
Availability = Disable AI System Functionality
Which triad is the following:
-Flood AI with requests, which demand more time
-Flood with incorrect classified objects to increase manual work
-Modify a model by retraining it with wrong examples
-Use computing power of an AI model for solving your own tasks
a. confidentiality
b. availability
c. integrity
b. availability
-Flood AI with requests, which demand more time
-Flood with incorrect classified objects to increase manual work
-Modify a model by retraining it with wrong examples
-Use computing power of an AI model for solving your own tasks
Integrity is associated with:
a. Gather System Insights
b. Disable AI System Functionality
c. Modify AI logic
c. Modify AI logic
Integrity = Modify AI Logic
Which triad is the following:
-Ex. Make autonomous cars, believe that there is a cat on the road, when in fact it is a car.
-2 different ways to interact with a system at the learning or production stage
1) poinsoning
2) evasion
a. confidentiality
b. availability
c. integrity
c. integrity
This attack is integrity because you modified the car to think it was a cat when it was really a car.
2 types of integrity (modify ai logic)
1. Poisoning - attackers poison some data in the training dataset
2. Evasion- attackers exploit vulnerabilities of an algorithm by showing modified picture at the production stage
Which integrity interaction is this?
________ attackers alter some data in the training dataset
a. poisoning
b. evasion
c. modify ai logic
a. poisoning
POSIONING- attackers poinson / alter some data in the training dataset
A attack form of Integrity
Which integrity interaction is this?
______ attackers exploit vulnerabilities of an algorithm by showing the modified picture at the production stage
a. poisoning
b. evasion
c. modify ai logic
b. evasion
EVASION - attackers exploit vulnerabilities of an algorithm by showing the modified picture at the production stage
A attack form of Integrity
_______ a procedure where someone is trying to exploit ML model, by injecting malicious data into the training dataset.
a. poisoning
b. evasion
c. modify ai logic
a. poisoning
Poisoning - a procedure where someone is trying to exploit ML model, by injecting malicious data into the training dataset.
_________ attacks change classification boundry while
_________ attacks change input examples
a. Poisoning, Adverarial
b. Adversarial, Poisoning
c. Posioning, Evasion
d. Evasion, Poisoning
a. Poisoning, Adverarial
Poisoning attacks - change classification boundry WHILE
Adversarial attacks - change input examples
True or False
If points are added to the training data, the decision boundry will change
True
If points are added to the training data, the decision boundry will change
______ attack allows an adversary to modify solely the labels in supervised learning datasets but for arbitrary data points
A. Label modification
B. Poisoning
C. Evasion
D. Data Injection
A. Label modification
label modification attack allows an adversary (enemy) to modify solely the labels in supervised learning datasets but for arbitrary (opposite) data points
______ An adversary (enemy) does not have access to the training data nor to the learning algorithm, but has the ability to add new data to the training set
A. Label modification
B. Poisoning
C. Data Injection
D. Adversarial
C. Data Injection
Data Injection - An adversary (enemy) does not have access to the training data nor to the learning algorithm, but has the ability to add new data to the training set
_______ An adversary does not have access to the learning algorithm but has full access to the training data
A. Label modification
B. Data Modification
C. Data Injection
D. Adversarial
B. Data Modification
Data modification - An adversary does not have access to the learning algorithm but has full access to the training data.