Domande Flashcards
According to the scientific visualization rules presented in class, is it possible to plot a graphical representation of the confidence level of one single figure of merit (like the accuracy) of your trained model?
- No, the confidence intervals data have different units and meaning and hence can not be represented in the same plot
- Yes, the confidence interval data have different units and meaning but they can be represented in the same plot using different visual attributes like “slope” and “area”
- Yes, the confidence interval data have the same units and meaning and they can be represented in the same plot
Yes, the confidence interval data have the same units and meaning and they can be represented in the same plot
The number of parameters to be fixed during a complete training in a deep learning model like the VGGNet presented in the course is about:
- < 100000
- > 100 Million
- about 1 Million
- about 10 Million
> 100 Million
Considering the class discussing about the basic metrics in data similarity, given a vector A, vector B, a real number alpha, and the cosine metrics cos(A,B) it is possible to say that
- alpha * cos(A,B) = cos(alpha*A, B)
- cos(A,B) = cos(alphaA, alphaB)*
- cos(A,B) = cos(alphaA, B) = cos(A, alphaB)
- alpha * cos(A,B) = cos(alphaA, alphaB)
cos(A,B) = cos(alphaA, alphaB)
Referring to the class discussion on data leakage what is the worst situation?
- The unwanted leakage of data from training dataset to test data set
- None of the other options since transferring data from test and/or training dataset is normal when the accuracy of the model is tested
- The unwanted leakage of data from test dataset to training data set since you are subtracting data to the generalization test, making the situation more pessimistic
- The unwanted leakage of data from test dataset to training data set since you are subtracting data to the generalization test, making the situation more optimistic
The unwanted leakage of data from test dataset to training data set since you are subtracting data to the generalization test, making the situation more optimistic
What task of an intelligent vision system is associated to following description: split or separate an image into regions using features, patterns and colors to facilitate recognition, understanding, and Region Of Interests (ROI) processing and measurements.
- Model training
- Post processing
- Enhancing
- Segmentation
- Feature engineering
Segmentation
According to the class discussion, text prefiltering is often used as input for a neural network to deal with a large text input making the networks able to classifiy the input.
- True, using the hamming distance as prefilering
- True, using the cosine distance as prefilering
- True, using the string approximate match distance as prefilering
- True, using the discrete gradient descent as prefilering
- True, using the so-called “word embeddings” technique
True, using the so-called “word embeddings” technique
According to the class discussion, what is Greedy Layer-Wise Training?
- A supervised training step to improve auto-encoders
- An unsupervised training step to classical feedforward networks
- An unsupervised training step to improve auto-encoders
- A supervised training step to classical feedforward networks
An unsupervised training step to improve auto-encoders
The following activity: a) Data Selection; b) Data Filtering; c) Data Enhancing …
- Are part of the classical machine learning approaches and they are (correctly) used also in deep learning applications
- All the other options are correct*
- Are part of the job of the artificial intelligent specialist in normal activities
- Contribute to keep lower the complexity of the learning task
a), b) and c) are extremely important in the final behavior of the trained model and the complexity of the training task
All the other options are correct
The Inception-v3 deep learning pretrained model discussed during the course is a model for
- Post processing
- None of the other options
- Segmentation
- Image enhancing
- Image classification*
Image classification
Intelligent vision systems can achieve Semantic segmentation by
- A hybrid approach by blob detection to select candidate ROIs and then image classification of the ROIs
- A complete fully convolutional solution
- A hybrid approach by blob detection to select candidate ROIs and then image segmentation of the ROIs
- None of the other options
A complete fully convolutional solution
Considering the possible Intelligent Vision tasks which is the correct option?
- Instance Segmentation is more complex than Object Detection
- Instance Segmentation is less complex than Object Detection
- Instance Segmentation and Object Detection have a similar complexity
- The other otpions are not Intelligent Vision tasks
Instance Segmentation is more complex than Object Detection
In a given picture ImmA you see 1 car and 5 people in a city background. Considering the Intelligent systems IS processing the image ImmA and producing in output the label “humans”, what Intelligent Vision task is performing?
- Image classification
- Instance segmentation
- Object detection
- Semantic segmentation
Image classification
According to the class discussion, considering the training of deep learning models on standard CPUs and standard commercial GPUs boards, what is the gain in training performance (time) and efficiency (energy) for a medium/large-size project?
- About 100x in performance and 10x in efficiency
- More than 100x in performance and more than 5x in efficiency
- About 10x in performance and 5x in efficiency
- About 2x in performance and 2x in efficiency
About 10x in performance and 5x in efficiency
A basic industrial setup for Intelligent vision systems is typically composed by the following elements
- Standard industrial smart camera with optics, external processing HW and SW units, illumination system
- Standard industrial camera with optics, illumination system
- Just a standard industrial camera with optics
- Standard industrial camera with optics, processing HW and SW units, illumination system
- Standard industrial camera with optics, processing HW and SW units
Standard industrial camera with optics, processing HW and SW units, illumination system
In agreement to the class discussion, what kind of labelling error is generally the worst case for the accuracy of the generalization of the model? ERR1 = Duplications with same labels, EER2 = Duplications with different labels
- ERR1 is equalt to EER2 by definition
- ERR2 is the worst case
- ERR1 is the worst case
- ERR1 is roughly equalt to EER2 in general
ERR2 is the worst case
According to the discussion presented in class about the data visualization, and considering the following steps of the design workflow 1) Get Data, 2) Clean Manipulate Data, 3) Train models, 4) Test Data, 5) Improve the design, which are the main step/steps where data visualization should be involved?
- # 2 and #5
- # 4
- # 3 and #5
- # 5
- # 1
2 and #5
According to the class discussion, the convolution/correlation operations are of foundamental relevance for many deep learning models. What is the characteristic of the autocorrelation map produced by a generic image?
- It is not possible to create an autocorrelation map from one single images, two different images are needed
- None of the other options
- A flat and noisy central plateau
- An evident spike at the center with a very well defined maximum
An evident spike at the center with a very well defined maximum
Considering the class discussion about feature preprocessing/engineering, alogarithmic scaling to one feature values is typically applied in a case of
- A very large range in the values (>0)
- Input coming from the preprocessing of long texts
- Negative values
- Outliers presence
- Feature values are integer numbers
Feedback
A very large range in the values (>0)
According to the notation used in class, which kind of a model is described by the equation
f(x) = sgn(w x + b)
- Liner regressor
- Soft-max neuron
- Liner classifier
- Sigmoidal neuron
- Gradient descent formula
- Number of the model’s parameters
Liner classifier
A tensor processing unit (TPU) is
- A part of a model of the Convolutional Neural Network used to process dedicated tensorial activation functions in the neurons
- An internal unit of the Arm processor architecture introduced to support 8-bit fixed-point matrix multiplication for deep learning models
- An AI accelerator application-specific integrated circuit (ASIC) and the related board developed specifically for neural network machine learning
- None of the other options
An AI accelerator application-specific integrated circuit (ASIC) and the related board developed specifically for neural network machine learning
You have a feature in your dataset with the following values F2 = [ -13 0 1 2 4 128 ], which normalization will give you the following F2_norm = [0 0 1 2 4 10 ]
- Z-score
- Min-MAX
- Clipping
- A different type of normalization
Clipping
You have a feature in your dataset with the following values F2 = [ -13 0 1 2 4 128 ], which normalization will give you the following F2_norm = [0 0 1 2 4 10 ]
- Z-score
- Min-MAX
- Clipping
- A different type of normalization
Clipping
The design of intelligent systems for Industry 4.0 applications should be compliant to the following main design principles.
- Interoperability, Information transparency, Improved technical assistance, Decentralized decisions
- Interoperability, Information transparency, Improved technical assistance
- Interoperability, Information transparency, Improved technical assistance, Wireless connectivity
- Interoperability, Information transparency, Decentralized decisions
La risposta corretta è: Interoperability, Information transparency, Improved technical assistance, Decentralized decisions
Machine Learning on CPUs offer the following advantages
- Ease of portability and use-case flexibility, Market availability at different performance and prices
- Ease of portability and use-case flexibility, Market availability at different performance and prices, Deployment across a wide spectrum of devices
- Ease of portability and use-case flexibility, Deployment across a wide spectrum of devices
- Market availability at different performance and prices, Deployment across a wide spectrum of devices
Ease of portability and use-case flexibility, Market availability at different performance and prices, Deployment across a wide spectrum of devices
The GoogLeNet deep learning pretrained model discussed during the course is model for
- Post processing
- None of the other options
- Image Enhancing
- Image classification
- Segmentation
Image classification
Considering IoT devices as source of data for external intelligent systems (IS is not intended to be embedded into the IoT device), what kind of IoT devices can be really used?
- Passive data IoT devices
- Active data IoT devices
- Dynamic data IoT devices
- All of the above
- None of the above
All of the above
Referring to the class discussion, the (correct) design practice for neural networks considers
- Start with deep learning models since they are the cutting edge and most advanced technology that we have now
- Start with deep learning models since they are the cutting edge and most advanced technology we have now, and then use classicals method as reference
- Start with simple neural networks before to consider deep learning models
Start with simple neural networks before to consider deep learning models
The missing values can also be occupied by computing mean, mode or median of the observed given values.
- This is very unusual and not common in practice
- This is a very simple and effective solution in case the learning method is not capable to deal with missing data
- This is not possible, since that is just descriptive statistics about the features, and cannot be used to fill missing data
This is a very simple and effective solution in case the learning method is not capable to deal with missing data
Referring to the class discussion on data leakage what is the worst situation?
- The unwanted leakage of data from test dataset to training data set
- The unwanted leakage of data from training dataset to test data set
- None of the above since transferring data from test and/or training dataset is normal when the accuracy of the model is tested
The unwanted leakage of data from test dataset to training data set
An additional information can allow the model to learn or know something that it otherwise would not know and in turn invalidate the estimated performance of the model being constructed. This is called:
- Data leakage
- Data pre-processing
- Data harmonization
- Data wrangling
Data leakage
The degrees of freedom for a given problem are the number of independent problem variables which must be specified to uniquely determine a solution. Hence the #DoF is important to be considered
- To design the number of vectors in the learning dataset.
- To avoid overfitting problem in the model
- All the above
- None of the above
All the above
About the cosine metrics it is possible to say that:
- Two vectors with the same orientation have a cosine similarity of 1
- Two vectors oriented at 90° relative to each other have a similarity of 0
- All of the above
- None of the above
All of the above
What similarity feature/features discussed in class offers/offer the property to allow a fast comparison based on a short 1D vector of elements or bits
- phash
- ahash
- All the above
- Cross-correlation
All the above
In agreement to the class discussion, which description better describes the design activity?
- Similarity in the dataset requires more space and processing time
- Similarity in the dataset can improve generalization
- Both of the above
- None of the above
Both of the above
In agreement to the class discussion, in a dataset of 1100 labelled images, the search for duplications is typically achieved…
- by manual exploration of the dataset for better results since the number of images is not critical
- by automatic iterations
by automatic iterations
In agreement to the class discussion, what kind of labelling error is generally the worst case for the accuracy of the generalization of the model? ERR1 = Duplications with same labels EER2 = Duplications with different labels
- ERR1
- ERR2
- ERR1 = EE2
ERR2
According to the class discussion, what is the characteristic of the self-correlation (𝑂 = 𝑥𝑐𝑜𝑟2(𝐴, 𝐴)) map produced by a generic image?
- A flat and noisy central plateau
- An evident spike at the center with a very well-defined maximum
- It is not possible to create an autocorrelation map from one single images, two different images are needed
An evident spike at the center with a very well-defined maximum
According to the class discussion, about the relationship between the operation of cross-correlation and convolution it is possible to say that:
- They are very similar in meaning and mathematical expression
- Despite the mathematical expression is similar, the meaning and their use is completely different
- There is no specific relationship since they are different in meaning and mathematical expressions
They are very similar in meaning and mathematical expression
If your data set contains extreme outliers, it better to use as preprocessing
- Feature clipping
- Min-max normalization
- Z’ norm
Feature clipping
A logarithmic scaling to one feature values is typically applied in a case of
- Outliers’ presence
- Negative values
- A very large range in the values (>0)
A very large range in the values (>0)
According to the scientific visualization rules presented in class, if you are plotting many figures of merit obtained by your trained neural network on a new dataset, which is the correct ranking of visual attributes to be used? Left: low accuracy Right: HIGH ACCURACY
- Color intensity > Hue > Length
- Area > Length > Hue
- Slope > Angle > Volume
- Hue > Area > Length
Hue > Area > Length
According to the scientific visualization rules presented in class, is it possible to plot a graphical representation of the confidence level of your figures of merit of your trained model?
- No, it is a statistical index with different units and meaning and hence cannot be represented in the same plot
- Yes, the confidence interval data have the same units and meaning, and they can be represented in the same plot
Yes, the confidence interval data have the same units and meaning, and they can be represented in the same plot
According to the discussion presented in class about the data visualization, and considering the following steps of the design workflow 1) Get Data, 2) Clean Manipulate Data, 3) Train models, 4) Test Data, 5) Improve the design, which are the main step/steps where data visualization should be involved?
- # 1
- # 5
- # 1 and #5
- # 2, #3 and #5
2, #3 and #5
According to the discussion presented in class about the similarity, consider an image 𝐴(𝑥, 𝑦) with internal similarity (repetitions of patterns). What happens to the output of the self-cross correlation (𝑂 = 𝑥𝑐𝑜𝑟𝑟2(𝐴, 𝐴))
- It is not possible to apply the cross correlation to the same image
- Output O tends to be a flat plateau with one clear central peak
- Output O tends to have many peaks and one evident maximum
- Output O tends to have many equivalent peaks with the same maximum value
Output O tends to have many peaks and one evident maximum