test1 Flashcards
- What is the primary goal of supervised learning?
A) To find hidden patterns in data
B) To predict outcomes based on labeled examples
C) To group data into clusters
D) To generate completely new data
B) To predict outcomes based on labeled examples
- What kind of problem is the βPunches-kicks-fallβ problem?
A) Regression
B) Classification
C) Clustering
D) Dimensionality Reduction
B) Classification
- Which of the following is NOT a component of a perceptron?
A) Inputs
B) Weights
C) Bias
D) Hidden Layers
D) Hidden Layers
- If the perceptron learning algorithm finds a solution, what does this indicate about the problem?
A) It is linearly separable
B) It is non-linearly separable
C) It has no possible solution
D) The algorithm is incorrect
A) It is linearly separable
- What does the perceptron training rule do when an error is detected?
A) Ignores the error
B) Updates weights using the error value
C) Adds random noise to the input
D) Decreases the learning rate
B) Updates weights using the error value
- What is the main limitation of the perceptron?
A) It can only classify numerical data
B) It cannot solve non-linearly separable problems
C) It requires a lot of computational power
D) It does not work with large datasets
B) It cannot solve non-linearly separable problems
- What is the advantage of using more hidden layers in a neural network?
A) It reduces the number of computations
B) It allows solving more complex problems
C) It makes training faster
D) It prevents overfitting
B) It allows solving more complex problems
- Which activation function is commonly used in perceptrons?
A) Sigmoid
B) ReLU
C) Hard threshold (step function)
D) Softmax
C) Hard threshold (step function)
- What is the key difference between the Perceptron and Adaline?
A) Adaline uses a different weight update rule based on error minimization
B) Perceptron uses sigmoid activation
C) Adaline cannot be trained
D) Perceptron uses gradient descent
A) Adaline uses a different weight update rule based on error minimization
- What is the formula for weight update in Adaline?
A)
π€
=
π€
+
πΌ
π₯
w=w+Ξ±x
B)
π€
=
π€
+
πΌ
π₯
πΈ
w=w+Ξ±xE
C)
π€
=
π€
+
π₯
πΈ
w=w+xE
D)
π€
=
π€
β
πΌ
π₯
w=wβΞ±x
B)
π€
=
π€
+
πΌ
π₯
πΈ
w=w+Ξ±xE
- What is the function of the learning rate (Ξ±) in weight updates?
A) It determines the speed of convergence
B) It increases the complexity of the model
C) It removes noise from the data
D) It changes the input values
A) It determines the speed of convergence
- Which type of neural network uses convolution operations for feature extraction?
A) Recurrent Neural Networks (RNNs)
B) Convolutional Neural Networks (CNNs)
C) Multi-layer Perceptrons (MLPs)
D) Support Vector Machines (SVMs)
B) Convolutional Neural Networks (CNNs)
- What is the gradient in gradient descent?
A) The sum of all errors
B) The vector of partial derivatives of the error function
C) The final weight values after training
D) A constant value used for optimization
B) The vector of partial derivatives of the error function
- Why does the perceptron fail to classify XOR correctly?
A) The XOR function is not linearly separable
B) XOR has too many inputs
C) The perceptron cannot handle binary inputs
D) The learning rate is too high
A) The XOR function is not linearly separable
- What is backpropagation used for?
A) To initialize weights randomly
B) To update weights in multi-layer neural networks
C) To generate new input data
D) To decrease the number of layers in a network
B) To update weights in multi-layer neural networks
- What is the difference between a Multi-Layer Perceptron (MLP) and a Deep Neural Network (DNN)?
A) MLPs have more than one hidden layer, while DNNs only have one
B) DNNs have multiple hidden layers, while MLPs typically have one
C) DNNs only perform classification, while MLPs perform regression
D) There is no difference; they are the same
B) DNNs have multiple hidden layers, while MLPs typically have one
- What does increasing the number of training examples generally do?
A) Reduces overfitting
B) Decreases accuracy
C) Makes training slower with no benefit
D) Eliminates the need for a learning rate
A) Reduces overfitting
- Which function is commonly used for regression problems in neural networks?
A) Softmax
B) ReLU
C) Sigmoid
D) Linear activation
D) Linear activation
- What does the sigmoid activation function do?
A) It transforms values into probabilities between 0 and 1
B) It outputs only binary values
C) It prevents gradient descent from working
D) It makes training slower
A) It transforms values into probabilities between 0 and 1
- What is the primary purpose of using multiple layers in neural networks?
A) To store more data
B) To reduce computation time
C) To learn complex representations of data
D) To make the network easier to interpret
C) To learn complex representations of data