Neural Networks in PyTorch QUIZ (MCQ QUESTIONS AND ANSWERS)

Total Correct: 0

Time:20:00

Question: 1

Which initialization strategy is suitable for preventing exploding gradients during training?

Question: 2

Which of the following activation functions is suitable for handling the vanishing gradient problem and is widely used in practice?

Question: 3

When initializing weights in a neural network, why might zero initialization not be a suitable choice?

Question: 4

In PyTorch, which function is used to initialize weights with a normal distribution?

Question: 5

In PyTorch, which activation function is applied element-wise to the input tensor, returning zero for negative values and maintaining positive values unchanged?

Question: 6

What happens when the dropout rate is set to 0.0 in a dropout layer?

Question: 7

Which of the following statements about dropout is true?

Question: 8

What is the purpose of the He initialization strategy in neural networks?

Question: 9

Which activation function is less likely to suffer from the vanishing gradient problem compared to the sigmoid and tanh functions?

Question: 10

When might dropout not be an appropriate regularization technique?

Question: 11

What is the primary advantage of using batch normalization in neural networks?

Question: 12

How does the choice of activation function impact the learning process in a neural network?

Question: 13

In PyTorch, what is the purpose of the torch.nn.init.normal_() function?

Question: 14

Which of the following activation functions is commonly used for multi-class classification problems at the output layer?

Question: 15

What role does the learning rate play in the training of neural networks?

Question: 16

Which activation function is commonly used for binary classification problems at the output layer of a neural network?

Question: 17

Which activation function is commonly used for the hidden layers of a neural network due to its ability to mitigate the vanishing gradient problem?

Question: 18

When should you consider using a smaller learning rate during neural network training?

Question: 19

What is the purpose of using different activation functions in different layers of a neural network?

Question: 20

In PyTorch, which function is used to apply dropout to a tensor?

Question: 21

What does the dropout rate represent in the context of a dropout layer in a neural network?

Question: 22

Which activation function is often used in the hidden layers of a neural network to address the vanishing gradient problem?

Question: 23

When might you encounter the issue of vanishing gradients during training?

Question: 24

What is the primary purpose of the rectified linear unit (ReLU) activation function?

Question: 25

When training a neural network with batch normalization, how does it affect the learning rate?

Question: 26

In PyTorch, which function is used to initialize the weights of a neural network with Xavier/Glorot initialization?

Question: 27

Which initialization strategy is commonly used for initializing weights in a neural network?

Question: 28

What is the purpose of the dropout layer in a neural network?

Question: 29

Which layer is commonly used to improve the convergence and training speed of a neural network by normalizing the inputs?

Question: 30

When might you prefer to use the Leaky ReLU activation function over the standard ReLU?