ImageNet Classification with Deep Convolutional Neural Networks

Architecture:

How many layers does AlexNet have in total?

1. 8 layers - 5 convolutional and 3 fully connected

What type of neural network is AlexNet (e.g., feedforward, recurrent, convolutional)?

1. Convolutional

What activation function did AlexNet popularize?

1. ReLU

Name one regularization technique used in AlexNet to prevent overfitting.

1. Dropout

Intermediate Level:

Explain the significance of AlexNet in the context of the deep learning revolution.

  • Made CNN popular, introduced ReLU, Dropout.

ow did AlexNet utilize GPUs, and why was this important?

  • It divided the work into two gpus which utilised parallelization.

Describe the purpose and mechanism of the dropout technique used in AlexNet.

  • It was used for regularisation.

How did AlexNet's use of ReLU activation functions contribute to its success?

  • Deep convolutional neural networks with ReLUs train several times faster than their equivalents with tanh units

Advanced Level:

Explain the update rule used in AlexNet's training process, including the role of momentum and weight decay.

  • Pasted image 20240719145219.png
  • i is iteration index, v is momentum variable, ϵ is learning rate
    Lw|wiDi[1] is the average over the ith batch Di of the derivative of the objective with respect to w, evaluated at wi .

  1. How did the splitting of the network across two GPUs affect the architecture and training process?
  2. Discuss the implications of visualizing the learned features of AlexNet's first convolutional layer.
  3. Compare and contrast AlexNet's approach to data augmentation with more modern techniques.

  1. Stochastic Gradient Descent ↩︎