跳转至

Ch 6. Deep Learning

约 240 个字 3 张图片 预计阅读时间 1 分钟

Convolutional Neural Networks

While MLPs are universal, they struggle with spatially structured data (like images) because they treat pixels as a flat vector, ignoring local topology. Chapter 6 introduces Convolutional Neural Networks (CNNs) and the transition to modern deep learning.

ch6/cnn.png

Convolution

Instead of connecting every input pixel to every hidden neuron, we connect neurons to small, localized regions of the input. All neurons in a single feature map use the same weights and bias.

It captures local features (edges, corners).

\[ \begin{align*} a^1 &= \sigma\left(b + \sum_{l=0}^4 \sum_{m=0}^4 w_{l,m} a_{j+l, k+m}^0\right) \\ &= \sigma(b + w \ast a^0). \end{align*} \]
ch6/convolution.png
A $5 \times 5$ window (receptive field) slides across the input image.
Tip

Backpropagation:

\[ \delta_{i,j}^{l-1} = \left( \sum_{m} \sum_{n} \delta_{i-m, j-n}^l \cdot w_{m,n} \right) \sigma'(z_{i,j}^{l-1}), \\ \frac{\partial C}{\partial w_{m,n}} = \sum_{i} \sum_{j} \delta_{i,j}^l \cdot a_{i+m, j+n}^{l-1}, \\ \frac{\partial C}{\partial b} = \sum_{i} \sum_{j} \delta_{i,j}^l. \]

Pooling

Pooling simplifies the information in the output from the convolutional layer.

ch6/max_pooling.png
Max pooling.

ReLU

Deep networks prefer Rectified Linear Units (ReLU) to Sigmoid:

\[f(z) = \max (0, z)\]

Pros:

  • Solves the vanishing gradient problem for \(z > 0\) (gradient is constant 1);
  • Much faster to compute than exponentials.

Modern Techniques in Deep Learning

  • Deep Networks
  • Ensemble Learning / Dropout
  • Data Augmentation
  • ...

评论区