In the convolutional neural network, take an image as an example, for example, the image is
First we use sparse connections, that is, each neuron is only connected to a part of the image, such as
Secondly, we use the second weapon, that is, weight sharing, then each of our neurons is connected to
Ok, so far we've used neurons to explain. Now imagine the neuron as a two-dimensional matrix, the only difference is that the original one-dimensional is changed to two-dimensional. And this shift just applies to the application of CNN to images. At this time, we also regard the previous parameters (also called weights) as a two-dimensional matrix, that is, a kernel . Use the convolution operation to operate. So corresponding to the previous, we said to connect only part of the image,
OK, at this time, only an image is output, that is, a feature map is obtained (we call the convolution operation in the above paragraph as a feature map). Usually in the image CNN, many feature maps will be generated. Suppose we generate M maps , that is, it is necessary to generate M side lengths of
The above is the core concept of the convolutional layer in CNN. The ideas used include: sparse interaction , weight sharing , and multi-feature mapping .