Convolutional neural networks: 2. zero paddings
Convolutional neural networks (CNN) In the previous posting , we discussed how a (two-dimensional) convolutional layer works in comparison to a fully connected layer, which is used for a basic building block for a feedforward neural network (FNN). A convolutional layer is a basic building block for a convolutional neural network (CNN) , but there are other components for a CNN, one of which we discuss in this posting. Reference Our main reference is the lecture notes by Smets , an excellent reference for mathematicians who pursue deep learning. Review: convolutional layers Let's first recall the general setting of a convolutional layer from the previous posting . We expect $c$ matrices $X = (X[0], X[1], \dots, X[c-1])$ of fixed size, say $h \times w$ to each such matrix $X[k]$, we have $c'$ trainable matrices $$K[0,k], K[1,k], \dots, K[c'-1,k]$$ which we filter $X[k]$ with. The recipe for the weight matrix $W : \mathbb{R}^{chw} \rightarrow \mathbb{R}^{c'(h-m...