0% found this document useful (0 votes)
186 views1 page

CNN

The visual cortex contains cells that are sensitive to visual fields and orientation of edges. This is similar to convolutional neural networks (CNNs) which contain neurons that respond to vertical, horizontal, and diagonal edges. CNNs are neural networks where the parameters are shared. An image can be represented as a cuboid with length, width, and color channels. A CNN uses small filters that are slid across the image to extract features, resulting in feature maps with reduced width and height but increased channels. The convolutional layers contain learnable filters that are convolved with patches of the input volume to calculate dot products, producing a 2D output for each filter which are then stacked to form the output volume.

Uploaded by

suryavamsi
Copyright
© © All Rights Reserved
We take content rights seriously. If you suspect this is your content, claim it here.
Available Formats
Download as TXT, PDF, TXT or read online on Scribd
0% found this document useful (0 votes)
186 views1 page

CNN

The visual cortex contains cells that are sensitive to visual fields and orientation of edges. This is similar to convolutional neural networks (CNNs) which contain neurons that respond to vertical, horizontal, and diagonal edges. CNNs are neural networks where the parameters are shared. An image can be represented as a cuboid with length, width, and color channels. A CNN uses small filters that are slid across the image to extract features, resulting in feature maps with reduced width and height but increased channels. The convolutional layers contain learnable filters that are convolved with patches of the input volume to calculate dot products, producing a 2D output for each filter which are then stacked to form the output volume.

Uploaded by

suryavamsi
Copyright
© © All Rights Reserved
We take content rights seriously. If you suspect this is your content, claim it here.
Available Formats
Download as TXT, PDF, TXT or read online on Scribd
You are on page 1/ 1

The visual cortex encompasses a small region of cells that are region sensitive to

visual fields. In case some certain orientation edges are present then only some
individual neuronal cells get fired inside the brain such as some neurons responds
as and when they get exposed to the vertical edges, however some responds when they
are shown to horizontal or diagonal edges, which is nothing but the motivation
behind Convolutional Neural Networks.

The Convolutional Neural Networks, which are also called as covnets, are nothing
but neural networks, sharing their parameters. Suppose that there is an image,
which is embodied as a cuboid, such that it encompasses length, width, and height.
Here the dimensions of the image are represented by the Red, Green, and Blue
channels, as shown in the image given below.

Convolutional Neural Network


Now assume that we have taken a small patch of the same image, followed by running
a small neural network on it, having k number of outputs, which is represented in a
vertical manner. Now when we slide our small neural network all over the image, it
will result in another image constituting different width, height as well as depth.
We will notice that rather than having R, G, B channels, we have come across some
more channels that, too, with less width and height, which is actually the concept
of Convolution. In case, if we accomplished in having similar patch size as that of
the image, then it would have been a regular neural network. We have some wights
due to this small patch.

Convolutional Neural Network


Mathematically it could be understood as follows;

The Convolutional layers encompass a set of learnable filters, such that each
filter embraces small width, height as well as depth as that of the provided input
volume (if the image is the input layer then probably it would be 3).
Suppose that we want to run the convolution over the image that comprises of
34x34x3 dimension, such that the size of a filter can be axax3. Here a can be any
of the above 3, 5, 7, etc. It must be small in comparison to the dimension of the
image.
Each filter gets slide all over the input volume during the forward pass. It slides
step by step, calling each individual step as a stride that encompasses a value of
2 or 3 or 4 for higher-dimensional images, followed by calculating a dot product in
between filter's weights and patch from input volume.
It will result in 2-Dimensional output for each filter as and when we slide our
filters followed by stacking them together so as to achieve an output volume to
have a similar depth value as that of the number of filters. And then, the network
will learn all the filters.

You might also like