<?xml version="1.0" encoding="UTF-8"?>
<rss version="2.0" xmlns:atom="http://www.w3.org/2005/Atom" xmlns:dc="http://purl.org/dc/elements/1.1/">
  <channel>
    <title>DEV Community: Tanuj shrivastava</title>
    <description>The latest articles on DEV Community by Tanuj shrivastava (@tanujtj).</description>
    <link>https://dev.to/tanujtj</link>
    <image>
      <url>https://media2.dev.to/dynamic/image/width=90,height=90,fit=cover,gravity=auto,format=auto/https:%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Fuser%2Fprofile_image%2F470247%2F01b49a87-aad5-4401-8378-e6bdcce85084.jpg</url>
      <title>DEV Community: Tanuj shrivastava</title>
      <link>https://dev.to/tanujtj</link>
    </image>
    <atom:link rel="self" type="application/rss+xml" href="https://dev.to/feed/tanujtj"/>
    <language>en</language>
    <item>
      <title>7 CNN Architectures Evolved from 2010–2015</title>
      <dc:creator>Tanuj shrivastava</dc:creator>
      <pubDate>Wed, 16 Sep 2020 17:56:35 +0000</pubDate>
      <link>https://dev.to/tanujtj/7-cnn-architectures-evolved-from-2010-2015-4jc1</link>
      <guid>https://dev.to/tanujtj/7-cnn-architectures-evolved-from-2010-2015-4jc1</guid>
      <description>&lt;p&gt;&lt;a href="https://res.cloudinary.com/practicaldev/image/fetch/s--Uui8CKSi--/c_limit%2Cf_auto%2Cfl_progressive%2Cq_auto%2Cw_880/https://dev-to-uploads.s3.amazonaws.com/i/h80o0a2x27b1i25zbvyq.png" class="article-body-image-wrapper"&gt;&lt;img src="https://res.cloudinary.com/practicaldev/image/fetch/s--Uui8CKSi--/c_limit%2Cf_auto%2Cfl_progressive%2Cq_auto%2Cw_880/https://dev-to-uploads.s3.amazonaws.com/i/h80o0a2x27b1i25zbvyq.png" alt="Alt Text"&gt;&lt;/a&gt;&lt;/p&gt;

&lt;p&gt;1.ILSVRC’10&lt;br&gt;
2.ILSVRC’11&lt;br&gt;
3.ILSVRC’12 (AlexNet)&lt;br&gt;
4.ILSVRC’13 (ZFNet)&lt;br&gt;
5.ILSVRC’14 (VGGNet)&lt;br&gt;
6.ILSVRC’14 (GoogleNet)&lt;br&gt;
7.ILSVRC’15 (ResNet)&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;ILSVRC stands for Imagenet Large Scale Visual Recognition 
Challenge or the Imagenet Challenge .&lt;/li&gt;
&lt;li&gt;ILSVRC is a challenge which is held annually to allow contenders 
to classify the images correctly and generate the best possible 
predictions .&lt;/li&gt;
&lt;/ul&gt;

&lt;h1&gt;
  
  
  AlexNet
&lt;/h1&gt;

&lt;p&gt;&lt;a href="https://res.cloudinary.com/practicaldev/image/fetch/s--S5rTMO6R--/c_limit%2Cf_auto%2Cfl_progressive%2Cq_auto%2Cw_880/https://dev-to-uploads.s3.amazonaws.com/i/7bz6clux8kqhi0yd28ui.png" class="article-body-image-wrapper"&gt;&lt;img src="https://res.cloudinary.com/practicaldev/image/fetch/s--S5rTMO6R--/c_limit%2Cf_auto%2Cfl_progressive%2Cq_auto%2Cw_880/https://dev-to-uploads.s3.amazonaws.com/i/7bz6clux8kqhi0yd28ui.png" alt="Alt Text"&gt;&lt;/a&gt;&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;AlexNet CNN architecture won the 2012 ImageNet ILSVRC challenge 
by a large margin : it achieve a top-five error rate of 17% 
while the second best achieved only 26%&lt;/li&gt;
&lt;li&gt;It was developed by Alex Krizhevsky , Ilya Sutskever and 
Geoffrey Hinton .&lt;/li&gt;
&lt;li&gt;It is quite similar to LeNet-5, only much larger and deeper, and 
it was the first to stack convolutional layers directly on top 
of each other, instead of stacking a pooling layer on top of 
each convolutional layer.&lt;/li&gt;
&lt;li&gt;It consist of 8 convolutiona land fully connected layers and 3 
max pooling layers .&lt;/li&gt;
&lt;/ul&gt;

&lt;p&gt;&lt;a href="https://res.cloudinary.com/practicaldev/image/fetch/s--vJlPoj6h--/c_limit%2Cf_auto%2Cfl_progressive%2Cq_auto%2Cw_880/https://dev-to-uploads.s3.amazonaws.com/i/vd3csdbjrtr64qon5tah.png" class="article-body-image-wrapper"&gt;&lt;img src="https://res.cloudinary.com/practicaldev/image/fetch/s--vJlPoj6h--/c_limit%2Cf_auto%2Cfl_progressive%2Cq_auto%2Cw_880/https://dev-to-uploads.s3.amazonaws.com/i/vd3csdbjrtr64qon5tah.png" alt="Alt Text"&gt;&lt;/a&gt;&lt;/p&gt;

&lt;p&gt;&lt;em&gt;To reduce overfitting, the authors used two regularization techniques: first they applied dropout with a 50% dropout rate during training to the outputs of layers F8 and F9. Second, they performed data augmentation by randomly shifting the training images by various offsets, flipping them horizontally, and changing the lighting conditions.&lt;/em&gt;&lt;/p&gt;

&lt;h1&gt;
  
  
  ZFNet
&lt;/h1&gt;

&lt;p&gt;&lt;a href="https://res.cloudinary.com/practicaldev/image/fetch/s--QE15rnGj--/c_limit%2Cf_auto%2Cfl_progressive%2Cq_auto%2Cw_880/https://dev-to-uploads.s3.amazonaws.com/i/sabyrbmou8levfj6dvmo.png" class="article-body-image-wrapper"&gt;&lt;img src="https://res.cloudinary.com/practicaldev/image/fetch/s--QE15rnGj--/c_limit%2Cf_auto%2Cfl_progressive%2Cq_auto%2Cw_880/https://dev-to-uploads.s3.amazonaws.com/i/sabyrbmou8levfj6dvmo.png" alt="Alt Text"&gt;&lt;/a&gt;&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;ZFNet is another 8-layer CNN architecture .&lt;/li&gt;
&lt;li&gt;ZFNet is largely similar to AlexNet, with the exception of a few 
of the layers .&lt;/li&gt;
&lt;li&gt;ZFNet CNN architecture won the 2013 ImageNet ILSVRC challenge.&lt;/li&gt;
&lt;li&gt;One major difference in the approaches was that ZF Net used 7x7 
sized filters whereas AlexNet used 11x11 filters.&lt;/li&gt;
&lt;li&gt;The intuition behind this is that by using bigger filters we 
were losing a lot of pixel information, which we can retain by 
having smaller filter sizes in the earlier convolution layers.&lt;/li&gt;
&lt;li&gt;The number of filters increase as we go deeper. This network 
also used ReLUs for their activation and trained using batch 
stochastic gradient descent.&lt;/li&gt;
&lt;/ul&gt;

&lt;h1&gt;
  
  
  VGGNet
&lt;/h1&gt;

&lt;p&gt;&lt;a href="https://res.cloudinary.com/practicaldev/image/fetch/s--phfvORnu--/c_limit%2Cf_auto%2Cfl_progressive%2Cq_auto%2Cw_880/https://dev-to-uploads.s3.amazonaws.com/i/wlego6b33r76i26wdnrc.png" class="article-body-image-wrapper"&gt;&lt;img src="https://res.cloudinary.com/practicaldev/image/fetch/s--phfvORnu--/c_limit%2Cf_auto%2Cfl_progressive%2Cq_auto%2Cw_880/https://dev-to-uploads.s3.amazonaws.com/i/wlego6b33r76i26wdnrc.png" alt="Alt Text"&gt;&lt;/a&gt;&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;The runner up in the ILSVRC 2014 challenge was VGGNet .&lt;/li&gt;
&lt;li&gt;It was developed by K. Simon‐yan and A. Zisserman.&lt;/li&gt;
&lt;li&gt;During the design of the VGGNet, it was found that alternating 
convolution &amp;amp; pooling layers were not required. So VGGnet uses 
multiple of Convolutional layers in sequence with pooling layers 
in between.&lt;/li&gt;
&lt;li&gt;It had a very simple and classical architecture, with 2 or 3 
convolutional layers, a pooling layer, then again 2 or 3 
convolutional layers, a pooling layer, and so on (with a total 
of just 16 convolutional layers), plus a final dense net‐work 
with 2 hidden layers and the output layer. It used only 3 × 3 
filters, but many filters.&lt;/li&gt;
&lt;/ul&gt;

&lt;h1&gt;
  
  
  GoogLeNet
&lt;/h1&gt;

&lt;p&gt;&lt;a href="https://res.cloudinary.com/practicaldev/image/fetch/s--cb9PbxLi--/c_limit%2Cf_auto%2Cfl_progressive%2Cq_auto%2Cw_880/https://dev-to-uploads.s3.amazonaws.com/i/v3dch442s13vzq4qvvx3.png" class="article-body-image-wrapper"&gt;&lt;img src="https://res.cloudinary.com/practicaldev/image/fetch/s--cb9PbxLi--/c_limit%2Cf_auto%2Cfl_progressive%2Cq_auto%2Cw_880/https://dev-to-uploads.s3.amazonaws.com/i/v3dch442s13vzq4qvvx3.png" alt="Alt Text"&gt;&lt;/a&gt;&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;The GoogLeNet architecture was developed by Christian Szegedy et 
al. from Google Research,12 and it won the ILSVRC 2014 challenge 
by pushing the top-5 error rate below 7%.&lt;/li&gt;
&lt;li&gt;This great performance came in large part from the fact that the 
network was much deeper than previous CNNs&lt;/li&gt;
&lt;li&gt;This was made possible by sub-networks called inception modules, 
which allow GoogLeNet to use parameters much more efficiently 
than previous architectures: GoogLeNet actually has 10 times 
fewer parameters than AlexNet (roughly 6 million instead of 60 
million).&lt;/li&gt;
&lt;/ul&gt;

&lt;p&gt;&lt;a href="https://res.cloudinary.com/practicaldev/image/fetch/s--vHnCQFXQ--/c_limit%2Cf_auto%2Cfl_progressive%2Cq_auto%2Cw_880/https://dev-to-uploads.s3.amazonaws.com/i/t4jg4vyky3e0ro5miiul.png" class="article-body-image-wrapper"&gt;&lt;img src="https://res.cloudinary.com/practicaldev/image/fetch/s--vHnCQFXQ--/c_limit%2Cf_auto%2Cfl_progressive%2Cq_auto%2Cw_880/https://dev-to-uploads.s3.amazonaws.com/i/t4jg4vyky3e0ro5miiul.png" alt="Alt Text"&gt;&lt;/a&gt;&lt;/p&gt;

&lt;p&gt;*GoogLeNet uses 9 inception module and it eliminates all fully connected layers using average pooling to go from 7x7x1024 to 1x1x1024. This saves a lot of parameters .&lt;/p&gt;

&lt;p&gt;Several variants of the GoogLeNet architecture were later proposed by Google researchers, including Inception-v3 and Inception-v4, using slightly different inception modules, and reaching even better performance.*&lt;/p&gt;

&lt;h1&gt;
  
  
  ResNet
&lt;/h1&gt;

&lt;p&gt;&lt;a href="https://res.cloudinary.com/practicaldev/image/fetch/s--kjS3I-9E--/c_limit%2Cf_auto%2Cfl_progressive%2Cq_auto%2Cw_880/https://dev-to-uploads.s3.amazonaws.com/i/w9ggbk6fufjd640e0j7l.png" class="article-body-image-wrapper"&gt;&lt;img src="https://res.cloudinary.com/practicaldev/image/fetch/s--kjS3I-9E--/c_limit%2Cf_auto%2Cfl_progressive%2Cq_auto%2Cw_880/https://dev-to-uploads.s3.amazonaws.com/i/w9ggbk6fufjd640e0j7l.png" alt="Alt Text"&gt;&lt;/a&gt;&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;The ILSVRC 2015 challenge was won using a Residual Network (or 
ResNet)&lt;/li&gt;
&lt;li&gt;It was Developed by Kaiming He et al .&lt;/li&gt;
&lt;li&gt;It delivered an astounding top-5 error rate under 3.6%, using an 
extremely deep CNN composed of 152 layers. It confirmed the 
general trend: models are getting deeper and deeper, with fewer 
and fewer parameters. The key to being able to train such a deep 
network is to use skip connections (also called shortcut 
connections): the signal feeding into a layer is also added to 
the output of a layer located a bit higher up the stack. Let’s 
see why this is useful.&lt;/li&gt;
&lt;/ul&gt;

&lt;p&gt;&lt;a href="https://res.cloudinary.com/practicaldev/image/fetch/s--gnaCSmHY--/c_limit%2Cf_auto%2Cfl_progressive%2Cq_auto%2Cw_880/https://dev-to-uploads.s3.amazonaws.com/i/1inzt5egkc2y41wd3mwr.png" class="article-body-image-wrapper"&gt;&lt;img src="https://res.cloudinary.com/practicaldev/image/fetch/s--gnaCSmHY--/c_limit%2Cf_auto%2Cfl_progressive%2Cq_auto%2Cw_880/https://dev-to-uploads.s3.amazonaws.com/i/1inzt5egkc2y41wd3mwr.png" alt="Alt Text"&gt;&lt;/a&gt;&lt;/p&gt;

&lt;p&gt;&lt;em&gt;When training a neural network, the goal is to make it model a target function h(x). If you add the input x to the output of the network (i.e., you add a skip connection),then the network will be forced to model f(x) = h(x)-x rather than h(x). This is called residual learning .&lt;/em&gt;&lt;/p&gt;

&lt;h3&gt;
  
  
  References :
&lt;/h3&gt;

&lt;ul&gt;
&lt;li&gt;
&lt;a href="https://www.guvi.in/"&gt;https://www.guvi.in/&lt;/a&gt; (Online Learning Platform)&lt;/li&gt;
&lt;li&gt;Handson Machine Learning with Scikit (Reference Book , PDF - 
&lt;a href="https://drive.google.com/file/d/16DdwF4KIGi47ky7Q_B-"&gt;https://drive.google.com/file/d/16DdwF4KIGi47ky7Q_B-&lt;/a&gt; 
4aApvMYW2evJZ/view?usp=sharing) .&lt;/li&gt;
&lt;/ul&gt;

</description>
      <category>machinelearning</category>
      <category>datascience</category>
    </item>
  </channel>
</rss>
