DEV Community

shangkyu shin
shangkyu shin

Posted on • Originally published at zeromathai.com

Evolution of Deep CNNs — From AlexNet to ResNet (Trade-offs Behind Modern Deep Learning)

Deep CNN evolution is not about deeper models — it’s about resolving engineering trade-offs under constraints.

Cross-posted from Zeromath. Original article:
https://zeromathai.com/en/deep-cnn-evolution-alexnet-resnet-en/


CNN Evolution = Constraint Evolution

Every CNN generation answers a different question:

  • AlexNet → can it work?
  • ZFNet → why does it work?
  • VGG → does depth help?
  • GoogLeNet → can we reduce compute?
  • ResNet → can we optimize deeper networks?

1. AlexNet — Feasibility

Solved:

deep CNNs can actually work at scale

Key ingredients:

  • GPU training
  • ReLU
  • Dropout
  • data augmentation

2. ZFNet — Interpretability

Solved:

understanding internal representations matters

Method:

  • feature visualization

Insight:

debugging models improves architecture design


3. VGG vs GoogLeNet — Real Trade-off

This is the key architectural tension.


VGG

  • simple architecture
  • stacked 3×3 conv
  • very deep

Problem:

compute cost explodes


GoogLeNet

  • Inception module
  • multi-scale processing
  • 1×1 conv compression

Problem:

more complex design


Trade-off

VGG GoogLeNet
simplicity efficiency
heavy compute optimized compute
depth scaling architectural branching

Insight

CNN progress is trade-off engineering, not scaling


4. ResNet — Optimization Fix

Problem:

deeper networks degrade performance

Solution:

:contentReference[oaicite:1]{index=1}

Why it works:

  • gradient flow improves
  • identity mapping preserved
  • optimization becomes easier

5. Big Picture

Model Problem solved
AlexNet feasibility
ZFNet interpretability
VGG depth scaling
GoogLeNet efficiency
ResNet optimization stability

Key Pattern

Every model follows:

  1. limitation appears
  2. root cause identified
  3. architecture changes
  4. scaling resumes

Final Insight

Deep learning is not model evolution.

It is:

continuous engineering under constraints


Discussion:

Which constraint mattered most in practice?

  • depth
  • efficiency
  • optimization

Top comments (0)