CS5720 - Week 5
Slide 88 of 100

DenseNet: Dense Connections

Dense Connectivity

DenseNet takes connectivity to the extreme: each layer receives inputs from all preceding layers and feeds into all subsequent layers.
The Core Idea:
β€’ If skip connections help, why not connect everything?
β€’ Layer L receives feature maps from layers 0, 1, 2, ..., L-1
β€’ Maximum information flow and feature reuse
β€’ Dramatic parameter reduction while maintaining performance
πŸ•ΈοΈ Dense Connectivity Pattern
Each layer concatenates feature maps from all previous layers as input, creating a densely connected network structure.
🌱 Growth Rate Concept
Each layer adds a fixed number of feature maps (growth rate k), leading to linear growth in channel dimensions.

Key Benefits

  • ⚑ Parameter Efficiency
    Fewer parameters than ResNet while achieving better performance
  • 🌊 Strong Gradient Flow
    Direct connections to all previous layers eliminate vanishing gradients
  • ♻️ Feature Reuse
    All layers have access to all previous features, maximizing information usage
  • πŸ“‘ Implicit Deep Supervision
    Each layer receives direct supervision signal through shorter paths
  • 🎨 Feature Diversity
    Encourages exploration of different feature representations

DenseNet Architecture Visualization

Input
β†’
Concat: [Input]
Layer 1
β†’
Concat: [Input, L1]
Layer 2
β†’
Concat: [Input, L1, L2]
Layer 3
β†’
Concat: [Input, L1, L2, L3]
Layer 4
β†’
Concat: [Input, L1, L2, L3, L4]
🌱 Growth Rate Demonstration (k=32)
64
Input Channels
96
After Layer 1
128
After Layer 2
160
After Layer 3
192
After Layer 4
Architecture Parameters CIFAR-10 Error ImageNet Top-1
ResNet-101 44.5M 6.43% 22.63%
ResNet-152 60.2M 6.41% 21.69%
DenseNet-121 8.0M 5.77% 25.35%
DenseNet-169 14.1M 5.25% 24.00%
DenseNet achieves better accuracy with significantly fewer parameters!
Prepared by Dr. Gorkem Kar